Skip to content

Support XLA_GPU and CUDA cowork by default #4713

Closed
@kevint324

Description

@kevint324

🚀 Feature

The feature request is pretty much the same as this issue Running regular PyTorch with PT-XLA CUDA Container.

Motivation

It is a little bit counter intuititve that XLA CUDA and CUDA cannot coexist.
Usually XLA_GPU users and CUDA users are highly overlapped.
And a very common scenario is to compare performance between the native CUDA mode and XLA mode.
Due to current design limitation we cannot do it within the same docker.

Support XLA_GPU/CUDA cowork will greatly improve PyTorch/XLA's usability and user experience.

Pitch

  1. Turn on CUDA support in PyTorch/XLA docker by default
  2. Add certain tests to ensure XLA_GPU and CUDA could work together.

Alternatives

Additional context

Same confusion raised in this post

We disabled the pytorch wheel with USE_CUDA=0 since there was a old bug that we can’t have both pytorch and pytorch/xla both trying to grab the CUDA device. That issue might or might not still be the case.

Same confusion 2

#3452

Metadata

Metadata

Assignees

No one assigned

    Labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions