Description
🚀 Feature
The feature request is pretty much the same as this issue Running regular PyTorch with PT-XLA CUDA Container.
Motivation
It is a little bit counter intuititve that XLA CUDA and CUDA cannot coexist.
Usually XLA_GPU users and CUDA users are highly overlapped.
And a very common scenario is to compare performance between the native CUDA mode and XLA mode.
Due to current design limitation we cannot do it within the same docker.
Support XLA_GPU/CUDA cowork will greatly improve PyTorch/XLA's usability and user experience.
Pitch
- Turn on CUDA support in PyTorch/XLA docker by default
- Add certain tests to ensure XLA_GPU and CUDA could work together.
Alternatives
Additional context
Same confusion raised in this post
We disabled the pytorch wheel with
USE_CUDA=0
since there was a old bug that we can’t have both pytorch and pytorch/xla both trying to grab the CUDA device. That issue might or might not still be the case.
Same confusion 2