-
Notifications
You must be signed in to change notification settings - Fork 811
[slimtensor] Enable CUDA tensor copy #16800
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Pull Request resolved: #16769 This diff adds CUDA storage infrastructure to SlimTensor, enabling GPU memory allocation and management. **Key changes:** 1. **`cuda/Guard.h`** - CUDAGuard RAII class: - Saves current CUDA device on construction, restores on destruction - Exception-safe device context switching - Constructors accept device index or Device object 2. **`core/Storage.h`** - Extended for CUDA support: - Added `DeviceTraits<DeviceType::CUDA>` specialization with: - `allocate()` - Uses cudaMalloc with CUDAGuard for device selection - `free()` - Uses cudaFree with warning on error - `memcpy()` - Supports Host↔Device and Device↔Device copies - Added `DEFAULT_CUDA_DEVICE` constant - Updated `MaybeOwningStorage` constructor to handle CUDA devices - Stub implementation when `CUDA_AVAILABLE` is not defined (throws error) ghstack-source-id: 335102161 @exported-using-ghexport Differential Revision: [D91202899](https://our.internmc.facebook.com/intern/diff/D91202899/)
Pull Request resolved: #16770 This diff enables CUDA tensor creation with basic tensor functionality and factory function support **Key changes:* 1. **`core/SlimTensor.h`** - Extended for CUDA support: - Added `is_cuda()` method to check if tensor is on CUDA device 2. **`factory/Empty.h`** - Supports CUDA: - `empty_strided()` and `empty()` work with CUDA device via `new_storage()` - Device routing is handled by `MaybeOwningStorage` constructor ghstack-source-id: 335102160 @exported-using-ghexport Differential Revision: [D91202897](https://our.internmc.facebook.com/intern/diff/D91202897/)
Pull Request resolved: #16771 This diff enables CUDA tensor copy operations in SlimTensor. **Key changes:** **`core/SlimTensor.h`** - Extended for CUDA support: - Updated `copy_()` to handle cross-device copies: - CPU→CUDA (cudaMemcpyHostToDevice) - CUDA→CPU (cudaMemcpyDeviceToHost) - CUDA→CUDA (cudaMemcpyDeviceToDevice, same device) - Cross-device copies require contiguous tensors - CPU-to-CPU copies continue to support non-contiguous (strided) tensors ghstack-source-id: 335102159 @exported-using-ghexport Differential Revision: [D91202900](https://our.internmc.facebook.com/intern/diff/D91202900/)
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/16800
Note: Links to docs will display an error until the docs builds have been completed. ⏳ No Failures, 91 PendingAs of commit ffb18f0 with merge base 8ab593b ( This comment was automatically generated by Dr. CI and updates every 15 minutes. |
This PR needs a
|
This PR was created by the merge bot to help merge the original PR into the main branch.
ghstack PR number: #16771 by @Gasoonjia
^ Please use this as the source of truth for the PR details, comments, and reviews
ghstack PR base: https://github.com/pytorch/executorch/tree/gh/gasoonjia/111/base
ghstack PR head: https://github.com/pytorch/executorch/tree/gh/gasoonjia/111/head
Merge bot PR base: https://github.com/pytorch/executorch/tree/gh/gasoonjia/110/orig
Merge bot PR head: https://github.com/pytorch/executorch/tree/gh/gasoonjia/111/orig
Differential Revision: D91202900
@diff-train-skip-merge