[ENH] Optimize PyTorch backend and refactor kernel functions for GPU performance#44
Draft
Leguark wants to merge 7 commits intooctree_improvementfrom
Draft
[ENH] Optimize PyTorch backend and refactor kernel functions for GPU performance#44Leguark wants to merge 7 commits intooctree_improvementfrom
Leguark wants to merge 7 commits intooctree_improvementfrom
Conversation
…r PyTorch backend - Ensure edge vectors are placed on the same device as input data - Add support for creating scalar tensors on the correct device and dtype based on backend tensor engine
…kend - Introduced `@torch.compile` to several kernel functions for performance optimization - Refactored cubic, exponential, and Matern functions for improved stability and reduced computational overhead - Enhanced code comments for better clarity and maintainability
…Torch - Implemented NumPy-based kernel functions as the default backend - Refactored PyTorch kernel functions with lazy loading for optional JIT optimizations - Updated `KernelFunction` to allow switching between NumPy and PyTorch implementations - Enhanced flexibility in `AvailableKernelFunctions` to support dual-backend configurations
Member
Author
|
Warning This pull request is not mergeable via GitHub because a downstack PR is open. Once all requirements are satisfied, merge this PR as a stack on Graphite.
This stack of pull requests is managed by Graphite. Learn more about stacking. |
_compress_binary_indices for PyTorch backend
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.

[ENH] Improve GPU performance and tensor handling for PyTorch backend
torch.set_default_device("cuda")when GPU is requested_arraymethodcdistandeinsum[ENH] Add JIT-compiled kernel functions for PyTorch backend
@torch.compileoptimized versions of kernel functions[ENH] Improve tensor device handling in data structures
_compress_binary_indices_secure_casthelper to safely handle tensor creation across backends[ENH] Optimize memory layout for tensor operations