- Notifications
You must be signed in to change notification settings - Fork26.3k
[ROCm] cpp_extension allow user to override default flags#152432
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to ourterms of service andprivacy statement. We’ll occasionally send you account related emails.
Already on GitHub?Sign in to your account
Uh oh!
There was an error while loading.Please reload this page.
Conversation
pytorch-botbot commentedApr 29, 2025 • edited
Loading Uh oh!
There was an error while loading.Please reload this page.
edited
Uh oh!
There was an error while loading.Please reload this page.
🔗 Helpful Links🧪 See artifacts and rendered test results athud.pytorch.org/pr/152432
Note: Links to docs will display an error until the docs builds have been completed. ❌ 1 New Failure, 1 Unrelated FailureAs of commit0bed116 with merge basee06a080 ( NEW FAILURE - The following job has failed:
UNSTABLE - The following job is marked as unstable, possibly due to flakiness on trunk:
This comment was automatically generated by Dr. CI and updates every 15 minutes. |
liligwu commentedApr 30, 2025
rocshmem shares the same global variable in different files, as deepEP uses CUDAExtention to build the projecthttps://github.com/deepseek-ai/DeepEP/blob/65e2a700f0330f3fb1c26f49a0250d1f9d0ac1e3/setup.py#L51 and depends on rocshmem, this |
jithunnair-amd commentedMay 4, 2025
Attempting to rebase so that we can get a clean run without a test_cpp_extensions* failure @pytorchbot rebase |
pytorchmergebot commentedMay 4, 2025
@pytorchbot started a rebase job ontorefs/remotes/origin/viable/strict. Check the current statushere |
pytorchmergebot commentedMay 4, 2025
Successfully rebased |
78ebe3d toa61a149Comparejithunnair-amd commentedMay 8, 2025
@malfet Can you please review? |
jeffdaily commentedMay 9, 2025
@pytorchbot rebase |
pytorchmergebot commentedMay 9, 2025
@pytorchbot started a rebase job ontorefs/remotes/origin/viable/strict. Check the current statushere |
pytorchmergebot commentedMay 9, 2025
Successfully rebased |
76c2c3b to372bc06Comparejithunnair-amd commentedMay 15, 2025
@pytorchbot merge -f "unrelated CI failures" |
pytorchmergebot commentedMay 15, 2025
Merge startedYour change will be merged immediately since you used the force (-f) flag,bypassing any CI checks (ETA: 1-5 minutes). Please use Learn more about merging in thewiki. Questions? Feedback? Please reach out to thePyTorch DevX Team |
…2432)We need -fgpu-rdc for projects such as DeepEP + rocSHMEM. The default of -no-gpu-rdc doesn't work for such cases.As perpytorch#152432 (comment):"rocshmem shares the same global variable in different files, as deepEP uses CUDAExtention to build the projecthttps://github.com/deepseek-ai/DeepEP/blob/65e2a700f0330f3fb1c26f49a0250d1f9d0ac1e3/setup.py#L51 and depends on rocshmem, this -fgpu-rdc is needed. The current logic in Pytorch prevents users from overriding this flag."Pull Requestresolved:pytorch#152432Approved by:https://github.com/jeffdailyCo-authored-by: Jeff Daily <jeff.daily@amd.com>
…2432) (#2374)cherry-pick ofpytorch@e4adf5dWe need -fgpu-rdc for projects such as DeepEP + rocSHMEM. The default of-no-gpu-rdc doesn't work for such cases.As perpytorch#152432 (comment):"rocshmem shares the same global variable in different files, as deepEPuses CUDAExtention to build the projecthttps://github.com/deepseek-ai/DeepEP/blob/65e2a700f0330f3fb1c26f49a0250d1f9d0ac1e3/setup.py#L51and depends on rocshmem, this -fgpu-rdc is needed. The current logic inPytorch prevents users from overriding this flag."Pull Requestresolved:pytorch#152432Approved by:https://github.com/jeffdailyCo-authored-by: Jithun Nair <37884920+jithunnair-amd@users.noreply.github.com>Co-authored-by: Jeff Daily <jeff.daily@amd.com>
…2432) (#2374)cherry-pick ofpytorch@e4adf5dWe need -fgpu-rdc for projects such as DeepEP + rocSHMEM. The default of-no-gpu-rdc doesn't work for such cases.As perpytorch#152432 (comment):"rocshmem shares the same global variable in different files, as deepEPuses CUDAExtention to build the projecthttps://github.com/deepseek-ai/DeepEP/blob/65e2a700f0330f3fb1c26f49a0250d1f9d0ac1e3/setup.py#L51and depends on rocshmem, this -fgpu-rdc is needed. The current logic inPytorch prevents users from overriding this flag."Pull Requestresolved:pytorch#152432Approved by:https://github.com/jeffdailyCo-authored-by: Jithun Nair <37884920+jithunnair-amd@users.noreply.github.com>Co-authored-by: Jeff Daily <jeff.daily@amd.com>
…2432) (#2374)cherry-pick ofpytorch@e4adf5dWe need -fgpu-rdc for projects such as DeepEP + rocSHMEM. The default of-no-gpu-rdc doesn't work for such cases.As perpytorch#152432 (comment):"rocshmem shares the same global variable in different files, as deepEPuses CUDAExtention to build the projecthttps://github.com/deepseek-ai/DeepEP/blob/65e2a700f0330f3fb1c26f49a0250d1f9d0ac1e3/setup.py#L51and depends on rocshmem, this -fgpu-rdc is needed. The current logic inPytorch prevents users from overriding this flag."Pull Requestresolved:pytorch#152432Approved by:https://github.com/jeffdailyCo-authored-by: Jithun Nair <37884920+jithunnair-amd@users.noreply.github.com>Co-authored-by: Jeff Daily <jeff.daily@amd.com>
…2432) (#2374)cherry-pick ofpytorch@e4adf5dWe need -fgpu-rdc for projects such as DeepEP + rocSHMEM. The default of-no-gpu-rdc doesn't work for such cases.As perpytorch#152432 (comment):"rocshmem shares the same global variable in different files, as deepEPuses CUDAExtention to build the projecthttps://github.com/deepseek-ai/DeepEP/blob/65e2a700f0330f3fb1c26f49a0250d1f9d0ac1e3/setup.py#L51and depends on rocshmem, this -fgpu-rdc is needed. The current logic inPytorch prevents users from overriding this flag."Pull Requestresolved:pytorch#152432Approved by:https://github.com/jeffdailyCo-authored-by: Jithun Nair <37884920+jithunnair-amd@users.noreply.github.com>Co-authored-by: Jeff Daily <jeff.daily@amd.com>
…2432) (#2374)cherry-pick ofpytorch@e4adf5dWe need -fgpu-rdc for projects such as DeepEP + rocSHMEM. The default of-no-gpu-rdc doesn't work for such cases.As perpytorch#152432 (comment):"rocshmem shares the same global variable in different files, as deepEPuses CUDAExtention to build the projecthttps://github.com/deepseek-ai/DeepEP/blob/65e2a700f0330f3fb1c26f49a0250d1f9d0ac1e3/setup.py#L51and depends on rocshmem, this -fgpu-rdc is needed. The current logic inPytorch prevents users from overriding this flag."Pull Requestresolved:pytorch#152432Approved by:https://github.com/jeffdailyCo-authored-by: Jithun Nair <37884920+jithunnair-amd@users.noreply.github.com>Co-authored-by: Jeff Daily <jeff.daily@amd.com>
Uh oh!
There was an error while loading.Please reload this page.
We need -fgpu-rdc for projects such as DeepEP + rocSHMEM. The default of -no-gpu-rdc doesn't work for such cases.
As per#152432 (comment):
"rocshmem shares the same global variable in different files, as deepEP uses CUDAExtention to build the projecthttps://github.com/deepseek-ai/DeepEP/blob/65e2a700f0330f3fb1c26f49a0250d1f9d0ac1e3/setup.py#L51 and depends on rocshmem, this -fgpu-rdc is needed. The current logic in Pytorch prevents users from overriding this flag."
cc@jeffdaily@sunway513@pruthvistony@ROCmSupport@dllehr-amd@jataylo@hongxiayang@naromero77amd