Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings

[CUDA][cuBLASLt] Fix scale setting forallowFP16AccumulationCuBLAStrue case#153083

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to ourterms of service andprivacy statement. We’ll occasionally send you account related emails.

Already on GitHub?Sign in to your account

Closed
eqy wants to merge3 commits intopytorch:mainfromeqy:fixscalefp16

Conversation

@eqy
Copy link
Collaborator

@eqyeqy commentedMay 7, 2025
edited by pytorch-botbot
Loading

Also add some missing@onlyCUDA / support check decorators intest_matmul_cuda.py
Should helpresolve#151890

cc@ptrblck@msaroufim@jerryzh168@csarofeen@xwang233

@eqyeqy requested a review fromsyed-ahmed as acode ownerMay 7, 2025 19:06
@eqyeqy added module: cudaRelated to torch.cuda, and CUDA support in general module: cublasProblem related to cublas support open source module: halfRelated to float16 half-precision floats topic: not user facingtopic category labelsMay 7, 2025
@pytorch-bot
Copy link

pytorch-botbot commentedMay 7, 2025
edited
Loading

🔗 Helpful Links

🧪 See artifacts and rendered test results athud.pytorch.org/pr/153083

Note: Links to docs will display an error until the docs builds have been completed.

✅ You can merge normally! (1 Unrelated Failure)

As of commit167c4a6 with merge base590965f (image):

BROKEN TRUNK - The following job failed but were present on the merge base:

👉Rebase onto the `viable/strict` branch to avoid these failures

This comment was automatically generated by Dr. CI and updates every 15 minutes.

cudaDeviceProp* prop =at::cuda::getCurrentDeviceProperties();
if (prop->major >=7 &&at::globalContext().allowFP16AccumulationCuBLAS()) {
computeType = CUBLAS_COMPUTE_16F;
scaleType = CUDA_R_16F;
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others.Learn more.

👀

Copy link
Contributor

@janeyx99janeyx99 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others.Learn more.

looks ok

@janeyx99janeyx99 added release notes: cudarelease notes category triagedThis issue has been looked at a team member, and triaged and prioritized into an appropriate module and removed topic: not user facingtopic category labelsMay 7, 2025
@eqy
Copy link
CollaboratorAuthor

eqy commentedMay 8, 2025

@pytorchmergebot merge

pytorch-bot[bot] reacted with thumbs up emoji

@pytorch-botpytorch-botbot added the ciflow/trunkTrigger trunk jobs on your pull request labelMay 8, 2025
@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged once all checks pass (ETA 0-4 Hours).

Learn more about merging in thewiki.

Questions? Feedback? Please reach out to thePyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

@pytorchmergebot
Copy link
Collaborator

Merge failed

Reason: Commandgit -C /home/runner/work/pytorch/pytorch rebase origin/main returned non-zero exit code 1

Rebasing (1/1)Auto-merging test/test_matmul_cuda.pyCONFLICT (content): Merge conflict in test/test_matmul_cuda.pyerror: could not apply c5d1ed74a17... [CUDA][cuBLASLt] Fix scale setting for `allowFP16AccumulationCuBLAS` `true` case (#153083)hint: Resolve all conflicts manually, mark them as resolved withhint: "git add/rm <conflicted_files>", then run "git rebase --continue".hint: You can instead skip this commit: run "git rebase --skip".hint: To abort and get back to the state before "git rebase", run "git rebase --abort".hint: Disable this message with "git config set advice.mergeConflict false"Could not apply c5d1ed74a17... [CUDA][cuBLASLt] Fix scale setting for `allowFP16AccumulationCuBLAS` `true` case (#153083)
Details for Dev Infra teamRaised byworkflow job

@eqy
Copy link
CollaboratorAuthor

eqy commentedMay 8, 2025

@pytorchmergebot merge

pytorch-bot[bot] reacted with thumbs up emoji

@pytorchmergebot
Copy link
Collaborator

Merge started

Your change will be merged once all checks pass (ETA 0-4 Hours).

Learn more about merging in thewiki.

Questions? Feedback? Please reach out to thePyTorch DevX Team

Advanced Debugging
Check the merge workflow status
here

Sign up for freeto join this conversation on GitHub. Already have an account?Sign in to comment

Reviewers

@janeyx99janeyx99janeyx99 approved these changes

@syed-ahmedsyed-ahmedAwaiting requested review from syed-ahmed

Assignees

No one assigned

Labels

ciflow/trunkTrigger trunk jobs on your pull requestMergedmodule: cublasProblem related to cublas supportmodule: cudaRelated to torch.cuda, and CUDA support in generalmodule: halfRelated to float16 half-precision floatsopen sourcerelease notes: cudarelease notes categorytriagedThis issue has been looked at a team member, and triaged and prioritized into an appropriate module

Projects

None yet

Milestone

No milestone

Development

Successfully merging this pull request may close these issues.

DISABLED test_cublas_and_lt_reduced_precision_fp16_accumulate_cuda (__main__.TestMatmulCudaCUDA)

3 participants

@eqy@pytorchmergebot@janeyx99

[8]ページ先頭

©2009-2025 Movatter.jp