- Notifications
You must be signed in to change notification settings - Fork566
flex_attention ver.#192
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to ourterms of service andprivacy statement. We’ll occasionally send you account related emails.
Already on GitHub?Sign in to your account
base:main
Are you sure you want to change the base?
Uh oh!
There was an error while loading.Please reload this page.
Conversation
joydddd commentedJul 30, 2024 • edited
Loading Uh oh!
There was an error while loading.Please reload this page.
edited
Uh oh!
There was an error while loading.Please reload this page.
sdpa_kernel: |
facebook-github-bot commentedAug 17, 2024
Hi@joydddd! Thank you for your pull request. Werequire contributors to sign ourContributor License Agreement, and yours needs attention. You currently have a record in our system, but the CLA is no longer valid, and will need to beresubmitted. ProcessIn order for us to review and merge your suggested changes, please sign athttps://code.facebook.com/cla.If you are contributing on behalf of someone else (eg your employer), the individual CLA may not be sufficient and your employer may need to sign the corporate CLA. Once the CLA is signed, our tooling will perform checks and validations. Afterwards, thepull request will be tagged with If you have received this in error or have any questions, please contact us atcla@meta.com. Thanks! |
Uh oh!
There was an error while loading.Please reload this page.
Implement gpt-fast using flex_attention HOP.
replies on this PR:pytorch/pytorch#132157