Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings

FB (Facebook) + GEMM (General Matrix-Matrix Multiplication) -https://code.fb.com/ml-applications/fbgemm/

License

NotificationsYou must be signed in to change notification settings

pytorch/FBGEMM

Repository files navigation

The FBGEMM Project is a repository of highly-optimized kernels used acrossdeep learning applications.

The codebase is organized and published as three related packages: FBGEMM,FBGEMM-GPU, and FBGEMM-GenAI. Each package has its own set of features anddocumentation.

Project Overview

  • FBGEMM: A low-precision, high-performance matrix multiplication andconvolution library for server-side inference. The documentation belowprovides an overview of FBGEMM, including its features, documentation, andcommunity resources.

  • FBGEMM_GPU: A collection of PyTorch GPU operator libraries built on top ofFBGEMM for training and inference, with focus on recommendation systemsapplications. Please seethe documentation for moreinformation.

  • FBGEMM_GPU GenAI: A collection of PyTorch GPU operator libraries that aredesigned for generative AI applications, such as FP8 row-wise quantization andcollective communications. Please seethe documentationfor more information.

FBGEMM

FBGEMM CI

FBGEMM (Facebook GEneral Matrix Multiplication) is a low-precision,high-performance matrix-matrix multiplications and convolution library forserver-side inference.

The library provides efficient low-precision general matrix multiplication forsmall batch sizes and support for accuracy-loss minimizing techniques such asrow-wise quantization and outlier-aware quantization. FBGEMM also exploitsfusion opportunities in order to overcome the unique challenges of matrixmultiplication at lower precision with bandwidth-bound operations.

FBGEMM is used as a backend of PyTorch quantized operators for x86 machines:

See the fullDocumentation for more informationon building, installing, and developing with FBGEMM, as well as the mostup-to-date support matrix and API documentation for this library.

What's New?

Citation

For a high-level overview, design philosophy and brief descriptions of variousparts of FBGEMM please seeour blog post.

For those looking for the appropriate article to cite regarding FBGEMM, werecommend citing ourpaper:

@article{fbgemm,  title={FBGEMM: Enabling High-Performance Low-Precision Deep Learning Inference},  author={Khudia, Daya and Huang, Jianyu and Basu, Protonu and Deng, Summer and Liu, Haixin and Park, Jongsoo and Smelyanskiy, Mikhail},  journal={arXiv preprint arXiv:2101.05615},  year={2021}}

Join the FBGEMM community

For questions, support, news updates, or feature requests, please feel free to:

For contributions, please see theCONTRIBUTING file forways to help out.

License

FBGEMM is BSD licensed, as found in theLICENSE file.

About

FB (Facebook) + GEMM (General Matrix-Matrix Multiplication) -https://code.fb.com/ml-applications/fbgemm/

Resources

License

Code of conduct

Contributing

Stars

Watchers

Forks

Packages

No packages published

[8]ページ先頭

©2009-2025 Movatter.jp