- Guangzhou, China
- 03:45
(UTC +08:00) - https://github.com/xlite-dev
🏢 Group: Owner. @xlite-dev | @vipshop | Prev. @PaddlePaddle 🏰
🛠 Creator: lite.ai.toolkit | Awesome-LLM-Inference | LeetCUDA | ffpa-attn 🎧
🖥 HGEMM | 🤗cache-dit | Awesome-DiT-Inference | torchlm🖱
🎉 Contributor: FastDeploy | vLLM | SGLang | Many Others ⚙️
✉️ Contact: qyjdef@163.com | GitHub: DefTruth | 知乎: DefTruth 🤖
PinnedLoading
- xlite-dev/LeetCUDA
xlite-dev/LeetCUDA Public📚LeetCUDA: Modern CUDA Learn Notes with PyTorch for Beginners🐑, 200+ CUDA Kernels, Tensor Cores, HGEMM, FA-2 MMA.🎉
- xlite-dev/lite.ai.toolkit
xlite-dev/lite.ai.toolkit Public🛠 A lite C++ AI toolkit: 100+ models with MNN, ORT and TRT, including Det, Seg, Stable-Diffusion, Face-Fusion, etc.🎉
- xlite-dev/Awesome-LLM-Inference
xlite-dev/Awesome-LLM-Inference Public📚A curated list of Awesome LLM/VLM Inference Papers with Codes: Flash-Attention, Paged-Attention, WINT8/4, Parallelism, etc.🎉
- vllm-project/vllm
vllm-project/vllm PublicA high-throughput and memory-efficient inference and serving engine for LLMs
- PaddlePaddle/FastDeploy
PaddlePaddle/FastDeploy PublicHigh-performance Inference and Deployment Toolkit for LLMs and VLMs based on PaddlePaddle
- vipshop/cache-dit
vipshop/cache-dit Public🤗CacheDiT: A Training-free and Easy-to-use Cache Acceleration Toolbox for Diffusion Transformers🔥
If the problem persists, check theGitHub status page orcontact support.
Uh oh!
There was an error while loading.Please reload this page.