gpt-2
Here are 971 public repositories matching this topic...
Language:All
Sort:Most stars
RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable). We are at RWKV-7 "Goose". So it's combining the best of RNN and transformer - great performance, linear time, constant space (no kv-cache), fast training, infinite ctx_len, and free sentence embedding.
- Updated
Dec 17, 2025 - Python
Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
- Updated
Dec 17, 2024 - Python
This repository contains demos I made with the Transformers library by HuggingFace.
- Updated
Jul 2, 2025 - Jupyter Notebook
AI Code Completions
- Updated
Sep 4, 2025 - Shell
[NeurIPS 2024 Best Paper Award][GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-simple, user-friendly yet state-of-the-art* codebase for autoregressive image generation!
- Updated
Nov 10, 2025 - Jupyter Notebook
An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library.
- Updated
Feb 25, 2022 - Python
Chinese version of GPT2 training code, using BERT tokenizer.
- Updated
Apr 25, 2024 - Python
An unnecessarily tiny implementation of GPT-2 in NumPy.
- Updated
Apr 24, 2023 - Python
Open Source Pre-training Model Framework in PyTorch & Pre-trained Model Zoo
- Updated
May 9, 2024 - Python
GPT2 for Chinese chitchat/用于中文闲聊的GPT2模型(实现了DialoGPT的MMI思想)
- Updated
Oct 30, 2023 - Python
Rust native ready-to-use NLP pipelines and transformer-based models (BERT, DistilBERT, GPT2,...)
- Updated
Dec 16, 2025 - Rust
Build, personalize and control your own LLMs. From data pre-processing to fine-tuning, xTuring provides an easy way to personalize open-source LLMs. Join our discord community:https://discord.gg/TgHXuSJEk6
- Updated
Dec 2, 2025 - Python
Large-scale pretraining for dialogue
- Updated
Oct 17, 2022 - Python
Kashgari is a production-level NLP Transfer learning framework built on top of tf.keras for text-labeling and text-classification, includes Word2Vec, BERT, and GPT2 Language Embedding.
- Updated
Sep 3, 2024 - Python
Toolkit for Machine Learning, Natural Language Processing, and Text Generation, in TensorFlow. This is part of the CASL project:http://casl-project.ai/
- Updated
Aug 26, 2021 - Python
Simple UI for LLM Model Finetuning
- Updated
Dec 21, 2023 - Jupyter Notebook
Guide to using pre-trained large language models of source code
- Updated
Jul 7, 2024 - Python
Improve this page
Add a description, image, and links to thegpt-2 topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with thegpt-2 topic, visit your repo's landing page and select "manage topics."