Fine tuning experiments for the GPT-2 model by OpenAI.
OTHER License
Ongoing research training transformer models at scale
Home of StarCoder2!
Chain-of-Hindsight, A Scalable RLHF Method
Ongoing research training transformer language models at scale, including: BERT & GPT-2
The simplest, fastest repository for training/finetuning medium-sized GPTs.
An implementation of training for GPT2, supports TPUs
GLM (General Language Model)
Unsupervised Language Modeling at scale for robust sentiment classification
Find better generation parameters for your LLM
TinyGPT-V: Efficient Multimodal Large Language Model via Small Backbones
Guide to using pre-trained large language models of source code
A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training
[GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregress...
Home of StarCoder: fine-tuning & inference!
Video+code lecture on building nanoGPT from scratch