[NeurIPS 2023] MeZO: Fine-Tuning Language Models with Just Forward Passes. https://arxiv.org/abs/2305.17333
MIT License
Official Pytorch Implementation of "OwLore: Outlier-weighed Layerwise Sampled Low-Rank Projection...
TinyGPT-V: Efficient Multimodal Large Language Model via Small Backbones
An open platform for training, serving, and evaluating large language models. Release repo for Vi...
Implementation of the training framework proposed in Self-Rewarding Language Model, from MetaAI
MiniCPM3-4B: An edge-side LLM that surpasses GPT-3.5-Turbo.
A family of open-sourced Mixture-of-Experts (MoE) Large Language Models
Meditron is a suite of open-source medical Large Language Models (LLMs).
GLM (General Language Model)
LongLLaMA is a large language model capable of handling long contexts. It is based on OpenLLaMA a...
Ongoing research training transformer models at scale
Official Pytorch repository for Extreme Compression of Large Language Models via Additive Quantiz...
minichatgpt - To Train ChatGPT In 5 Minutes
code for the ICLR'22 paper: On Robust Prefix-Tuning for Text Classification
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.