A Jax-based library for designing and training transformer models from scratch.
MIT License
A high-performance inference system for large language models, designed for production environments.
AirLLM 70B inference with single 4GB GPU
Interact with NanoGPT's API for pay-per-prompt interaction with AI models
A throughput-oriented high-performance serving framework for LLMs
An efficient, flexible and full-featured toolkit for fine-tuning LLM (InternLM2, Llama3, Phi3, Qw...
🦖 X—LLM: Cutting Edge & Easy LLM Finetuning
Large language models (LLMs) made easy, EasyLM is a one stop solution for pre-training, finetunin...
KoAlpaca: 한국어 명령어를 이해하는 오픈소스 언어모델
[ICLR 2024] Sheared LLaMA: Accelerating Language Model Pre-training via Structured Pruning
Build, customize and control you own LLMs. From data pre-processing to fine-tuning, xTuring provi...
A lightweight library that leverages Language Models (LLMs) to enable natural language interactio...
High-speed Large Language Model Serving on PCs with Consumer-grade GPUs
Easy and Efficient Finetuning LLMs. (Supported LLama, LLama2, LLama3, Qwen, Baichuan, GLM , Fal...
LightLLM is a Python-based LLM (Large Language Model) inference and serving framework, notable fo...
Lightweight inference library for ONNX files, written in C++. It can run Stable Diffusion XL 1.0 ...