Finetune llama2-70b and codellama on MacBook Air without quantization
MIT License
We unified the interfaces of instruction-tuning data (e.g., CoT data), multiple LLMs and paramete...
Run any Llama 2 locally with gradio UI on GPU or CPU from anywhere (Linux/Windows/Mac). Use `llam...
High-speed Large Language Model Serving on PCs with Consumer-grade GPUs
AirLLM 70B inference with single 4GB GPU
Easy and Efficient Finetuning LLMs. (Supported LLama, LLama2, LLama3, Qwen, Baichuan, GLM , Fal...
Train Llama 2 & 3 on the SQuAD v2 task as an example of how to specialize a generalized (foundati...
Inference Llama 2 in one file of pure Zig
LLaMA-2 in native Go
Tensor parallelism is all you need. Run LLMs on an AI cluster at home using any device. Distribut...
Build, customize and control you own LLMs. From data pre-processing to fine-tuning, xTuring provi...
llama.go is like llama.cpp in pure Golang!
Finetune Llama 3.2, Mistral, Phi & Gemma LLMs 2-5x faster with 80% less memory
LLM inference in Fortran
Chinese-Vicuna: A Chinese Instruction-following LLaMA-based Model —— 一个中文低资源的llama+lora方案,结构参考alpaca
🦖 X—LLM: Cutting Edge & Easy LLM Finetuning