A clone of InfiniteCraft (AI!!! LLMs!!) you can run on a laptop _without_ a good GPU!!
MIT License
AirLLM 70B inference with single 4GB GPU
This project creates a real-time conversational AI, either serverless via SvelteKit/Static or usi...
AubAI brings you on-device gen-AI capabilities, including offline text generation and more, direc...
Function-calling API for LLM from multiple providers
Simple LLM library for JavaScript
Simple large language model playground app
🏗️ Fine-tune, build, and deploy open-source LLMs easily!
LLM inference in Fortran
Maid is a cross-platform Flutter app for interfacing with GGUF / llama.cpp models locally, and wi...
llama.go is like llama.cpp in pure Golang!
A high-performance inference system for large language models, designed for production environments.
A set of bash scripts to automate deployment of GGML/GGUF models [default: RWKV] with the use of ...
A lightweight library that leverages Language Models (LLMs) to enable natural language interactio...
glai - GGUF LLAMA AI - Package for simplified model handling and text generation with Llama model...
LightLLM is a Python-based LLM (Large Language Model) inference and serving framework, notable fo...