LLMs Run Riot in Production. Get Back in The Driving Seat. Build Your Own Evals, Iterate Quickly, and Go from Prototype to Production in No Time ⚡
APACHE-2.0 License
Run LLMs locally with as little friction as possible.
One .NET library to consume OpenAI, Anthropic, Cohere, Google, Azure, Groq, and self-hosed APIs.
Simple LLM library for JavaScript
GPT-4 level function calling models for real-world tool using use cases
Query LLM with Chain-of-Tought
LightLLM is a Python-based LLM (Large Language Model) inference and serving framework, notable fo...
Easy "1-line" calling of all LLMs from OpenAI, MS Azure, AWS Bedrock, GCP Vertex, and Ollama
🦙 Free and Open Source Large Language Model (LLM) chatbot web UI and API. Self-hosted, offline ca...
💬 Chatbot web app + HTTP and Websocket endpoints for LLM inference with the Petals client
A Discord LLM chat bot that supports any OpenAI compatible API. Run a local model with ollama, oo...
LLaMA Server combines the power of LLaMA C++ with the beauty of Chatbot UI.
telegram bot for self-hosted local inference of stable diffusion, text-to-speech and large langua...
The tool package for Microsoft's Prompt flow and the VS Code extension
Rust multiprovider generative AI client (Ollama, OpenAi, Anthropic, Groq, Gemini, Cohere, ...)
Friendli: the fastest serving engine for generative AI