🦙 Ollama interfaces for Neovim
MIT License
WebAssembly binding for llama.cpp - Enabling in-browser LLM inference
llama.go is like llama.cpp in pure Golang!
This repository contains a web application designed to execute relatively compact, locally-operat...
Chat with your favourite LLaMA models in a native macOS app
This repository demonstrates how to do inference with llama-2-7b-chat using llama.cpp on a machin...
LightLLM is a Python-based LLM (Large Language Model) inference and serving framework, notable fo...
LLaMa 7b with CUDA acceleration implemented in rust. Minimal GPU memory needed!
A simple "Be My Eyes" web app with a llama.cpp/llava backend
Yet another operator for running large language models on Kubernetes with ease. Powered by Ollama! 🐫
LLaMA Server combines the power of LLaMA C++ with the beauty of Chatbot UI.
Node Llama Cpp wrapper for Node JS
Run any Large Language Model behind a unified API
A self-hosted, offline, ChatGPT-like chatbot. Powered by Llama 2. 100% private, with no data leav...
A Web Interface for chatting with your local LLMs via the ollama API
vnc-lm is a Discord bot that lets you talk with and configure language models in your server. It ...