mindmeld

A simple-to-use, open source GUI for local AI chat on desktop and mobile. Powered by llama.cpp.

GPL-3.0 License

Stars
2
Committers
1

MindMeld (Alpha Release)

A simple-to-use, open source GUI for local AI chat.

Supported platforms: iOS, Android (un-accelerated), Windows, MacOS, Linux

Features

  • Based on llama.cpp and supports any model supported by that project in GGUF form.
  • Graphical user interface, tailored towards chatting with AI, which is mobile compatible.
  • Create multiple chat logs, each with their own parameters, story context and characters
  • Customizable character portrait images.
  • Automatic model downloading from HuggingFace.
  • Built-in 'narrator' using the slash-command '/narrator' at the start of a message.
  • Lorebook support: customizable lists of 'lore' text that gets added behind the scenes
    to the LLM prompt when a pattern is matched in the chat log context or recent messages.
    Lorebooks are active across all chats and are enabled if a character name from the chat
    log pattern matches.
  • Fast regeneration of AI chat replies with prompt caching.

Instructions

See the documentation for the quick start guide.

Building From Source

See the documentation for the build instructions for all supported platforms.

Downloading Official Builds

Currently shipping the first official release binaries off to the right hand side of the main page or via that Link.

The author is new at shipping apps with more complicated, non-static library dependencies, the first attempt at deploying an AppImage for Linux builds and this is the first major deployment of a Flutter app for them, so if you try a release binary and it doesn't work, please create an issue or start a discussion!

License

This project is licensed under the GPL v3 terms, as specified in the LICENSE file.

The project is built around woolycore, woolydart for the language bindings and of course the great llama.cpp library. All three of these libraries are licensed under the MIT license.