Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema on the model output on the generation level
MIT License
Bot releases are visible (Hide)
Published by github-actions[bot] 14 days ago
Shipped with llama.cpp
release b3889
To use the latest
llama.cpp
release available, runnpx -n node-llama-cpp source download --release latest
. (learn more)
Published by github-actions[bot] 15 days ago
Shipped with llama.cpp
release b3887
To use the latest
llama.cpp
release available, runnpx -n node-llama-cpp source download --release latest
. (learn more)
Published by github-actions[bot] 25 days ago
node-llama-cpp
3.0 is here! ✨Read about the release in the blog post
Shipped with llama.cpp
release b3825
To use the latest
llama.cpp
release available, runnpx -n node-llama-cpp source download --release latest
. (learn more)
Published by github-actions[bot] 25 days ago
node-llama-cpp
3.0 is here! ✨Read about the release in the blog post
Shipped with llama.cpp
release b3821
To use the latest
llama.cpp
release available, runnpx -n node-llama-cpp source download --release latest
. (learn more)
Published by github-actions[bot] 27 days ago
node-llama-cpp
3.0 is here! ✨Read about the release in the blog post
Shipped with llama.cpp
release b3808
To use the latest
llama.cpp
release available, runnpx -n node-llama-cpp source download --release latest
. (learn more)
Published by github-actions[bot] 27 days ago
node-llama-cpp
3.0 is here! ✨Read about the release in the blog post
pull
command (#214) (453c162)inspect gpu
command (#175) (5a70576)inspect gguf
command (#182) (35e6f50)inspect estimate
command (#309) (4b3ad61)inspect measure
command (#182) (35e6f50)init
command to scaffold a new project from a template (with node-typescript
and electron-typescript-react
templates) (#217) (d6a0f43)download
, build
and clear
commands to be subcommands of a source
command (#309) (4b3ad61)seed
option to the prompt level (#309) (4b3ad61)TemplateChatWrapper
: custom history template for each message role (#309) (4b3ad61)onTextChunk
option (#273) (e3e0994)stopOnAbortSignal
and customStopTriggers
on LlamaChat
and LlamaChatSession
(#214) (453c162)--gpu
flag in generation CLI commands (#205) (ef501f9)specialTokens
parameter on model.detokenize
(#205) (ef501f9)LlamaModel
(#182) (35e6f50)tokenizer.chat_template
header from the gguf
file when available - use it to find a better specialized chat wrapper or use JinjaTemplateChatWrapper
with it as a fallback (#182) (35e6f50)chat
, complete
, infill
(#182) (35e6f50)getLlama
when using "lastBuild"
(#164) (ede69c1)chatWrapper
getter on a LlamaChatSession
(#161) (46235a2)LlamaChat
(#139) (5fcdf9b)LlamaText
util (#139) (5fcdf9b)llama.cpp
release in GitHub releases (#142) (36c779d)Shipped with llama.cpp
release b3808
To use the latest
llama.cpp
release available, runnpx -n node-llama-cpp source download --release latest
. (learn more)
Published by github-actions[bot] 27 days ago
Shipped with llama.cpp
release b3804
To use the latest
llama.cpp
release available, runnpx -n node-llama-cpp source download --release latest
. (learn more)
Published by github-actions[bot] about 1 month ago
defineChatSessionFunction
types and docs (#322) (2204e7a)electron-builder
version used in Electron template (#323) (6c644ff)Shipped with llama.cpp
release b3787
To use the latest
llama.cpp
release available, runnpx -n node-llama-cpp source download --release latest
. (learn more)
Published by github-actions[bot] about 1 month ago
llama.cpp
sampling refactor (#309) (4b3ad61)chat
command when using --printTimings
or --meter
(#309) (4b3ad61)inspect estimate
command (#309) (4b3ad61)seed
option to the prompt level (#309) (4b3ad61)autoDisposeSequence
default to false
(#309) (4b3ad61)download
, build
and clear
commands to be subcommands of a source
command (#309) (4b3ad61)TokenBias
(#309) (4b3ad61)threads
default value (#309) (4b3ad61)LlamaEmbedding
an object (#309) (4b3ad61)HF_TOKEN
env var support for reading GGUF file metadata (#309) (4b3ad61)TemplateChatWrapper
: custom history template for each message role (#309) (4b3ad61)inspect gpu
command (#309) (4b3ad61)--gpuLayers max
and --contextSize max
flag support for inspect estimate
command (#309) (4b3ad61)Shipped with llama.cpp
release b3785
To use the latest
llama.cpp
release available, runnpx -n node-llama-cpp source download --release latest
. (learn more)
Shipped with llama.cpp
release b3543
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)
Published by github-actions[bot] 2 months ago
Shipped with llama.cpp
release b3560
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)
Published by github-actions[bot] 2 months ago
Shipped with llama.cpp
release b3541
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)
Published by github-actions[bot] 3 months ago
Shipped with llama.cpp
release b3504
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)
Published by github-actions[bot] 3 months ago
Shipped with llama.cpp
release b3488
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)
Published by github-actions[bot] 3 months ago
llama.cpp
breaking changes (#273) (e3e0994)onTextChunk
option (#273) (e3e0994)Shipped with llama.cpp
release b3479
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)
Published by github-actions[bot] 3 months ago
llama.cpp
breaking changes (#266) (c35ff5a)Shipped with llama.cpp
release b3347
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)