node-llama-cpp

Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema on the model output on the generation level

MIT License

Downloads
37.6K
Stars
905
Committers
6

Bot releases are visible (Hide)

node-llama-cpp - v3.0.0-beta.19

Published by github-actions[bot] 5 months ago

3.0.0-beta.19 (2024-05-12)

Bug Fixes

Features


Shipped with llama.cpp release b2861

To use the latest llama.cpp release available, run npx --no node-llama-cpp download --release latest. (learn more)

node-llama-cpp - v3.0.0-beta.18

Published by github-actions[bot] 5 months ago

3.0.0-beta.18 (2024-05-09)

Bug Fixes

  • more efficient max context size finding algorithm (#214) (453c162)
  • make embedding-only models work correctly (#214) (453c162)
  • perform context shift on the correct token index on generation (#214) (453c162)
  • make context loading work for all models on Electron (#214) (453c162)

Features


Shipped with llama.cpp release b2834

To use the latest llama.cpp release available, run npx --no node-llama-cpp download --release latest. (learn more)

node-llama-cpp - v2.8.10

Published by github-actions[bot] 6 months ago

2.8.10 (2024-04-27)

Bug Fixes

node-llama-cpp - v3.0.0-beta.17

Published by github-actions[bot] 6 months ago

3.0.0-beta.17 (2024-04-24)

Bug Fixes

  • FunctionaryChatWrapper bugs (#205) (ef501f9)
  • function calling syntax bugs (#205) ([ef501f9]
  • show GPU layers in the Model line in CLI commands (#205) ([ef501f9]
  • refactor: rename LlamaChatWrapper to Llama2ChatWrapper

Features


Shipped with llama.cpp release b2717

To use the latest llama.cpp release available, run npx --no node-llama-cpp download --release latest. (learn more)

node-llama-cpp - v3.0.0-beta.16

Published by github-actions[bot] 6 months ago

3.0.0-beta.16 (2024-04-13)

Bug Fixes

Features


Shipped with llama.cpp release b2665

To use the latest llama.cpp release available, run npx --no node-llama-cpp download --release latest. (learn more)

node-llama-cpp - v3.0.0-beta.15

Published by github-actions[bot] 7 months ago

3.0.0-beta.15 (2024-04-04)

Bug Fixes

Features

  • automatically adapt to current free VRAM state (#182) (35e6f50)
  • inspect gguf command (#182) (35e6f50)
  • inspect measure command (#182) (35e6f50)
  • readGgufFileInfo function (#182) (35e6f50)
  • GGUF file metadata info on LlamaModel (#182) (35e6f50)
  • JinjaTemplateChatWrapper (#182) (35e6f50)
  • use the tokenizer.chat_template header from the gguf file when available - use it to find a better specialized chat wrapper or use JinjaTemplateChatWrapper with it as a fallback (#182) (35e6f50)
  • simplify generation CLI commands: chat, complete, infill (#182) (35e6f50)
  • Windows on Arm prebuilt binary (#181) (f3b7f81)

Shipped with llama.cpp release b2608

To use the latest llama.cpp release available, run npx --no node-llama-cpp download --release latest. (learn more)

node-llama-cpp - v2.8.9

Published by github-actions[bot] 7 months ago

2.8.9 (2024-03-21)

Bug Fixes

node-llama-cpp - v3.0.0-beta.14

Published by github-actions[bot] 7 months ago

3.0.0-beta.14 (2024-03-16)

Bug Fixes

  • DisposedError was thrown when calling .dispose() (#178) (315a3eb)
  • adapt to breaking llama.cpp changes (#178) (315a3eb)

Features

  • async model and context loading (#178) (315a3eb)
  • automatically try to resolve Failed to detect a default CUDA architecture CUDA compilation error (#178) (315a3eb)
  • detect cmake binary issues and suggest fixes on detection (#178) (315a3eb)

Shipped with llama.cpp release b2440

To use the latest llama.cpp release available, run npx --no node-llama-cpp download --release latest. (learn more)

node-llama-cpp - v3.0.0-beta.13

Published by github-actions[bot] 8 months ago

3.0.0-beta.13 (2024-03-03)

Bug Fixes

Features


Shipped with llama.cpp release b2329

To use the latest llama.cpp release available, run npx --no node-llama-cpp download --release latest. (learn more)

node-llama-cpp - v3.0.0-beta.12

Published by github-actions[bot] 8 months ago

3.0.0-beta.12 (2024-02-24)

Bug Fixes

Features


Shipped with llama.cpp release b2254

To use the latest llama.cpp release available, run npx --no node-llama-cpp download --release latest. (learn more)

node-llama-cpp - v2.8.8

Published by github-actions[bot] 8 months ago

2.8.8 (2024-02-23)

Bug Fixes

node-llama-cpp - v2.8.7

Published by github-actions[bot] 8 months ago

2.8.7 (2024-02-18)

Bug Fixes

node-llama-cpp - v3.0.0-beta.11

Published by github-actions[bot] 8 months ago

3.0.0-beta.11 (2024-02-18)

Features

  • completion and infill (#164) (ede69c1)
  • support configuring more options for getLlama when using "lastBuild" (#164) (ede69c1)
  • export resolveChatWrapperBasedOnWrapperTypeName (#165) (624fa30)

Shipped with llama.cpp release b2174

To use the latest llama.cpp release available, run npx --no node-llama-cpp download --release latest. (learn more)

node-llama-cpp - v3.0.0-beta.10

Published by github-actions[bot] 8 months ago

3.0.0-beta.10 (2024-02-11)

Features


Shipped with llama.cpp release b2127

To use the latest llama.cpp release available, run npx --no node-llama-cpp download --release latest. (learn more)

node-llama-cpp - v2.8.6

Published by github-actions[bot] 9 months ago

2.8.6 (2024-02-06)

node-llama-cpp - v3.0.0-beta.9

Published by github-actions[bot] 9 months ago

3.0.0-beta.9 (2024-02-05)

Bug Fixes

  • don't block a node process from exiting (#157) (74fb35c)
  • fix: respect logLevel and logger params when using "lastBuild" (#157) (74fb35c)
  • fix: print logs on the same event loop cycle (#157) (74fb35c)

Shipped with llama.cpp release b2074

To use the latest llama.cpp release available, run npx --no node-llama-cpp download --release latest. (learn more)

node-llama-cpp - v3.0.0-beta.8

Published by github-actions[bot] 9 months ago

3.0.0-beta.8 (2024-02-05)

Bug Fixes


Shipped with llama.cpp release b2060

To use the latest llama.cpp release available, run npx --no node-llama-cpp download --release latest. (learn more)

node-llama-cpp - v3.0.0-beta.7

Published by github-actions[bot] 9 months ago

3.0.0-beta.7 (2024-02-05)

Bug Fixes


Shipped with llama.cpp release b2060

To use the latest llama.cpp release available, run npx --no node-llama-cpp download --release latest. (learn more)

node-llama-cpp - v3.0.0-beta.6

Published by github-actions[bot] 9 months ago

3.0.0-beta.6 (2024-02-04)

Bug Fixes

Features


Shipped with llama.cpp release b2060

To use the latest llama.cpp release available, run npx --no node-llama-cpp download --release latest. (learn more)

node-llama-cpp - v3.0.0-beta.5

Published by github-actions[bot] 9 months ago

3.0.0-beta.5 (2024-01-24)

Bug Fixes

  • build Metal by default for Apple silicone devices (#150) (187627e)

Features

  • minor: add more flags to the chat command (#149) (cab617a)

Shipped with llama.cpp release b1961

To use the latest llama.cpp release available, run npx --no node-llama-cpp download --release latest. (learn more)