Releases: withcatai/node-llama-cpp
v2.8.14
v2.8.13
v3.0.0-beta.38
3.0.0-beta.38 (2024-07-09)
Bug Fixes
- adapt to
llama.cpp
breaking changes (#266) (c35ff5a) - Llama 3 Instruct function calling (#266) (c35ff5a)
Features
Shipped with llama.cpp
release b3347
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)
v3.0.0-beta.37
3.0.0-beta.37 (2024-07-05)
Features
- flash attention (#264) (c2e322c)
- manually exclude specific GPU types from
gpu: "auto"
(#264) (c2e322c)
Shipped with llama.cpp
release b3322
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)
v3.0.0-beta.36
3.0.0-beta.36 (2024-06-30)
Bug Fixes
Shipped with llama.cpp
release b3267
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)
v3.0.0-beta.35
3.0.0-beta.35 (2024-06-30)
Bug Fixes
Shipped with llama.cpp
release b3266
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)
v3.0.0-beta.34
3.0.0-beta.34 (2024-06-30)
Bug Fixes
Shipped with llama.cpp
release b3265
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)
v3.0.0-beta.33
3.0.0-beta.33 (2024-06-29)
Bug Fixes
- long
LlamaText
tokenization (#249) (9cab784) - improve GGUF metadata reading performance (#249) (9cab784)
- adapt to breaking
llama.cpp
changes (#249) (9cab784)
Features
Shipped with llama.cpp
release b3265
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)
v2.8.12
v3.0.0-beta.32
3.0.0-beta.32 (2024-06-18)
Bug Fixes
Shipped with llama.cpp
release b3166
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)