3.2.0 (2024-10-31)
Bug Fixes
- Electron crash with some models on macOS when not using Metal (#375) (ea12dc5)
- adapt to
llama.cpp
breaking changes (#375) (ea12dc5) - support
rejectattr
in Jinja templates (#376) (ea12dc5) - build warning on macOS (#377) (6405ee9)
Features
- chat session response prefix (#375) (ea12dc5)
- improve context shift strategy (#375) (ea12dc5)
- use RAM and swap sizes in memory usage estimations (#375) (ea12dc5)
- faster building from source (#375) (ea12dc5)
- improve CPU compatibility score (#375) (ea12dc5)
inspect gguf
command: print a single key flag (#375) (ea12dc5)
Shipped with llama.cpp
release b3995
To use the latest
llama.cpp
release available, runnpx -n node-llama-cpp source download --release latest
. (learn more)