Sourced from llama-cpp-python's changelog.
[0.2.32]
- feat: Update llama.cpp to ggerganov/llama.cpp@504dc37be8446fb09b1ede70300250ad41be32a2
- fix: from_json_schema oneof/anyof bug by
@jndiogo
in d3f5528ca8bcb9d69d4f27e21631e911f1fb9bfe- fix: pass chat handler not chat formatter for huggingface autotokenizer and tokenizer_config formats by
@abetlen
in 24f39454e91cf5dddbc4b6041aead4accc7c7a2d- feat: Add add_generation_prompt option for jinja2chatformatter by
@abetlen
in 7f3209b1eb4ad3260ba063801fab80a8c25a2f4c- feat: Add Jinja2ChatFormatter by
@abetlen
in be09318c26add8674ce494ae7cc480cce72a4146- feat: Expose gguf model metadata in metadata property by
@abetlen
in 5a34c57e5479e50c99aba9b38218cc48e6560b81[0.2.31]
- feat: Update llama.cpp to ggerganov/llama.cpp@a5cacb22b2114fd9adf61c00cbb237384d86bced
- fix: Mirostat sampling now passes correct type to ctypes and tracks state during generation by
@abetlen
in 3babe3512cb95743108f2b595210c38ed6f1b904- fix: Python3.8 support in server by
@abetlen
in 141293a75b564a8699e0acba1da24d9aa1cf0ab1[0.2.30]
- feat: Update llama.cpp to ggerganov/llama.cpp@57e2a7a52a819883f40dada8a2edc24ecf48186b
- feat(server): Add ability to load chat format from huggingface autotokenizer or tokenizer_config.json files by
@abetlen
in b8fc1c7d83ad4a9207c707ba1d954fe580286a01- feat: Integration of Jinja2 Templating for chat formats by
@teleprint-me
in #875- fix: Offload KQV by default by
@abetlen
in 48c3b77e6f558a9899de0e1155c7dc0c7958d8e8- fix: Support Accept text/event-stream in chat and completion endpoints, resolves #1083 by
@aniljava
in #1088- fix(cli): allow passing n_ctx=0 to openAI API server args to use model n_ctx_train field per #1015 by
@K-Mistele
in #1093[0.2.29]
- feat: Update llama.cpp to ggerganov/llama.cpp@4483396751c79dea540808b9cb9238245d06da2b
- feat: Add split_mode option by
@abetlen
in 84615adbc6855c8384807c42f0130f9a1763f99d- feat: Implement GGUF metadata KV overrides by
@phiharri
in #1011- fix: Avoid "LookupError: unknown encoding: ascii" when open() called in a destructor by
@yieldthought
in #1012- fix: Fix low_level_api_chat_cpp example to match current API by
@aniljava
in #1086- fix: Fix Pydantic model parsing by
@DeNeutoy
in #1087[0.2.28]
- feat: Update llama.cpp to ggerganov/llama.cpp@6efb8eb30e7025b168f3fda3ff83b9b386428ad6
- feat: Add ability to pass in penalize_nl param by
@shankinson
in #1068- fix: print_grammar to stderr by
@turian
in #1052
2ce0b8a
Bump versiond3f5528
fix: from_json_schema oneof/anyof bug. Closes #10978eefdbc
Update llama.cpp88fbcca
docs: Add macosx wrong arch fix to README24f3945
fix: pass chat handler not chat formatter for huggingface autotokenizer
and t...7f3209b
feat: Add add_generation_prompt option for jinja2chatformatter.ac2e96d
Update llama.cppbe09318
feat: Add Jinja2ChatFormatter5a34c57
feat: Expose gguf model metadata in metadata property833a7f1
Bump version