Skip to content

v0.5.0

Compare
Choose a tag to compare
@Tomas2D Tomas2D released this 13 Nov 21:46
· 142 commits to main since this release
c2a6cf6

🚀 Features / Enhancements

  • Added integration for LangChain Chat Models; see an example of generation and streaming.
  • Added support for LangChain Model Serialization (saving and loading models); see an example.
  • Added support for the Chat endpoint in Model class; see an example.
  • Added support for new moderation models (HAP, STIGMA, Implicit Hate) - not released on API yet but will be available soon.
  • Added type validation for input_tokens property in generate response.
  • Extend LangChain generation information / LLM Output (token_usage structure, generated tokens, stop_reason, conversation_id, created_at, ...).
  • Add optional raw_response=True/False parameter to generate_stream / generate_as_complete and generate methods to receive a raw response instead of unwrapped results.

🐛 Bug fixes

  • LangChain extension now correctly tokenizes the inputs (previously, the GPT2 tokenizer had been used).
  • Improve general error handling.

Full Changelog: v0.4.1...v0.5.0