BREAKING CHANGES
- Renamed `Kani.get_truncated_chat_history` to `Kani.get_prompt`
Additions & Improvements
- Added `CTransformersEngine` and `LlamaCTransformersEngine` (thanks Maknee!)
- Added a lower-level `Kani.get_model_completion` to make a prediction at the current chat state (without modifying the chat history)
- Added the `auto_truncate` param to `ai_function` to opt in to kani trimming long responses from a function (i.e., responses that do not fit in a model's context)
- Improved the internal handling of tokens when the chat history is directly modified
- `ChatMessage.[role]()` classmethods now pass kwargs to the constructor
- LLaMA: Improved the fidelity of non-strict-mode LLaMA prompting
- OpenAI: Added support for specifying an OpenAI organization and configuring retry
- Many documentation improvements
Fixes
- OpenAI message length could return too short on messages with no content
- Other minor fixes and improvements