Releases: intentee/paddler
Releases · intentee/paddler
v3.0.1
v3.0.1-rc1
- Fix invalid max tokens count
- Add more unit tests
- Add integration tests
- Add code coverage validation in CI
v3.0.0
ci: build everything, create release from all the artifacts
v3.0.0-rc2
ci: build everything, create release from all the artifacts
v3.0.0-rc1
bump paddler to 3.0.0
Paddler API Client
Features
- Add
paddler_client- Rust library to connect to Paddler - Embedding, and inference performance improvements
- Update llama.cpp bindings to the latest version
Fixes
OpenAI compatibility improvements
OpenAI compatibility improvements
v2.1.0
License switch
The license was switched from MIT to Apache-2.0, which is more permissive (has explicit patent grants), so it should be easier to adopt in organizations.
Features
- OpenAI compatibility endpoint:
- Support for
max_completion_tokensparameter in /v1/chat/completions endpoint - Support for
messagesparameter in /v1/chat/completions endpoint - Support for
streamparameter in /v1/chat/completions endpoint
- Support for
Documentation: https://paddler.intentee.com/docs/migrating-to-paddler/openai-compatibility/
Full Changelog: v2.0.0...v2.1.0
v2.0.0
What's Changed
Long story short, we rewrote most of the llama-server, made it scalable, and bundled that with Paddler. This means you do not have to deploy llama-server alongside Paddler anymore (the fewer moving parts, the better). :)
We also have a new, vastly improved admin panel, and a lot of other fixes, improvements, and changes.
Full Changelog: v1.2.1-rc1...v2.0.0