Skip to content

Release v0.4.0

Latest

Choose a tag to compare

@jjk-g jjk-g released this 06 Feb 22:48
e3e690b

Release v0.4.0

This release contains several feature improvements and various bug fixes:

  • mTLS support in vllm client
  • Multilora support
  • E2e testing against inference sim
  • Multiple report analysis
  • New aliases for shared_prefix config fields
  • Dependency updates

What's Changed

New Contributors

Full Changelog: v0.3.0...v0.4.0

Docker Image

quay.io/inference-perf/inference-perf:v0.4.0

Python Package

pip install inference-perf==v0.4.0