Releases: michaelfeil/infinity
Releases · michaelfeil/infinity
0.0.13
What's Changed
- update-ci-and-readme by @michaelfeil in #48
- fix-fp16 by @michaelfeil in #49
Full Changelog: 0.0.12...0.0.13
0.0.12
What's Changed
- add classification pipeline by @michaelfeil in #44 -> See update in Readme.md
- make fp16 default for CUDA by @michaelfeil in #47
- add bettertransformer by @michaelfeil in #47
Breaking:
- make fp16 default for CUDA by @michaelfeil in #47 -> This may lower your precision for a 2x speedup.
Full Changelog: 0.0.11...0.0.12
0.0.11
What's Changed
- Refactoring of API. by @michaelfeil in #43
- initial support for ReRanker models by @michaelfeil in #43
Full Changelog: 0.0.10...0.0.11
0.0.10
0.0.9
What's Changed
- Update README.md by @BrianPulfer in #37
- Update README.md by @BrianPulfer in #38
- add the device feature by @michaelfeil in #39
New Contributors
- @BrianPulfer made their first contribution in #37 also first contributor to this repo - thanks!
Full Changelog: 0.0.8...0.0.9
0.0.8
0.0.7
What's Changed
- Docker: Cuda11.8, make dependencies optional by @michaelfeil in #33
- Breaking changes: New install groups
pip install infinity-emb[server,logging,onnx-gpu-runtime]
- add onnx-gpu by @michaelfeil in #26
Full Changelog: 0.0.6...0.0.7
0.0.6
0.0.5
What's Changed
- Docker image multi by @michaelfeil in #24
- patch missing event -> 200ms to 7ms inference at Batch size 1
Full Changelog: 0.0.4...0.0.5
0.0.4
What's Changed
PRs:
- Fastembed v2 by @michaelfeil in #21 :
Issues:
Closes #5 ONNX Support via https://github.com/qdrant/fastembed/
Closes #22 making pytorch and optional dependency
tl,dr
fastembed
as backend besides ct2 or torch- v1/models returns "backend"
- makes torch an optional dependency
- calculates "min" sleep time dynamically on startup _> slightly optimized.
- default model is now "BAAI/bge-small-en-v1.5"
Full Changelog: 0.0.3...0.0.4