Skip to content

Releases: qdrant/fastembed

v0.7.4

05 Dec 12:18
428381c

Choose a tag to compare

Changelog

Features 🏎️

  • #577 - don't do any network calls if model is loaded from cache @joein
  • #578 - expose enable_cpu_mem_arena onnx session option to handle onnxruntime memory allocation by @joein
  • #582 - unlock hf hub and pillow versions by @joein
  • #583 - add a new token_count method to text models by @joein @dancixx

Fixes 🕸️

  • #585 - more granular numpy versioning

Thanks to everyone who contributed to this release @dancixx @joein and @tbung for all the reviews

v0.7.2

25 Aug 15:05

Choose a tag to compare

Changelog

Features 🏎️

Fixes 🕸️

  • #547 - avoid adding padding token embeddings to the result in batched inference in late interaction models by @generall

Thanks to everyone who contributed to this release @kacperlukawski @generall @joein

v0.7.1

16 Jun 09:06
acec312

Choose a tag to compare

Changelog

Features 🏎️

  • #532 - improved warnings for the recently changed models by @joein
  • #522 - raise exceptions in case of incorrect pooling in custom models by @joein
  • #521 - add property embedding_size and cls method get_embedding_size for easier access to the model dim by @joein

Fixes 🕸️

  • #524 - fix propagation of specific_model_path and local_files_only to embed with parallel > 1 by @joein

v0.7.0

13 May 14:32
c85e8c2

Choose a tag to compare

Changelog

Features 🏎️

  • #513 - New sparse embeddings model with semantic understanding: MiniCOIL (Qdrant/minicoil-v1) by @generall

Fixes 🕸️

  • #506 - fix list of supported languages in bm25 by @joein

v0.6.1

10 Apr 13:51

Choose a tag to compare

Changelog

Features 🏎️

  • #490 - deprecate old archive struct when loading from custom urls in favour of model_name.tar.gz to ease adding custom models by @joein
  • #492 - preserve embeddings in a type set by their model to allow lower precision outputs by @joein
  • #496 - custom rerankers support by @I8dNLo

Fixes 🕸️

  • #499 - fix splade hf source, causing model downloading errors in some cases by @joein

v0.6.0

26 Feb 13:55

Choose a tag to compare

Changelog

Features 🏎️

Fixes 🕸️

  • #486 - fix thenlper/gte-large, use mean pooling of token embeddings instead of CLS embedding by @joein
  • #436 - fix paraphrase-multilingual-MiniLM-L12-v2, do not normalize and use mean pooling of token embeddings instead of CLS embedding by @hh-space-invader
  • #445 - fix paraphrase-multilingual-mpnet-base-v2 and intfloat/multilingual-e5-large, do not normalize and use mean pooling of token embeddings instead of CLS embedding by @I8dNLo

Thanks everyone who contributed to the current release!

v0.5.1

20 Jan 10:43

Choose a tag to compare

Change Log

Fixes 🐛

  • #439 - fix python3.13 installation by making onnx an optional dependency, keep onnxruntime mandatory by @joein

v0.5.0

24 Dec 19:53

Choose a tag to compare

Features 📖

#403 - Drop Python 3.8 support by @joein
#404 - Add Python 3.13 support by @joein
#406 - Improve models cache progress bar by @hh-space-invader
#422 - Add multi-GPU example by @hh-space-invader
#425 - Provide user warning when specifying providers and CUDA by @hh-space-invader


Models 🧠

#405 - Support jina embeddings v2 models by @hh-space-invader
#408 - Add jina clip v1 model by @hh-space-invader
#415 - Added support for the thenlper/gte-base model by @hh-space-invader
#419 - Introduced parallel processing and pair-wise API for cross-encoders by @I8dNLo
#429 - All models now support Hugging Face (hf) compatibility by @I8dNLo


Fixes 🐛

#413 - Fix ColBERT model shape mismatch by @hh-space-invader

v0.4.2

13 Nov 13:41
9841666

Choose a tag to compare

Change Log

Features 📖

Models 🧠

Fixes 🐛

v0.4.1

21 Oct 20:30

Choose a tag to compare

Change Log

Features 📢

  • #366 - replace pystemmer with py-rust-stemmers by @I8dNLo