What does normalizing embeddings do? #5067
ShelbyJenkins
started this conversation in
General
Replies: 2 comments
-
Hi @ShelbyJenkins sorry to hear that the large amount of memory required by Haystack's dependencies is an issue. Maybe this discussion is relevant to you: #5032 Vector length normalization ensures that all the vectors have the same length so that when we compare vectors, we are only comparing their direction. |
Beta Was this translation helpful? Give feedback.
0 replies
-
@julian-risch thanks for the update. Good to know about that option. FWIW my little chat bot's image size without haystack ended up at 150mb. I'm new to deploying apps in containers, but that seems reasonable. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I really liked haystack, but I'm trying to migrate off of haystack for my chat bot project due to the large size (5g compressed) of containers.
Without haystack I can't retrieve embeddings from pinecone. Queries return no results just using the openai embedding endpoint and pinecone-client package. I think I tracked the issue to haystack performing operations on embedded vectors. I'm sure there is a way to do it without haystack, but the levels of abstraction are too much of a time sink for me to parse. So I'll just recreate my pinecone index from scratch.
But I am curious what the value this sort of normalization brings!
from:
haystack/haystack/document_stores/base.py
Line 313 in a9a49e2
Beta Was this translation helpful? Give feedback.
All reactions