-
Notifications
You must be signed in to change notification settings - Fork 5k
Description
Could you help clarify why this warning is necessary? It clutters the console, and the purpose isn't entirely clear.
Additionally, I've noticed that when using the Vision API, all pages from documents (like PDFs) are stored as PNGs, even if there isn’t a single image on the page. Is there a reason for this? Couldn’t we apply the Vision API selectively, using it only for pages containing images? This would avoid the extra processing effort and the token usage involved in storing simple text pages as images. Converting text to images for Vision seems to double the runtime, increase blob storage, and create unnecessary index chunks for text-only pages.
Unless I’ve misunderstood, it would make sense to use Vision solely on image-containing pages, achieving the best of both approaches without doubling token consumption. Could you provide some insight into this approach? I’m still exploring the code and would appreciate a better understanding.
There is also some dead code, where i am not sure if its leftover or not finished feature.
"has_image_embeddings" is not used at
self.has_image_embeddings = has_image_embeddings |
so its also not needed at
search_images: bool = False, |
I wonder if its ok to enable vision in the app but selectively run prepdocs only for sources with heavy image content is ok. Will the app be able to deal with mixed embedding? If prepdocs would decide on the fly per page if its image heavy or not this could improve speed and costs and only use vision where useful.