Skip to content

Commit 406571f

Browse files
committed
update meta data
Signed-off-by: will-hwang <[email protected]>
1 parent 8051b55 commit 406571f

File tree

1 file changed

+5
-4
lines changed

1 file changed

+5
-4
lines changed

_posts/2025-05-15-optimized-inference-processors.md

Lines changed: 5 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -3,14 +3,15 @@ layout: post
33
title: "Optimizing inference processors for cost efficiency and performance"
44
authors:
55
- will-hwang
6-
- heemin
6+
- heemin-kim
77
- kolchfa
8-
date: 2025-05-15
8+
date: 2025-05-29
99
has_science_table: true
1010
categories:
1111
- technical-posts
12-
meta_keywords: OpenSearch, inference processors, vector embeddings, text embedding, sparse encoding, image embedding, ingest pipeline optimization, skip_existing, performance tuning, semantic search, multimodal search, machine learning inference, cost reduction, bulk API, document updates
13-
meta_description: Learn how to optimize inference processors in OpenSearch to reduce redundant model calls, lower costs, and improve ingestion performance.
12+
meta_keywords: inference processors, vector embeddings, OpenSearch text embedding, text image embedding, sparse encoding, caching mechanism, ingest pipeline, OpenSearch optimization
13+
meta_description: Learn about a new OpenSearch optimization for inference processors that reduces redundant calls, lowering costs and improving performance in vector embedding generation.
14+
1415
---
1516

1617
Inference processors, such as `text_embedding`, `text_image_embedding`, and `sparse_encoding`, enable the generation of vector embeddings during document ingestion or updates. Today, these processors invoke model inference every time a document is ingested or updated, even if the embedding source fields remain unchanged. This can lead to unnecessary compute usage and increased costs.

0 commit comments

Comments
 (0)