Skip to content

Commit fc11815

Browse files
committed
After merge conflict resolution clean-up
1 parent 3164c6c commit fc11815

File tree

1 file changed

+21
-27
lines changed

1 file changed

+21
-27
lines changed

x-pack/plugin/inference/src/main/java/org/elasticsearch/xpack/inference/services/elastic/ElasticInferenceService.java

Lines changed: 21 additions & 27 deletions
Original file line numberDiff line numberDiff line change
@@ -353,21 +353,21 @@ protected void doChunkedInfer(
353353
TimeValue timeout,
354354
ActionListener<List<ChunkedInference>> listener
355355
) {
356-
if(model instanceof ElasticInferenceServiceDenseTextEmbeddingsModel denseTextEmbeddingsModel){
357-
var actionCreator = new ElasticInferenceServiceActionCreator(getSender(), getServiceComponents(), getCurrentTraceInfo());
358-
359-
List<EmbeddingRequestChunker.BatchRequestAndListener> batchedRequests = new EmbeddingRequestChunker<>(
360-
inputs.getInputs(),
361-
DENSE_TEXT_EMBEDDINGS_MAX_BATCH_SIZE,
362-
denseTextEmbeddingsModel.getConfigurations().getChunkingSettings()
363-
).batchRequestsWithListeners(listener);
364-
365-
for (var request : batchedRequests) {
366-
var action = denseTextEmbeddingsModel.accept(actionCreator, taskSettings);
367-
action.execute(EmbeddingsInput.fromStrings(request.batch().inputs().get(), inputType), timeout, request.listener());
368-
}
369-
370-
return;
356+
if (model instanceof ElasticInferenceServiceDenseTextEmbeddingsModel denseTextEmbeddingsModel) {
357+
var actionCreator = new ElasticInferenceServiceActionCreator(getSender(), getServiceComponents(), getCurrentTraceInfo());
358+
359+
List<EmbeddingRequestChunker.BatchRequestAndListener> batchedRequests = new EmbeddingRequestChunker<>(
360+
inputs.getInputs(),
361+
DENSE_TEXT_EMBEDDINGS_MAX_BATCH_SIZE,
362+
denseTextEmbeddingsModel.getConfigurations().getChunkingSettings()
363+
).batchRequestsWithListeners(listener);
364+
365+
for (var request : batchedRequests) {
366+
var action = denseTextEmbeddingsModel.accept(actionCreator, taskSettings);
367+
action.execute(EmbeddingsInput.fromStrings(request.batch().inputs().get(), inputType), timeout, request.listener());
368+
}
369+
370+
return;
371371
}
372372

373373
if (model instanceof ElasticInferenceServiceSparseEmbeddingsModel sparseTextEmbeddingsModel) {
@@ -423,8 +423,7 @@ public void parseRequestConfig(
423423
serviceSettingsMap,
424424
elasticInferenceServiceComponents,
425425
TaskType.unsupportedTaskTypeErrorMsg(taskType, NAME),
426-
ConfigurationParseContext.REQUEST,
427-
chunkingSettings
426+
ConfigurationParseContext.REQUEST
428427
);
429428

430429
throwIfNotEmptyMap(config, NAME);
@@ -461,8 +460,7 @@ private static ElasticInferenceServiceModel createModel(
461460
@Nullable Map<String, Object> secretSettings,
462461
ElasticInferenceServiceComponents elasticInferenceServiceComponents,
463462
String failureMessage,
464-
ConfigurationParseContext context,
465-
ChunkingSettings chunkingSettings
463+
ConfigurationParseContext context
466464
) {
467465
return switch (taskType) {
468466
case SPARSE_EMBEDDING -> new ElasticInferenceServiceSparseEmbeddingsModel(
@@ -534,8 +532,7 @@ public Model parsePersistedConfigWithSecrets(
534532
taskSettingsMap,
535533
chunkingSettings,
536534
secretSettingsMap,
537-
parsePersistedConfigErrorMsg(inferenceEntityId, NAME),
538-
chunkingSettings
535+
parsePersistedConfigErrorMsg(inferenceEntityId, NAME)
539536
);
540537
}
541538

@@ -556,8 +553,7 @@ public Model parsePersistedConfig(String inferenceEntityId, TaskType taskType, M
556553
taskSettingsMap,
557554
chunkingSettings,
558555
null,
559-
parsePersistedConfigErrorMsg(inferenceEntityId, NAME),
560-
chunkingSettings
556+
parsePersistedConfigErrorMsg(inferenceEntityId, NAME)
561557
);
562558
}
563559

@@ -573,8 +569,7 @@ private ElasticInferenceServiceModel createModelFromPersistent(
573569
Map<String, Object> taskSettings,
574570
ChunkingSettings chunkingSettings,
575571
@Nullable Map<String, Object> secretSettings,
576-
String failureMessage,
577-
ChunkingSettings chunkingSettings
572+
String failureMessage
578573
) {
579574
return createModel(
580575
inferenceEntityId,
@@ -585,8 +580,7 @@ private ElasticInferenceServiceModel createModelFromPersistent(
585580
secretSettings,
586581
elasticInferenceServiceComponents,
587582
failureMessage,
588-
ConfigurationParseContext.PERSISTENT,
589-
chunkingSettings
583+
ConfigurationParseContext.PERSISTENT
590584
);
591585
}
592586

0 commit comments

Comments
 (0)