Skip to content

Commit fb66cab

Browse files
committed
Fix API URLs
1 parent f3daa03 commit fb66cab

File tree

8 files changed

+14
-32
lines changed

8 files changed

+14
-32
lines changed

deploy-manage/autoscaling/trained-model-autoscaling.md

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -46,8 +46,7 @@ If you set the minimum number of allocations to 1, you will be charged even if t
4646

4747
You can enable adaptive allocations by using:
4848

49-
* the create inference endpoint API for ELSER, E5 and models uploaded through Eland that are used as inference services.
50-
%TBD URL for APIs
49+
* the create inference endpoint API for [ELSER](https://www.elastic.co/docs/api/doc/elasticsearch/operation/operation-inference-put-elser), [E5 and models uploaded through Eland](https://www.elastic.co/docs/api/doc/elasticsearch/operation/operation-inference-put-elasticsearch) that are used as inference services.
5150
* the [start trained model deployment](https://www.elastic.co/docs/api/doc/elasticsearch/operation/operation-ml-start-trained-model-deployment) or [update trained model deployment](https://www.elastic.co/docs/api/doc/elasticsearch/operation/operation-ml-update-trained-model-deployment) APIs for trained models that are deployed on {{ml}} nodes.
5251

5352
If the new allocations fit on the current {{ml}} nodes, they are immediately started. If more resource capacity is needed for creating new model allocations, then your {{ml}} node will be scaled up if {{ml}} autoscaling is enabled to provide enough resources for the new allocation. The number of model allocations can be scaled down to 0. They cannot be scaled up to more than 32 allocations, unless you explicitly set the maximum number of allocations to more. Adaptive allocations must be set up independently for each deployment and [{{infer}} endpoint](https://www.elastic.co/docs/api/doc/elasticsearch/group/endpoint-inference).

explore-analyze/machine-learning/nlp/ml-nlp-e5.md

Lines changed: 2 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -13,8 +13,7 @@ EmbEddings from bidirEctional Encoder rEpresentations - or E5 - is a {{nlp}} mo
1313

1414
[Semantic search](../../../solutions/search/semantic-search.md) provides you search results based on contextual meaning and user intent, rather than exact keyword matches.
1515

16-
E5 has two versions: one cross-platform version which runs on any hardware and one version which is optimized for Intel® silicon. The **Model Management** > **Trained Models** page shows you which version of E5 is recommended to deploy based on your cluster’s hardware. However, the recommended way to use E5 is through the {{infer}} API as a service which makes it easier to download and deploy the model and you don’t need to select from different versions.
17-
% TBD URL for API
16+
E5 has two versions: one cross-platform version which runs on any hardware and one version which is optimized for Intel® silicon. The **Model Management** > **Trained Models** page shows you which version of E5 is recommended to deploy based on your cluster’s hardware. However, the recommended way to use E5 is through the [{{infer}} API](https://www.elastic.co/docs/api/doc/elasticsearch/operation/operation-inference-put-elasticsearch) as a service which makes it easier to download and deploy the model and you don’t need to select from different versions.
1817

1918
Refer to the model cards of the [multilingual-e5-small](https://huggingface.co/elastic/multilingual-e5-small) and the [multilingual-e5-small-optimized](https://huggingface.co/elastic/multilingual-e5-small-optimized) models on HuggingFace for further information including licensing.
2019

@@ -45,8 +44,7 @@ PUT _inference/text_embedding/my-e5-model
4544

4645
The API request automatically initiates the model download and then deploy the model.
4746

48-
Refer to the `elasticsearch` {{infer}} service documentation to learn more about the available settings.
49-
% TBD URL for API
47+
Refer to the `elasticsearch` [{{infer}} service documentation](https://www.elastic.co/docs/api/doc/elasticsearch/operation/operation-inference-put-elasticsearch) to learn more about the available settings.
5048

5149
After you created the E5 {{infer}} endpoint, it’s ready to be used for semantic search. The easiest way to perform semantic search in the {{stack}} is to [follow the `semantic_text` workflow](../../../solutions/search/semantic-search/semantic-search-semantic-text.md).
5250

explore-analyze/machine-learning/nlp/ml-nlp-elser.md

Lines changed: 2 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -39,8 +39,7 @@ Enabling trained model autoscaling for your ELSER deployment is recommended. Ref
3939

4040
Compared to the initial version of the model, ELSER v2 offers improved retrieval accuracy and more efficient indexing. This enhancement is attributed to the extension of the training data set, which includes high-quality question and answer pairs and the improved FLOPS regularizer which reduces the cost of computing the similarity between a query and a document.
4141

42-
ELSER v2 has two versions: one cross-platform version which runs on any hardware and one version which is optimized for Intel® silicon. The **Model Management** > **Trained Models** page shows you which version of ELSER v2 is recommended to deploy based on your cluster’s hardware. However, the recommended way to use ELSER is through the {{infer}} API as a service which makes it easier to download and deploy the model and you don't need to select from different versions.
43-
% TBD URL for API
42+
ELSER v2 has two versions: one cross-platform version which runs on any hardware and one version which is optimized for Intel® silicon. The **Model Management** > **Trained Models** page shows you which version of ELSER v2 is recommended to deploy based on your cluster’s hardware. However, the recommended way to use ELSER is through the [{{infer}} API](https://www.elastic.co/docs/api/doc/elasticsearch/operation/operation-inference-put-elasticsearch) as a service which makes it easier to download and deploy the model and you don't need to select from different versions.
4443

4544
If you want to learn more about the ELSER V2 improvements, refer to [this blog post](https://www.elastic.co/search-labs/blog/introducing-elser-v2-part-1).
4645

@@ -75,8 +74,7 @@ PUT _inference/sparse_embedding/my-elser-endpoint
7574

7675
The API request automatically initiates the model download and then deploy the model. This example uses [autoscaling](../../../deploy-manage/autoscaling/trained-model-autoscaling.md) through adaptive allocation.
7776

78-
Refer to the ELSER {{infer}} integration documentation to learn more about the available settings.
79-
% TBD URL for API
77+
Refer to the [ELSER {{infer}} integration documentation](https://www.elastic.co/docs/api/doc/elasticsearch/operation/operation-inference-put-elser) to learn more about the available settings.
8078

8179
After you created the ELSER {{infer}} endpoint, it’s ready to be used for semantic search. The easiest way to perform semantic search in the {{stack}} is to [follow the `semantic_text` workflow](../../../solutions/search/semantic-search/semantic-search-semantic-text.md).
8280

@@ -309,7 +307,6 @@ To gain the biggest value out of ELSER trained models, consider to follow this l
309307

310308
::::{important}
311309
The recommended way to use ELSER is through the {{infer}} API as a service.
312-
% TBD URL for API
313310
::::
314311

315312
The following sections provide information about how ELSER performs on different hardwares and compares the model performance to {{es}} BM25 and other strong baselines.

explore-analyze/machine-learning/nlp/ml-nlp-rerank.md

Lines changed: 2 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -44,8 +44,7 @@ Elastic Rerank is available in Elastic Stack version 8.17+:
4444

4545
## Download and deploy [ml-nlp-rerank-deploy]
4646

47-
To download and deploy Elastic Rerank, use the create inference API to create an {{es}} service `rerank` endpoint.
48-
% TBD URL for API
47+
To download and deploy Elastic Rerank, use the [create inference API](https://www.elastic.co/docs/api/doc/elasticsearch/operation/operation-inference-put-elasticsearch) to create an {{es}} service `rerank` endpoint.
4948

5049
::::{tip}
5150
Refer to this [Python notebook](https://github.com/elastic/elasticsearch-labs/blob/main/notebooks/search/12-semantic-reranking-elastic-rerank.ipynb) for an end-to-end example using Elastic Rerank.
@@ -281,7 +280,7 @@ For detailed benchmark information, including complete dataset results and metho
281280
**Documentation**:
282281

283282
* [Semantic re-ranking in {{es}} overview](../../../solutions/search/ranking/semantic-reranking.md#semantic-reranking-in-es)
284-
% TBD URL for API * [Inference API example](../../elastic-inference/inference-api/elasticsearch-inference-integration.md#inference-example-elastic-reranker)
283+
* [Inference API example](https://www.elastic.co/docs/api/doc/elasticsearch/operation/operation-inference-put-elasticsearch)
285284

286285
**Blogs**:
287286

solutions/search/hybrid-semantic-text.md

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -14,9 +14,7 @@ This tutorial demonstrates how to perform hybrid search, combining semantic sear
1414

1515
In hybrid search, semantic search retrieves results based on the meaning of the text, while full-text search focuses on exact word matches. By combining both methods, hybrid search delivers more relevant results, particularly in cases where relying on a single approach may not be sufficient.
1616

17-
The recommended way to use hybrid search in the {{stack}} is following the `semantic_text` workflow. This tutorial uses the `elasticsearch` service for demonstration, but you can use any service and their supported models offered by the {{infer-cap}} API.
18-
% TBD URL for API
19-
17+
The recommended way to use hybrid search in the {{stack}} is following the `semantic_text` workflow. This tutorial uses the [`elasticsearch` service](https://www.elastic.co/docs/api/doc/elasticsearch/operation/operation-inference-put-elasticsearch) for demonstration, but you can use any service and their supported models offered by the {{infer-cap}} API.
2018

2119
## Create an index mapping [hybrid-search-create-index-mapping]
2220

solutions/search/ranking/semantic-reranking.md

Lines changed: 4 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -92,14 +92,10 @@ To use semantic re-ranking in {{es}}, you need to:
9292

9393
1. **Select and configure a re-ranking model**. You have the following options:
9494

95-
1. Use the Elastic Rerank cross-encoder model via the inference API's {{es}} service.
96-
% TBD URL for API
97-
2. Use the Cohere Rerank inference endpoint to create a `rerank` endpoint.
98-
% TBD URL for API
99-
3. Use the Google Vertex AI inference endpoint to create a `rerank` endpoint.
100-
% TBD URL for API
101-
4. Upload a model to {{es}} from Hugging Face with [Eland](eland://reference/machine-learning.md#ml-nlp-pytorch). You’ll need to use the `text_similarity` NLP task type when loading the model using Eland. Then set up an {{es}} service inference endpoint with the `rerank` endpoint type.
102-
% TBD URL for API
95+
1. Use the Elastic Rerank cross-encoder model via the [inference API's {{es}} service](https://www.elastic.co/docs/api/doc/elasticsearch/operation/operation-inference-put-elasticsearch).
96+
2. Use the [Cohere Rerank inference endpoint](https://www.elastic.co/docs/api/doc/elasticsearch/operation/operation-inference-put-cohere) to create a `rerank` endpoint.
97+
3. Use the [Google Vertex AI inference endpoint](https://www.elastic.co/docs/api/doc/elasticsearch/operation/operation-inference-put-googlevertexai) to create a `rerank` endpoint.
98+
4. Upload a model to {{es}} from Hugging Face with [Eland](eland://reference/machine-learning.md#ml-nlp-pytorch). You’ll need to use the `text_similarity` NLP task type when loading the model using Eland. Then set up an [{{es}} service inference endpoint](https://www.elastic.co/docs/api/doc/elasticsearch/operation/operation-inference-put-elasticsearch) with the `rerank` endpoint type.
10399

104100
Refer to [the Elastic NLP model reference](../../../explore-analyze/machine-learning/nlp/ml-nlp-model-ref.md#ml-nlp-model-ref-text-similarity) for a list of third party text similarity models supported by {{es}} for semantic re-ranking.
105101

solutions/search/semantic-search/semantic-search-inference.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,7 @@ The following examples use the:
2929
* `ops-text-embedding-zh-001` model for [AlibabaCloud AI](https://help.aliyun.com/zh/open-search/search-platform/developer-reference/text-embedding-api-details)
3030

3131
You can use any Cohere and OpenAI models, they are all supported by the {{infer}} API.
32-
% TBD URL: For a list of recommended models available on HuggingFace, refer to [the supported model list](../../../explore-analyze/elastic-inference/inference-api/huggingface-inference-integration.md#inference-example-hugging-face-supported-models).
32+
For a list of recommended models available on HuggingFace, refer to the supported model list in the [API documentation](https://www.elastic.co/docs/api/doc/elasticsearch/operation/operation-inference-put-hugging-face).
3333

3434
Click the name of the service you want to use on any of the widgets below to review the corresponding instructions.
3535

solutions/search/semantic-search/semantic-search-semantic-text.md

Lines changed: 1 addition & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -15,15 +15,11 @@ Semantic text simplifies the {{infer}} workflow by providing {{infer}} at ingest
1515

1616
The recommended way to use [semantic search](../semantic-search.md) in the {{stack}} is following the `semantic_text` workflow. When you need more control over indexing and query settings, you can still use the complete {{infer}} workflow (refer to [this tutorial](../../../explore-analyze/elastic-inference/inference-api.md) to review the process).
1717

18-
This tutorial uses the `elasticsearch` service for demonstration, but you can use any service and their supported models offered by the {{infer-cap}} API.
19-
% TBD URL for API
20-
18+
This tutorial uses the [`elasticsearch` service](https://www.elastic.co/docs/api/doc/elasticsearch/operation/operation-inference-put-elasticsearch) for demonstration, but you can use any service and their supported models offered by the {{infer-cap}} API.
2119

2220
## Requirements [semantic-text-requirements]
2321

2422
This tutorial uses the `elasticsearch` service for demonstration, which is created automatically as needed. To use the `semantic_text` field type with an {{infer}} service other than `elasticsearch` service, you must create an inference endpoint using the [Create {{infer}} API](https://www.elastic.co/docs/api/doc/elasticsearch/operation/operation-inference-put).
25-
% TBD URL for API
26-
2723

2824
## Create the index mapping [semantic-text-index-mapping]
2925

@@ -44,7 +40,6 @@ PUT semantic-embeddings
4440

4541
1. The name of the field to contain the generated embeddings.
4642
2. The field to contain the embeddings is a `semantic_text` field. Since no `inference_id` is provided, the default endpoint `.elser-2-elasticsearch` for the `elasticsearch` service is used. To use a different {{infer}} service, you must create an {{infer}} endpoint first using the [Create {{infer}} API](https://www.elastic.co/docs/api/doc/elasticsearch/operation/operation-inference-put) and then specify it in the `semantic_text` field mapping using the `inference_id` parameter.
47-
% TBD URL for API
4843

4944
::::{note}
5045
If you’re using web crawlers or connectors to generate indices, you have to [update the index mappings](https://www.elastic.co/docs/api/doc/elasticsearch/operation/operation-indices-put-mapping) for these indices to include the `semantic_text` field. Once the mapping is updated, you’ll need to run a full web crawl or a full connector sync. This ensures that all existing documents are reprocessed and updated with the new semantic embeddings, enabling semantic search on the updated data.

0 commit comments

Comments
 (0)