Skip to content

Commit 8eaf6b0

Browse files
committed
small edits to semantic docs
1 parent 09c238a commit 8eaf6b0

File tree

4 files changed

+9
-9
lines changed

4 files changed

+9
-9
lines changed

articles/search/includes/quickstarts/semantic-ranker-csharp.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -357,7 +357,7 @@ Press Enter to continue to the next query...
357357

358358
### Output for a semantic query with captions
359359

360-
This query adds captions with hit highlighting.
360+
Here are the results for the query that adds captions with hit highlighting.
361361

362362
```
363363
Caption: Chic hotel near the city. High-rise hotel in downtown, within walking distance to<em> theaters, </em>art galleries, restaurants and shops. Visit<em> Seattle Art Museum </em>by day, and then head over to<em> Benaroya Hall </em>to catch the evening's concert performance.
@@ -400,13 +400,13 @@ Press Enter to continue to the next query...
400400

401401
### Output for semantic answers
402402

403-
In this final query, return semantic answers. Notice that we changed the query string for this example: 'what's a good hotel for people who like to read'.
403+
The final query returns a semantic answer. Notice that we changed the query string for this example: 'what's a good hotel for people who like to read'.
404404

405405
Semantic ranker can produce an answer to a query string that has the characteristics of a question. The generated answer is extracted verbatim from your content so it won't include composed content like what you might expect from a chat completion model. If the semantic answer isn't useful for your scenario, you can omit `semantic_answers` from your code.
406406

407-
To get a semantic answer, the question and answer must be closely aligned, and the model must find content that clearly answers the question. If potential answers fail to meet a confidence threshold, the model doesn't return an answer. For demonstration purposes, the question in this example is designed to get a response so that you can see the syntax.
407+
To produce a semantic answer, the question and answer must be closely aligned, and the model must find content that clearly answers the question. If potential answers fail to meet a confidence threshold, the model doesn't return an answer. For demonstration purposes, the question in this example is designed to get a response so that you can see the syntax.
408408

409-
Recall that answers are verbatim content pulled from your index and might be missing phrases that a user would expect to see. To get composed answers as generated by a chat completion model, considering using a [RAG pattern](../../retrieval-augmented-generation-overview.md) or [agentic retrieval](../../search-agentic-retrieval-concept.md).
409+
Recall that answers are *verbatim content* pulled from your index and might be missing phrases that a user would expect to see. To get *composed answers* as generated by a chat completion model, considering using a [RAG pattern](../../retrieval-augmented-generation-overview.md) or [agentic retrieval](../../search-agentic-retrieval-concept.md).
410410

411411
```bash
412412
Extractive Answers:

articles/search/includes/quickstarts/semantic-ranker-intro.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -69,7 +69,7 @@ This quickstart assumes an existing index, modified to include a semantic config
6969

7070
Your response should be similar to the following example, as scored by the default BM25 L1 ranker for full text search. For readability, the example selects just the "HotelName" "HotelId", and "Description" fields.
7171

72-
This query is a keyword search. The matches contain verbatim matches on the query terms (walking, distance, live, music) or a linguistic variant of one of the terms (walk, living).
72+
This query is a keyword search. The results contain verbatim matches on the query terms (walking, distance, live, music) or on a linguistic variant (walk, living).
7373

7474
```json
7575
"@odata.count": 13,

articles/search/includes/quickstarts/semantic-ranker-python.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -261,7 +261,7 @@ In this final query, return semantic answers.
261261

262262
Semantic ranker can produce an answer to a query string that has the characteristics of a question. The generated answer is extracted verbatim from your content so it won't include composed content like what you might expect from a chat completion model. If the semantic answer isn't useful for your scenario, you can omit `semantic_answers` from your code.
263263

264-
To get a semantic answer, the question and answer must be closely aligned, and the model must find content that clearly answers the question. If potential answers fail to meet a confidence threshold, the model doesn't return an answer. For demonstration purposes, the question in this example is designed to get a response so that you can see the syntax.
264+
To produce a semantic answer, the question and answer must be closely aligned, and the model must find content that clearly answers the question. If potential answers fail to meet a confidence threshold, the model doesn't return an answer. For demonstration purposes, the question in this example is designed to get a response so that you can see the syntax.
265265

266266
1. Add `answers` to the query.
267267

@@ -297,7 +297,7 @@ To get a semantic answer, the question and answer must be closely aligned, and t
297297

298298
1. Output should look similar to the following example, where the best answer to question is pulled from one of the results.
299299

300-
Recall that answers are verbatim content pulled from your index and might be missing phrases that a user would expect to see. To get composed answers as generated by a chat completion model, considering using a [RAG pattern](../../retrieval-augmented-generation-overview.md) or [agentic retrieval](../../search-agentic-retrieval-concept.md).
300+
Recall that answers are *verbatim content* pulled from your index and might be missing phrases that a user would expect to see. To get *composed answers* as generated by a chat completion model, considering using a [RAG pattern](../../retrieval-augmented-generation-overview.md) or [agentic retrieval](../../search-agentic-retrieval-concept.md).
301301

302302
```bash
303303
Semantic Answer: Nature is Home on the beach. Explore the shore by day, and then come home to our shared living space to relax around a stone fireplace, sip something warm, and explore the<em> library </em>by night. Save up to 30 percent. Valid Now through the end of the year. Restrictions and blackouts may apply.

articles/search/includes/quickstarts/semantic-ranker-rest.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -285,7 +285,7 @@ In this final query, return semantic answers.
285285

286286
Semantic ranker can produce an answer to a query string that has the characteristics of a question. The generated answer is extracted verbatim from your content so it won't include composed content like what you might expect from a chat completion model. If the semantic answer isn't useful for your scenario, you can omit `semantic_answers` from your code.
287287

288-
To get a semantic answer, the question and answer must be closely aligned, and the model must find content that clearly answers the question. If potential answers fail to meet a confidence threshold, the model doesn't return an answer. For demonstration purposes, the question in this example is designed to get a response so that you can see the syntax.
288+
To produce a semantic answer, the question and answer must be closely aligned, and the model must find content that clearly answers the question. If potential answers fail to meet a confidence threshold, the model doesn't return an answer. For demonstration purposes, the question in this example is designed to get a response so that you can see the syntax.
289289

290290
1. Formulate the request using a search string that asks a question.
291291

@@ -306,7 +306,7 @@ To get a semantic answer, the question and answer must be closely aligned, and t
306306
307307
1. Output consists of 41 results for the new query, with "@search.answers" for the question in the query about hotels for people who like to read.
308308
309-
Recall that answers are verbatim content pulled from your index and might be missing phrases that a user would expect to see. To get composed answers as generated by a chat completion model, considering using a [RAG pattern](../../retrieval-augmented-generation-overview.md) or [agentic retrieval](../../search-agentic-retrieval-concept.md).
309+
Recall that answers are *verbatim content* pulled from your index and might be missing phrases that a user would expect to see. To get *composed answers* as generated by a chat completion model, considering using a [RAG pattern](../../retrieval-augmented-generation-overview.md) or [agentic retrieval](../../search-agentic-retrieval-concept.md).
310310
311311
In this example, the answer is considered as a strong fit for the question.
312312

0 commit comments

Comments
 (0)