Skip to content

Commit 184bbb4

Browse files
authored
Remove static, outdated lists of available models throughout the UI reference docs (#740)
1 parent 2864275 commit 184bbb4

File tree

7 files changed

+36
-174
lines changed

7 files changed

+36
-174
lines changed

ui/embedding.mdx

Lines changed: 4 additions & 36 deletions
Original file line numberDiff line numberDiff line change
@@ -59,41 +59,9 @@ on Hugging Face:
5959

6060
## Generate embeddings
6161

62-
To generate embeddings, choose one of the following embedding providers and models in the **Select Embedding Model** section of an **Embedder** node in a workflow:
62+
To generate embeddings, choose one of the available embedding providers and models in the **Select Embedding Model** section of an **Embedder** node in a workflow.
6363

64-
<Note>You can change a workflow's preconfigured provider only through [Custom](/ui/workflows#create-a-custom-workflow) workflow settings.</Note>
65-
66-
- **Azure OpenAI**: Use [Azure OpenAI](https://azure.microsoft.com/products/ai-services/openai-service) to generate embeddings with one of the following models:
67-
68-
- **text-embedding-3-small**, with 1536 dimensions.
69-
- **text-embedding-3-large**, with 3072 dimensions.
70-
- **Ada 002 (Text)** (`text-embedding-ada-002`), with 1536 dimensions.
71-
72-
[Learn more](https://learn.microsoft.com/azure/ai-services/openai/concepts/models#embeddings).
73-
74-
- **Amazon Bedrock**: Use [Amazon Bedrock](https://aws.amazon.com/bedrock) to generate embeddings with one of the following models:
75-
76-
- **Titan Text Embeddings V2**, with 1024 dimensions. [Learn more](https://docs.aws.amazon.com/bedrock/latest/userguide/titan-embedding-models.html).
77-
- **Titan Embeddings G1 - Text**, with 1536 dimensions. [Learn more](https://docs.aws.amazon.com/bedrock/latest/userguide/titan-embedding-models.html).
78-
- **Titan Multimodal Embeddings G1**, with 1024 dimensions. [Learn more](https://docs.aws.amazon.com/bedrock/latest/userguide/titan-multiemb-models.html).
79-
- **Cohere Embed English**, with 1024 dimensions. [Learn more](https://docs.aws.amazon.com/bedrock/latest/userguide/model-parameters-embed.html).
80-
- **Cohere Embed Multilingual**, with 1024 dimensions. [Learn more](https://docs.aws.amazon.com/bedrock/latest/userguide/model-parameters-embed.html).
64+
When choosing an embedding model, be sure to pay attention to the number of dimensions listed next to each model. This number must match the number of dimensions in the
65+
embeddings field of your destination connector's table, collection, or index.
8166

82-
- **TogetherAI**: Use [TogetherAI](https://www.together.ai/) to generate embeddings with one of the following models:
83-
84-
- **M2-BERT-80M-32K-Retrieval**, with 768 dimensions.
85-
86-
[Learn more](https://docs.together.ai/docs/serverless-models#embedding-models).
87-
88-
- **Voyage AI**: Use [Voyage AI](https://www.voyageai.com/) to generate embeddings with one of the following models:
89-
90-
- **voyage-code-2**, with 1536 dimensions.
91-
- **voyage-3**, with 1024 dimensions.
92-
- **voyage-3-large**, with 1024 dimensions.
93-
- **voyage-3-lite**, with 512 dimensions.
94-
- **voyage-code-3**, with 1024 dimensions.
95-
- **voyage-finance-2**, with 1024 dimensions.
96-
- **voyage-law-2**, with 1024 dimensions.
97-
- **voyage-multimodal-3**, with 1024 dimensions.
98-
99-
[Learn more](https://docs.voyageai.com/docs/embeddings).
67+
<Note>You can change a workflow's preconfigured provider only through [Custom](/ui/workflows#create-a-custom-workflow) workflow settings.</Note>

ui/enriching/image-descriptions.mdx

Lines changed: 2 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -4,11 +4,7 @@ title: Image descriptions
44

55
After partitioning, you can have Unstructured generate text-based summaries of detected images.
66

7-
This summarization is done by using models offered through these providers:
8-
9-
- [GPT-4o](https://openai.com/index/hello-gpt-4o/), provided through OpenAI.
10-
- [Claude 3.5 Sonnet](https://www.anthropic.com/news/claude-3-5-sonnet), provided through Anthropic.
11-
- [Claude 3.5 Sonnet](https://aws.amazon.com/bedrock/claude/), provided through Amazon Bedrock.
7+
This summarization is done by using models offered through various model providers.
128

139
Here is an example of the output of a detected image using GPT-4o. Note specifically the `text` field that is added.
1410
Line breaks have been inserted here for readability. The output will not contain these line breaks.
@@ -64,7 +60,7 @@ Any embeddings that are produced after these summaries are generated will be bas
6460

6561
import EnrichmentImageSummaryHiResOnly from '/snippets/general-shared-text/enrichment-image-summary-hi-res-only.mdx';
6662

67-
To generate image descriptions, in an **Enrichment** node in a workflow, specify the following:
63+
To generate image descriptions, in an **Enrichment** node in a workflow, select **Image**, and then choose one of the available provider (and model) combinations that are shown.
6864

6965
<Note>
7066
You can change a workflow's image description settings only through [Custom](/ui/workflows#create-a-custom-workflow) workflow settings.
@@ -74,9 +70,3 @@ To generate image descriptions, in an **Enrichment** node in a workflow, specify
7470
</Note>
7571

7672
<EnrichmentImageSummaryHiResOnly />
77-
78-
Select **Image**, and then choose one of the following provider (and model) combinations to use:
79-
80-
- **OpenAI (GPT-4o)**. [Learn more](https://openai.com/index/hello-gpt-4o/).
81-
- **Anthropic (Claude 3.5 Sonnet)**. [Learn more](https://www.anthropic.com/news/claude-3-5-sonnet).
82-
- **Amazon Bedrock (Claude 3.5 Sonnet)**. [Learn more](https://aws.amazon.com/bedrock/claude/).

ui/enriching/ner.mdx

Lines changed: 2 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -5,10 +5,7 @@ title: Named entity recognition (NER)
55
After partitioning, you can have Unstructured generate a list of recognized entities and their types (such as the names of organizations, products, and people) in the content, through a process known as _named entity recognition_ (NER).
66
You can also have Unstructured generate a list of relationships between the entities that are recognized.
77

8-
This NER is done by using models offered through these providers:
9-
10-
- [GPT-4o](https://openai.com/index/hello-gpt-4o/), provided through OpenAI.
11-
- [Claude 3.5 Sonnet](https://www.anthropic.com/news/claude-3-5-sonnet), provided through Anthropic.
8+
This NER is done by using models offered through various model providers.
129

1310
Here is an example of a list of recognized entities and their entity types, along with a list of relationships between those
1411
entities and their relationship types, using GPT-4o. Note specifically the `entities` field that is added to the `metadata` field.
@@ -147,7 +144,7 @@ To generate a list of recognized entities and their relationships, in an **Enric
147144
</Note>
148145

149146
1. Select **Text**.
150-
2. For **Model**, select either **OpenAI (GPT-4o)** or **Anthropic (Claude 3.5 Sonnet)**.
147+
2. For **Model**, select one of the available models that are shown.
151148
3. The selected model will follow a default set of instructions (called a _prompt_) to perform NER using a set of predefined entity types and relationships. To experiment
152149
with running the default prompt against some sample data, click **Edit**, and then click **Run Prompt**. The selected **Model** uses the
153150
**Prompt** to run NER on the **Input sample** and shows the results in the **Output**. Look specifically at the `response_json` field for the

ui/enriching/table-descriptions.mdx

Lines changed: 5 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -4,11 +4,7 @@ title: Table descriptions
44

55
After partitioning, you can have Unstructured generate text-based summaries of detected tables.
66

7-
This summarization is done by using models offered through these providers:
8-
9-
- [GPT-4o](https://openai.com/index/hello-gpt-4o/), provided through OpenAI.
10-
- [Claude 3.5 Sonnet](https://www.anthropic.com/news/claude-3-5-sonnet), provided through Anthropic.
11-
- [Claude 3.5 Sonnet](https://aws.amazon.com/bedrock/claude/), provided through Amazon Bedrock.
7+
This summarization is done by using models offered through various model providers.
128

139
Here is an example of the output of a detected table using GPT-4o. Note specifically the `text` field that is added.
1410
Line breaks have been inserted here for readability. The output will not contain these line breaks.
@@ -74,7 +70,10 @@ Any embeddings that are produced after these summaries are generated will be bas
7470

7571
import EnrichmentTableSummaryHiResOnly from '/snippets/general-shared-text/enrichment-table-summary-hi-res-only.mdx';
7672

77-
To generate table descriptions, in an **Enrichment** node in a workflow, specify the following:
73+
To generate table descriptions, in an **Enrichment** node in a workflow, select **Table**, and then choose one of the available provider (and model) combinations that are shown.
74+
75+
Make sure after you choose the provider and model, that **Table Description** is also displayed. If **Table Description** and **Table to HTML** are both
76+
displayed, be sure to select **Table Description**.
7877

7978
<Note>
8079
You can change a workflow's table description settings only through [Custom](/ui/workflows#create-a-custom-workflow) workflow settings.
@@ -85,14 +84,7 @@ To generate table descriptions, in an **Enrichment** node in a workflow, specify
8584

8685
<EnrichmentTableSummaryHiResOnly />
8786

88-
Select **Table**, and then choose one of the following provider (and model) combinations to use:
8987

90-
- **OpenAI (GPT-4o)**. [Learn more](https://openai.com/index/hello-gpt-4o/).
91-
- **Anthropic (Claude 3.5 Sonnet)**. [Learn more](https://www.anthropic.com/news/claude-3-5-sonnet).
92-
- **Amazon Bedrock (Claude 3.5 Sonnet)**. [Learn more](https://aws.amazon.com/bedrock/claude/).
93-
94-
Make sure after you choose the provider and model, that **Table Description** is also displayed. If **Table Description** and **Table to HTML** are both
95-
displayed, be sure to select **Table Description**.
9688

9789
## Learn more
9890

ui/partitioning.mdx

Lines changed: 10 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -98,26 +98,23 @@ The **VLM** strategy also has great support for recognizing multilanguage charac
9898
Language support for **VLM** depends on the model used. The list of supported languages for a particular model is maintained by
9999
that model's provider. For the list of languages that each model supports, see the following, where provided:
100100

101-
- Anthropic
101+
- Anthropic:
102102

103-
- Claude 3.5 Sonnet: Arabic, Bengali, Chinese (Simplified), English, French, German, Hindi, Indonesian, Italian, Japanese, Korean,
104-
Portuguese (Brazil), Spanish, Swahili, and Yoruba are mentioned. ([Source](https://docs.anthropic.com/en/docs/build-with-claude/multilingual-support))
103+
- [Claude](https://docs.anthropic.com/en/docs/build-with-claude/multilingual-support)
105104

106105
- OpenAI
107106

108-
- GPT-4o: Arabic, Chinese, English, French, German, Gujarati, Hindi, Italian, Japanese, Korean, Marathi, Persian,
109-
Portuguese, Russian, Spanish, Tamil, Telugu, Turkish, Urdu, and Vietnamese are mentioned. ([Source](https://openai.com/index/hello-gpt-4o/))
107+
- [GPT](https://help.openai.com/en/articles/8357869-how-to-change-your-language-setting-in-chatgpt#h_513834920e)
110108

111109
- Amazon Bedrock
112110

113-
- Claude 3.5 Sonnet: "English, Spanish, Japanese, and multiple other languages" ([Source](https://aws.amazon.com/bedrock/claude/))
114-
- Claude 3 Opus: "English, Spanish, Japanese, and multiple other languages" ([Source](https://aws.amazon.com/bedrock/claude/))
115-
- Claude 3 Haiku: "English, Spanish, Japanese, and multiple other languages" ([Source](https://aws.amazon.com/bedrock/claude/))
116-
- Claude 3 Sonnet: "English, Spanish, Japanese, and multiple other languages" ([Source](https://aws.amazon.com/bedrock/claude/))
117-
- Amazon Nova Pro: "200+ languages" ([Source](https://aws.amazon.com/ai/generative-ai/nova/))
118-
- Amazon Nova Lite: "200+ languages" ([Source](https://aws.amazon.com/ai/generative-ai/nova/))
119-
- Meta Llama 3.2 90B Instruct: "English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai" ([Source](https://aws.amazon.com/bedrock/llama/))
120-
- Meta Llama 3.2 11B Instruct: "English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai" ([Source](https://aws.amazon.com/bedrock/llama/))
111+
- [Claude](https://aws.amazon.com/bedrock/claude/)
112+
- [Nova](https://aws.amazon.com/ai/generative-ai/nova/)
113+
- [Llama](https://aws.amazon.com/bedrock/llama/)
114+
115+
- Vertex AI
116+
117+
- [Gemini](https://cloud.google.com/vertex-ai/generative-ai/docs/models#expandable-1)
121118

122119
## Learn more
123120

ui/summarizing.mdx

Lines changed: 3 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -3,11 +3,7 @@ title: Summarizing
33
---
44

55
After partitioning, _summarizing_ generates text-based summaries of images and tables.
6-
This summarization is done by using models offered through these providers:
7-
8-
- [GPT-4o](https://openai.com/index/hello-gpt-4o/), provided through OpenAI.
9-
- [Claude 3.5 Sonnet](https://www.anthropic.com/news/claude-3-5-sonnet), provided through Anthropic.
10-
- [Claude 3.5 Sonnet](https://aws.amazon.com/bedrock/claude/), provided through Amazon Bedrock.
6+
This summarization is done by using models offered through various model providers.
117

128
Here is an example of the output of image summarization using GPT-4o. Note specifically the `text` field that is added.
139
Line breaks have been inserted here for readability. The output will not contain these line breaks.
@@ -86,14 +82,6 @@ To summarize images or tables, in the **Task** drop-down list of an **Enrichment
8682

8783
<EnrichmentImagesTablesHiResOnly />
8884

89-
For image summarization, select **Image Description**, and then choose one of the following provider (and model) combinations to use:
90-
91-
- **OpenAI (GPT-4o)**. [Learn more](https://openai.com/index/hello-gpt-4o/).
92-
- **Anthropic (Claude 3.5 Sonnet)**. [Learn more](https://www.anthropic.com/news/claude-3-5-sonnet).
93-
- **Amazon Bedrock (Claude 3.5 Sonnet)**. [Learn more](https://aws.amazon.com/bedrock/claude/).
94-
95-
For table summarization, select **Table Description**, and then choose one of the following provider (and model) combinations to use:
85+
For image summarization, select **Image Description**, and then choose one of the available provider (and model) combinations that are shown.
9686

97-
- **OpenAI (GPT-4o)**. [Learn more](https://openai.com/index/hello-gpt-4o/).
98-
- **Anthropic (Claude 3.5 Sonnet)**. [Learn more](https://www.anthropic.com/news/claude-3-5-sonnet).
99-
- **Amazon Bedrock (Claude 3.5 Sonnet)**. [Learn more](https://aws.amazon.com/bedrock/claude/).
87+
For table summarization, select **Table Description**, and then choose one of the available provider (and model) combinations that are shown.

0 commit comments

Comments
 (0)