Skip to content

Commit 9561e63

Browse files
authored
Merge branch 'main' into rishibommasani/genie-2-1
2 parents 837634e + 3f1bc0c commit 9561e63

File tree

16 files changed

+1524
-834
lines changed

16 files changed

+1524
-834
lines changed

assets/amazon.yaml

Lines changed: 73 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -86,3 +86,76 @@
8686
prohibited_uses: ''
8787
monitoring: ''
8888
feedback: https://github.com/amazon-science/chronos-forecasting/discussions
89+
- type: model
90+
name: Amazon Nova (Understanding)
91+
organization: Amazon Web Services (AWS)
92+
description: A new generation of state-of-the-art foundation models (FMs) that
93+
deliver frontier intelligence and industry leading price performance, available
94+
exclusively in Amazon Bedrock. Amazon Nova understanding models excel in Retrieval-Augmented
95+
Generation (RAG), function calling, and agentic applications.
96+
created_date: 2024-12-03
97+
url: https://aws.amazon.com/blogs/aws/introducing-amazon-nova-frontier-intelligence-and-industry-leading-price-performance/
98+
model_card: unknown
99+
modality:
100+
explanation: Amazon Nova understanding models accept text, image, or video inputs
101+
to generate text output.
102+
value: text, image, video; text
103+
analysis: Amazon Nova Pro is capable of processing up to 300K input tokens and
104+
sets new standards in multimodal intelligence and agentic workflows that require
105+
calling APIs and tools to complete complex workflows. It achieves state-of-the-art
106+
performance on key benchmarks including visual question answering ( TextVQA
107+
) and video understanding ( VATEX ).
108+
size: unknown
109+
dependencies: []
110+
training_emissions: unknown
111+
training_time: unknown
112+
training_hardware: unknown
113+
quality_control: All Amazon Nova models include built-in safety controls and creative
114+
content generation models include watermarking capabilities to promote responsible
115+
AI use.
116+
access:
117+
explanation: available exclusively in Amazon Bedrock
118+
value: limited
119+
license: unknown
120+
intended_uses: You can build on Amazon Nova to analyze complex documents and videos,
121+
understand charts and diagrams, generate engaging video content, and build sophisticated
122+
AI agents, from across a range of intelligence classes optimized for enterprise
123+
workloads.
124+
prohibited_uses: unknown
125+
monitoring: unknown
126+
feedback: unknown
127+
- type: model
128+
name: Amazon Nova (Creative Content Generation)
129+
organization: Amazon Web Services (AWS)
130+
description: A new generation of state-of-the-art foundation models (FMs) that
131+
deliver frontier intelligence and industry leading price performance, available
132+
exclusively in Amazon Bedrock.
133+
created_date: 2024-12-03
134+
url: https://aws.amazon.com/blogs/aws/introducing-amazon-nova-frontier-intelligence-and-industry-leading-price-performance/
135+
model_card: unknown
136+
modality:
137+
explanation: Amazon creative content generation models accept text and image
138+
inputs to generate image or video output.
139+
value: text, image;image, video
140+
analysis: Amazon Nova Canvas excels on human evaluations and key benchmarks such
141+
as text-to-image faithfulness evaluation with question answering (TIFA) and
142+
ImageReward.
143+
size: unknown
144+
dependencies: []
145+
training_emissions: unknown
146+
training_time: unknown
147+
training_hardware: unknown
148+
quality_control: All Amazon Nova models include built-in safety controls and creative
149+
content generation models include watermarking capabilities to promote responsible
150+
AI use.
151+
access:
152+
explanation: available exclusively in Amazon Bedrock
153+
value: limited
154+
license: unknown
155+
intended_uses: You can build on Amazon Nova to analyze complex documents and videos,
156+
understand charts and diagrams, generate engaging video content, and build sophisticated
157+
AI agents, from across a range of intelligence classes optimized for enterprise
158+
workloads.
159+
prohibited_uses: unknown
160+
monitoring: unknown
161+
feedback: unknown

assets/anthropic.yaml

Lines changed: 42 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -608,15 +608,17 @@
608608
speed of its predecessor, Claude 3 Opus, and is designed to tackle tasks like
609609
context-sensitive customer support, orchestrating multi-step workflows, interpreting
610610
charts and graphs, and transcribing text from images.
611-
created_date: 2024-06-21
612-
url: https://www.anthropic.com/news/claude-3-5-sonnet
611+
created_date:
612+
explanation: Claude 3.5 Sonnet updated on Oct. 22, initially released on June
613+
20 of the same year.
614+
value: 2024-10-22
615+
url: https://www.anthropic.com/news/3-5-models-and-computer-use
613616
model_card: unknown
614617
modality: text; image, text
615618
analysis: The model has been evaluated on a range of tests including graduate-level
616619
reasoning (GPQA), undergraduate-level knowledge (MMLU), coding proficiency (HumanEval),
617-
and standard vision benchmarks. In an internal agentic coding evaluation, Claude
618-
3.5 Sonnet solved 64% of problems, outperforming the previous version, Claude
619-
3 Opus, which solved 38%.
620+
and standard vision benchmarks. Claude 3.5 Sonnet demonstrates state-of-the-art
621+
performance on most benchmarks.
620622
size: Unknown
621623
dependencies: []
622624
training_emissions: Unknown
@@ -637,3 +639,38 @@
637639
integrated to ensure robustness of evaluations.
638640
feedback: Feedback on Claude 3.5 Sonnet can be submitted directly in-product to
639641
inform the development roadmap and improve user experience.
642+
- type: model
643+
name: Claude 3.5 Haiku
644+
organization: Anthropic
645+
description: Claude 3.5 Haiku is Anthropic's fastest model, delivering advanced
646+
coding, tool use, and reasoning capability, surpassing the previous Claude 3
647+
Opus in intelligence benchmarks. It is designed for critical use cases where
648+
low latency is essential, such as user-facing chatbots and code completions.
649+
created_date: 2024-10-22
650+
url: https://www.anthropic.com/claude/haiku
651+
model_card: unknown
652+
modality:
653+
explanation: Claude 3.5 Haiku is available...initially as a text-only model
654+
and with image input to follow.
655+
value: text; unknown
656+
analysis: Claude 3.5 Haiku offers strong performance and speed across a variety
657+
of coding, tool use, and reasoning tasks. Also, it has been tested in extensive
658+
safety evaluations and exceeded expectations in reasoning and code generation
659+
tasks.
660+
size: unknown
661+
dependencies: []
662+
training_emissions: unknown
663+
training_time: unknown
664+
training_hardware: unknown
665+
quality_control: During Claude 3.5 Haiku’s development, we conducted extensive
666+
safety evaluations spanning multiple languages and policy domains.
667+
access:
668+
explanation: Claude 3.5 Haiku is available across Claude.ai, our first-party
669+
API, Amazon Bedrock, and Google Cloud’s Vertex AI.
670+
value: open
671+
license: unknown
672+
intended_uses: Critical use cases where low latency matters, like user-facing
673+
chatbots and code completions.
674+
prohibited_uses: unknown
675+
monitoring: unknown
676+
feedback: unknown

assets/cohere.yaml

Lines changed: 22 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -592,3 +592,25 @@
592592
prohibited_uses: unknown
593593
monitoring: unknown
594594
feedback: https://huggingface.co/CohereForAI/aya-23-35B/discussions
595+
- type: model
596+
name: Command R+
597+
organization: Cohere
598+
description: Command R+ is a state-of-the-art RAG-optimized model designed to
599+
tackle enterprise-grade workloads, and is available first on Microsoft Azure.
600+
created_date: 2024-04-04
601+
url: https://cohere.com/blog/command-r-plus-microsoft-azure
602+
model_card: unknown
603+
modality: unknown
604+
analysis: unknown
605+
size: unknown
606+
dependencies: []
607+
training_emissions: unknown
608+
training_time: unknown
609+
training_hardware: unknown
610+
quality_control: unknown
611+
access: ''
612+
license: unknown
613+
intended_uses: unknown
614+
prohibited_uses: unknown
615+
monitoring: unknown
616+
feedback: unknown

assets/genmo.yaml

Lines changed: 43 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,43 @@
1+
---
2+
- type: model
3+
name: Mochi 1
4+
organization: Genmo
5+
description: Mochi 1 is an open-source video generation model designed to produce
6+
high-fidelity motion and strong prompt adherence in generated videos, setting
7+
a new standard for open video generation systems.
8+
created_date: 2025-01-14
9+
url: https://www.genmo.ai/blog
10+
model_card: unknown
11+
modality:
12+
explanation: Mochi 1 generates smooth videos... Measures how accurately generated
13+
videos follow the provided textual instructions
14+
value: text; video
15+
analysis: Mochi 1 sets a new best-in-class standard for open-source video generation.
16+
It also performs very competitively with the leading closed models... We benchmark
17+
prompt adherence with an automated metric using a vision language model as a
18+
judge following the protocol in OpenAI DALL-E 3. We evaluate generated videos
19+
using Gemini-1.5-Pro-002.
20+
size:
21+
explanation: featuring a 10 billion parameter diffusion model
22+
value: 10B parameters
23+
dependencies: [DDPM, DreamFusion, Emu Video, T5-XXL]
24+
training_emissions: unknown
25+
training_time: unknown
26+
training_hardware: unknown
27+
quality_control: robust safety moderation protocols in the playground to ensure
28+
that all video generations remain safe and aligned with ethical guidelines.
29+
access:
30+
explanation: open state-of-the-art video generation model... The weights and
31+
architecture for Mochi 1 are open
32+
value: open
33+
license:
34+
explanation: We're releasing the model under a permissive Apache 2.0 license.
35+
value: Apache 2.0
36+
intended_uses: Advance the field of video generation and explore new methodologies.
37+
Build innovative applications in entertainment, advertising, education, and
38+
more. Empower artists and creators to bring their visions to life with AI-generated
39+
videos. Generate synthetic data for training AI models in robotics, autonomous
40+
vehicles and virtual environments.
41+
prohibited_uses: unknown
42+
monitoring: unknown
43+
feedback: unknown

assets/google.yaml

Lines changed: 65 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1931,4 +1931,69 @@
19311931
prohibited_uses: Unknown
19321932
monitoring: Unknown
19331933
feedback: Unknown
1934+
- type: model
1935+
name: Veo 2
1936+
organization: Google DeepMind
1937+
description: Veo 2 is a state-of-the-art video generation model that creates videos
1938+
with realistic motion and high-quality output, up to 4K, with extensive camera
1939+
controls. It simulates real-world physics and offers advanced motion capabilities
1940+
with enhanced realism and fidelity.
1941+
created_date: 2024-12-16
1942+
url: https://deepmind.google/technologies/veo/veo-2/
1943+
model_card: unknown
1944+
modality:
1945+
explanation: Our state-of-the-art video generation model ... text-to-image model
1946+
Veo 2
1947+
value: text; video
1948+
analysis: Veo 2 outperforms other leading video generation models, based on human
1949+
evaluations of its performance.
1950+
size: unknown
1951+
dependencies: []
1952+
training_emissions: unknown
1953+
training_time: unknown
1954+
training_hardware: unknown
1955+
quality_control: Veo 2 includes features that enhance realism, fidelity, detail,
1956+
and artifact reduction to ensure high-quality output.
1957+
access: limited
1958+
license: unknown
1959+
intended_uses: Creating high-quality videos with realistic motion, different styles,
1960+
camera controls, shot styles, angles, and movements.
1961+
prohibited_uses: unknown
1962+
monitoring: unknown
1963+
feedback: unknown
19341964

1965+
- type: model
1966+
name: Gemini 2.0
1967+
organization: Google DeepMind
1968+
description: Google DeepMind introduces Gemini 2.0, a new AI model designed for
1969+
the 'agentic era.'
1970+
created_date: 2024-12-11
1971+
url: https://blog.google/technology/google-deepmind/google-gemini-ai-update-december-2024/#ceo-message
1972+
model_card: unknown
1973+
modality:
1974+
explanation: The first model built to be natively multimodal, Gemini 1.0 and
1975+
1.5 drove big advances with multimodality and long context to understand information
1976+
across text, video, images, audio and code...
1977+
value: text, video, image, audio; image, text
1978+
analysis: unknown
1979+
size: unknown
1980+
dependencies: []
1981+
training_emissions: unknown
1982+
training_time: unknown
1983+
training_hardware:
1984+
explanation: It’s built on custom hardware like Trillium, our sixth-generation
1985+
TPUs.
1986+
value: custom hardware like Trillium, our sixth-generation TPUs
1987+
quality_control: Google is committed to building AI responsibly, with safety and
1988+
security as key priorities.
1989+
access:
1990+
explanation: Gemini 2.0 Flash is available to developers and trusted testers,
1991+
with wider availability planned for early next year.
1992+
value: limited
1993+
license: unknown
1994+
intended_uses: Develop more agentic models, meaning they can understand more about
1995+
the world around you, think multiple steps ahead, and take action on your behalf,
1996+
with your supervision.
1997+
prohibited_uses: unknown
1998+
monitoring: unknown
1999+
feedback: unknown

assets/ibm.yaml

Lines changed: 42 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -75,3 +75,45 @@
7575
prohibited_uses: ''
7676
monitoring: ''
7777
feedback: ''
78+
- type: model
79+
name: IBM Granite 3.0
80+
organization: IBM
81+
description: IBM Granite 3.0 models deliver state-of-the-art performance relative
82+
to model size while maximizing safety, speed and cost-efficiency for enterprise
83+
use cases.
84+
created_date: 2024-10-21
85+
url: https://www.ibm.com/new/ibm-granite-3-0-open-state-of-the-art-enterprise-models
86+
model_card: unknown
87+
modality:
88+
explanation: IBM Granite 3.0 8B Instruct model for classic natural language
89+
use cases including text generation, classification, summarization, entity
90+
extraction and customer service chatbots
91+
value: text; text
92+
analysis: Granite 3.0 8B Instruct matches leading similarly-sized open models
93+
on academic benchmarks while outperforming those peers on benchmarks for enterprise
94+
tasks and safety.
95+
size:
96+
explanation: 'Dense, general purpose LLMs: Granite-3.0-8B-Instruct'
97+
value: 8B parameters
98+
dependencies: [Hugging Face’s OpenLLM Leaderboard v2]
99+
training_emissions: unknown
100+
training_time: unknown
101+
training_hardware: unknown
102+
quality_control: The entire Granite family of models are trained on carefully
103+
curated enterprise datasets, filtered for objectionable content with critical
104+
concerns like governance, risk, privacy and bias mitigation in mind
105+
access:
106+
explanation: In keeping with IBM’s strong historical commitment to open source
107+
, all Granite models are released under the permissive Apache 2.0 license
108+
value: open
109+
license:
110+
explanation: In keeping with IBM’s strong historical commitment to open source
111+
, all Granite models are released under the permissive Apache 2.0 license
112+
value: Apache 2.0
113+
intended_uses: classic natural language use cases including text generation, classification,
114+
summarization, entity extraction and customer service chatbots, programming
115+
language use cases such as code generation, code explanation and code editing,
116+
and for agentic use cases requiring tool calling
117+
prohibited_uses: unknown
118+
monitoring: ''
119+
feedback: unknown

assets/inflection.yaml

Lines changed: 26 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -93,3 +93,29 @@
9393
prohibited_uses: ''
9494
monitoring: ''
9595
feedback: none
96+
- type: model
97+
name: Inflection 3.0
98+
organization: Inflection AI
99+
description: Inflection for Enterprise, powered by our industry-first, enterprise-grade
100+
AI system, Inflection 3.0.
101+
created_date: 2024-10-07
102+
url: https://inflection.ai/blog/enterprise
103+
model_card: unknown
104+
modality: unknown
105+
analysis: unknown
106+
size: unknown
107+
dependencies: []
108+
training_emissions: unknown
109+
training_time: unknown
110+
training_hardware: unknown
111+
quality_control: unknown
112+
access:
113+
explanation: Developers can now access Inflection AI’s Large Language Model
114+
through its new commercial API.
115+
value: open
116+
license: unknown
117+
intended_uses: unknown
118+
prohibited_uses: unknown
119+
monitoring: unknown
120+
feedback: So please drop us a line. We want to keep hearing from enterprises about
121+
how we can help solve their challenges and make AI a reality for their business.

assets/maya.yaml

Lines changed: 24 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -22,3 +22,27 @@
2222
prohibited_uses: ''
2323
monitoring: unknown
2424
feedback: none
25+
- type: model
26+
name: voyage-code-3
27+
organization: Voyage AI
28+
description: "Introducing voyage-code-3, our next-generation embedding model optimized for code retrieval."
29+
created_date: 2024-12-04
30+
url: https://blog.voyageai.com/2024/12/04/voyage-code-3/
31+
model_card: unknown
32+
modality: unknown
33+
analysis: "We evaluated voyage-code-3 using an enhanced suite of evaluation datasets designed to address the shortcomings of existing benchmarks and deliver practical, robust results."
34+
size: unknown
35+
dependencies: []
36+
training_emissions: unknown
37+
training_time: unknown
38+
training_hardware: unknown
39+
quality_control: unknown
40+
access:
41+
explanation: "The first 200 million tokens are free."
42+
value: limited
43+
license: unknown
44+
intended_uses: "optimized for code retrieval"
45+
prohibited_uses: unknown
46+
monitoring: unknown
47+
feedback: "If you’re also interested in fine-tuned embedding models, we’d love to hear from you—please email us at contact@voyageai.com."
48+

0 commit comments

Comments
 (0)