You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: resources/all_assets.csv
+1Lines changed: 1 addition & 0 deletions
Original file line number
Diff line number
Diff line change
@@ -840,6 +840,7 @@ model,Gen-3 Alpha,"Runway AI, Inc.","Gen-3 Alpha is a foundation model trained f
840
840
model,EXAONE 3.0 Instruction Tuned Language Model,LG AI Research,EXAONE 3.0 is an instruction-tuned large language model developed by LG AI Research. It demonstrates notably robust performance across a range of tasks and benchmarks. It has been fine-tuned to be capable of complex reasoning and has a particular proficiency in Korean. The released 7.8B parameter model is designed to promote open research and innovation.,2024-09-08,https://arxiv.org/pdf/2408.03541,unknown,text; text,The model was evaluated extensively across a wide range of public and in-house benchmarks. The comparative analysis showed that the performance of EXAONE 3.0 was competitive in English and excellent in Korean compared to other large language models of a similar size.,7.8B parameters (dense),['MeCab'],Unknown,Unknown,Unknown,"Extensive pre-training on a diverse dataset, and advanced post-training techniques were employed to enhance instruction-following capabilities. The model was also trained to fully comply with data handling standards.",open,Unknown,"The model was intended for non-commercial and research purposes. The capabilities of the model allow for use cases that involve advanced AI and language processing tasks, particularly in fields requiring proficiency in English and Korean.",Commercial use is not intended for this model. Its intended use is for non-commercial research and innovation.,Unknown,Unknown,,,,,,,,,,,
841
841
model,CausalLM,CausalLM,CausalLM is an LLM based on the model weights of Qwen and trained on a model architecture identical to LLaMA 2.,2023-10-21,https://huggingface.co/CausalLM/14B,https://huggingface.co/CausalLM/14B,text; text,Evaluated on standard benchmarks across a range of tasks.,14B parameters (dense),"['Qwen', 'OpenOrca', 'Open Platypus']",unknown,unknown,unknown,,open,WTFPL,,,unknown,,,,,,,,,,,,
842
842
model,Aurora-M,"Tokyo Institute of Technology, MIT-IBM Watson Lab, Sapienza University of Rome","Aurora-M is a 15B parameter multilingual open-source model trained on English, Finnish, Hindi, Japanese, Vietnamese, and code.",2024-04-23,https://arxiv.org/pdf/2404.00399,,text; text,"Evaluated on all language datasets compared to similarly sized SOTA models, with Aurora-M achieving strong performance in most.",15B parameters,['StarCoderPlus'],unknown,48 days,"LUMI supercomputer, using 128 AMD MI250X GPUs",,open,unknown,,,unknown,,,,,,,,,,,,
843
+
model,Reflection Llama-3.1 70B,Unknown,"Reflection Llama-3.1 70B is an open-source LLM, trained with a new technique called Reflection-Tuning that teaches a LLM to detect mistakes in its reasoning and correct course. The model was trained on synthetic data generated by Glaive.",2024-09-28,https://huggingface.co/mattshumer/Reflection-70B,https://huggingface.co/mattshumer/Reflection-70B,text; text,Unknown,70B parameters,"['Glaive', 'Llama 3.1']",Unknown,Unknown,Unknown,The model uses a Reflection-Tuning technique which allows it to notice mistakes in its reasoning and correct them.,open,Llama 3.1 Community License Agreement,The model is intended for complex reasoning and reflection tasks. It is designed to separate its internal thoughts and reasoning from its final answer.,Unknown,Unknown,Unknown,,,,,,,,,,,
843
844
model,Pixtral 12B,Unknown,"Pixtral 12B is a 12-billion-parameter multimodal language model trained to understand both natural images and documents, achieving leading performance on various multimodal benchmarks without compromising on natural language performance.",2024-10-10,https://arxiv.org/pdf/2410.07073,unknown,text; image,"The model outperforms other models of similar and larger sizes on multimodal benchmarks. An open-source benchmark, MM-MT-Bench, is contributed for evaluating vision-language models.",12B parameters,['Mistral Nemo 12B'],Unknown,Unknown,Unknown,Evaluation protocols for multimodal language models were standardized and analysis was conducted to improve the reliability of model evaluations.,open,Apache 2.0,"Multimodal instruction following tasks, capable of multi-turn, multi-image conversations.",Unknown,Evaluation protocols and benchmarks are open-sourced to establish fair and standardized testing.,Unknown,,,,,,,,,,,
844
845
application,My AI for Snapchat,Snap,"My AI offers Snapchatters a friendly, customizable chatbot at their fingertips that offers recommendations, and can even write a haiku for friends in seconds. Snapchat, where communication and messaging is a daily behavior, has 750 million monthly Snapchatters.",2023-03-01,https://openai.com/blog/introducing-chatgpt-and-whisper-apis,,,,,['ChatGPT API'],,,,,open,custom,,,,,,,https://snap.com/terms,,,,,,,,
845
846
application,Auto-GPT,Auto-GPT,Auto-GPT is an experimental open-source application showcasing the capabilities of the GPT-4 language model.,2023-04-16,https://news.agpt.co/,,,,,['GPT-4 API'],,,,,open,MIT,,,,,"GPT-4 adapted to run autonomously by chaining together LLM ""thoughts""",text,,,,,,,,,
0 commit comments