diff --git a/tasks/src/text-to-image/data.ts b/tasks/src/text-to-image/data.ts new file mode 100644 index 000000000..2798a38a4 --- /dev/null +++ b/tasks/src/text-to-image/data.ts @@ -0,0 +1,34 @@ +import type { TaskDataCustom } from "../Types"; + +const taskData: TaskDataCustom = { + datasets: [ + { + description: "RedCaps is a large-scale dataset of 12M image-text pairs collected from Reddit.", + id: "red_caps", + }, + { + description: "Conceptual Captions is a dataset consisting of ~3.3M images annotated with captions.", + id: "conceptual_captions", + }, + ], + demo: { + inputs: [], + outputs: [], + }, + metrics: [], + models: [ + { + description: "A model that can be used to generate images based on text prompts. The DALLĀ·E Mega model is the largest version of DALLE Mini.", + id: "dalle-mini/dalle-mega", + }, + { + description: "A latent text-to-image diffusion model capable of generating photo-realistic images given any text input.", + id: "CompVis/stable-diffusion-v1-4", + }, + ], + summary: "Generates images from input text. These models can be used to generate and modify images based on text prompts.", + widgetModels: [], + youtubeId: "", +}; + +export default taskData;