Skip to content

Commit 7a48e35

Browse files
authored
cornstarch arxiv (#301)
1 parent 2bc70b6 commit 7a48e35

File tree

1 file changed

+19
-0
lines changed

1 file changed

+19
-0
lines changed

source/_data/SymbioticLab.bib

Lines changed: 19 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -2006,3 +2006,22 @@ @Article{curie:arxiv25
20062006
Scientific experimentation, a cornerstone of human progress, demands rigor in reliability, methodical control, and interpretability to yield meaningful results. Despite the growing capabilities of large language models (LLMs) in automating different aspects of the scientific process, automating rigorous experimentation remains a significant challenge. To address this gap, we propose Curie, an AI agent framework designed to embed rigor into the experimentation process through three key components: an intra-agent rigor module to enhance reliability, an inter-agent rigor module to maintain methodical control, and an experiment knowledge module to enhance interpretability. To evaluate Curie, we design a novel experimental benchmark composed of 46 questions across four computer science domains, derived from influential research papers, and widely adopted open-source projects. Compared to the strongest baseline tested, we achieve a 3.4× improvement in correctly answering experimental questions. Curie is open-sourced at https://github.com/Just-Curieous/Curie.
20072007
}
20082008
}
2009+
2010+
@Article{cornstarch:arxiv25,
2011+
author = {Insu Jang and Runyu Lu and Nikhil Bansal and Ang Chen and Mosharaf Chowdhury},
2012+
title = {Cornstarch: Distributed Multimodal Training Must Be Multimodality-Aware },
2013+
year = {2025},
2014+
month = {March},
2015+
volume = {abs/2503.11367},
2016+
archivePrefix = {arXiv},
2017+
eprint = {2503.11367},
2018+
url = {https://arxiv.org/abs/2503.11367},
2019+
publist_link = {code || https://github.com/cornstarch-org/Cornstarch},
2020+
publist_confkey = {arXiv:2503.11367},
2021+
publist_link = {paper || https://arxiv.org/abs/2503.11367},
2022+
publist_topic = {Systems + AI},
2023+
publist_abstract = {
2024+
Multimodal large language models (MLLMs) extend the capabilities of large language models (LLMs) by combining heterogeneous model architectures to handle diverse modalities like images and audio. However, this inherent heterogeneity in MLLM model structure and data types makes makeshift extensions to existing LLM training frameworks unsuitable for efficient MLLM training.
2025+
In this paper, we present Cornstarch, the first general-purpose distributed MLLM training framework. Cornstarch facilitates modular MLLM construction, enables composable parallelization of constituent models, and introduces MLLM-specific optimizations to pipeline and context parallelism for efficient distributed MLLM training. Our evaluation shows that Cornstarch outperforms state-of-the-art solutions by up to 1.57x in terms of training throughput.
2026+
}
2027+
}

0 commit comments

Comments
 (0)