Skip to content

Commit c75d72d

Browse files
authored
fedtrans mlsys'24 (#252)
* fedtrans * put FedTrans inside bracket * {FedTrans}
1 parent 55cdf1e commit c75d72d

File tree

3 files changed

+21
-0
lines changed

3 files changed

+21
-0
lines changed

source/_data/SymbioticLab.bib

Lines changed: 17 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1722,3 +1722,20 @@ @Article{crosslayer-energy:arxiv24
17221722
publist_abstract = {The enormous energy consumption of machine learning (ML) and generative AI workloads shows no sign of waning, taking a toll on operating costs, power delivery, and environmental sustainability. Despite a long line of research on energy-efficient hardware, we found that software plays a critical role in ML energy optimization through two recent works: Zeus and Perseus. This is especially true for large language models (LLMs) because their model sizes and, therefore, energy demands are growing faster than hardware efficiency improvements. Therefore, we advocate for a cross-layer approach for energy optimizations in ML systems, where hardware provides architectural support that pushes energy-efficient software further, while software leverages and abstracts the hardware to develop techniques that bring hardware-agnostic energy-efficiency gains.
17231723
}
17241724
}
1725+
1726+
@InProceedings{fedtrans:mlsys24,
1727+
author = {Yuxuan Zhu and Jiachen Liu and Fan Lai and Mosharaf Chowdhury},
1728+
booktitle = {MLSys},
1729+
title = {{FedTrans}: Efficient Federated Learning via Multi-Model Transformation},
1730+
year = {2024},
1731+
publist_confkey = {MLSys'24},
1732+
publist_topic = {Systems + AI},
1733+
publist_topic = {Wide-Area Computing},
1734+
publist_abstract = {
1735+
Federated learning (FL) aims to train machine learning (ML) models across potentially millions of edge client devices. Yet, training and customizing models for FL clients is notoriously challenging due to the heterogeneity of client data, device capabilities, and the massive scale of clients, making individualized model exploration prohibitively expensive. State-of-the-art FL solutions personalize a globally trained model or concurrently train multiple models, but they often incur suboptimal model accuracy and huge training costs.
1736+
1737+
In this paper, we introduce FedTrans, a multi-model FL training framework that automatically produces and trains high-accuracy, hardware-compatible models for individual clients at scale. FedTrans begins with a basic global model, identifies accuracy bottlenecks in model architectures during training, and then employs model transformation to derive new models for heterogeneous clients on the fly. It judiciously assigns models to individual clients while performing soft aggregation on multi-model updates to minimize total training costs. Our evaluations using realistic settings show that FedTrans improves individual client model accuracy by 13% while slashing training costs by 4x over state-of-the-art solutions.
1738+
}
1739+
}
1740+
1741+
1.79 MB
Binary file not shown.

source/publications/index.md

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -139,6 +139,10 @@ venues:
139139
MLSys:
140140
category: Conferences
141141
occurrences:
142+
- key: MLSys'24
143+
name: The 7th Conference on Machine Learning and Systems
144+
date: 2024-05-13
145+
url: https://mlsys.org/Conferences/2024
142146
- key: MLSys'23
143147
name: The 6th Conference on Machine Learning and Systems
144148
date: 2023-06-04

0 commit comments

Comments
 (0)