Skip to content

Commit 472effa

Browse files
authored
Memstrata camera-ready (#259)
1 parent 0292755 commit 472effa

File tree

3 files changed

+21
-0
lines changed

3 files changed

+21
-0
lines changed

source/_data/SymbioticLab.bib

Lines changed: 16 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1778,3 +1778,19 @@ @Article{andes:arxiv24
17781778
}
17791779
}
17801780
1781+
@InProceedings{memstrata:osdi24,
1782+
author = {Yuhong Zhong and Daniel S. Berger and Carl Waldspurger and Ryan Wee and Ishwar Agarwal and Rajat Agarwal and Frank Hady and Karthik Kumar and Mark D. Hill and Mosharaf Chowdhury and Asaf Cidon},
1783+
booktitle = {OSDI},
1784+
title = {Managing Memory Tiers with {CXL} in Virtualized Environments},
1785+
year = {2024},
1786+
publist_confkey = {OSDI'24},
1787+
publist_topic = {Disaggregation},
1788+
publist_link = {paper || memstrata-osdi24.pdf},
1789+
publist_abstract = {
1790+
Cloud providers seek to deploy CXL-based memory to increase aggregate memory capacity, reduce costs, and lower carbon emissions. However, CXL accesses incur higher latency than local DRAM. Existing systems use software to manage data placement across memory tiers at page granularity. Cloud providers are reluctant to deploy software-based tiering due to high overheads in virtualized environments. Hardware-based memory tiering could place data at cacheline granularity, mitigating these drawbacks. However, hardware is oblivious to application-level performance.
1791+
1792+
We propose combining hardware-managed tiering with software-managed performance isolation to overcome the pitfalls of either approach. We introduce Intel® Flat Memory Mode, the first hardware-managed tiering system for CXL. Our evaluation on a full-system prototype demonstrates that it provides performance close to regular DRAM, with no more than 5% degradation for more than 82% of workloads. Despite such small slowdowns, we identify two challenges that can still degrade performance by up to 34% for “outlier” workloads: (1) memory contention across tenants, and (2) intra-tenant contention due to conflicting access patterns.
1793+
1794+
To address these challenges, we introduce Memstrata, a lightweight multi-tenant memory allocator. Memstrata employs page coloring to eliminate inter-VM contention. It improves performance for VMs with access patterns that are sensitive to hardware tiering by allocating them more local DRAM using an online slowdown estimator. In multi-VM experiments on prototype hardware, Memstrata is able to identify performance outliers and reduce their degradation from above 30% to below 6%, providing consistent performance across a wide range of workloads.
1795+
}
1796+
}
1.93 MB
Binary file not shown.

source/publications/index.md

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -205,6 +205,11 @@ venues:
205205
OSDI:
206206
category: Conferences
207207
occurrences:
208+
- key: OSDI'24
209+
name: The 18th USENIX Symposium on Operating Systems Design and Implementation
210+
date: 2024-07-10
211+
url: https://www.usenix.org/conference/osdi24
212+
acceptance: 15.6%
208213
- key: OSDI'23
209214
name: The 17th USENIX Symposium on Operating Systems Design and Implementation
210215
date: 2023-07-10

0 commit comments

Comments
 (0)