Skip to content

Commit fc65072

Browse files
authored
Add journal articles from 2020 to 2023
1 parent 313c539 commit fc65072

File tree

14 files changed

+480
-0
lines changed

14 files changed

+480
-0
lines changed
Lines changed: 18 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,18 @@
1+
@article{DBLP:journals/corr/abs-2011-09501,
2+
author = {Yixin Guo and
3+
Pengcheng Li and
4+
Yingwei Luo and
5+
Xiaolin Wang and
6+
Zhenlin Wang},
7+
title = {{GRAPHSPY:} Fused Program Semantic-Level Embedding via Graph Neural
8+
Networks for Dead Store Detection},
9+
journal = {CoRR},
10+
volume = {abs/2011.09501},
11+
year = {2020},
12+
url = {https://arxiv.org/abs/2011.09501},
13+
eprinttype = {arXiv},
14+
eprint = {2011.09501},
15+
timestamp = {Mon, 22 Jul 2024 08:26:51 +0200},
16+
biburl = {https://dblp.org/rec/journals/corr/abs-2011-09501.bib},
17+
bibsource = {dblp computer science bibliography, https://dblp.org}
18+
}
Lines changed: 49 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,49 @@
1+
---
2+
title: "GRAPHSPY: Fused Program Semantic-Level Embedding via Graph Neural Networks for Dead Store Detection"
3+
4+
authors:
5+
- Yixin Guo
6+
- Pengcheng Li
7+
- Yingwei Luo
8+
- Xiaolin Wang
9+
- Zhenlin Wang
10+
11+
date: "2020-11-01T00:00:00Z"
12+
doi: ""
13+
14+
# Schedule page publish date (NOT publication's date).
15+
publishDate: "2020-11-01T00:00:00Z"
16+
17+
# Publication type.
18+
# Accepts a single type but formatted as a YAML list (for Hugo requirements).
19+
# Enter a publication type from the CSL standard.
20+
publication_types: ["article-journal"]
21+
22+
# Publication name and optional abbreviated publication name.
23+
publication: "*Computing Research Repository*"
24+
publication_short: ""
25+
26+
abstract: 'Production software oftentimes suffers from the issue of performance inefficiencies caused by inappropriate use of data structures, programming abstractions, and conservative compiler optimizations. It is desirable to avoid unnecessary memory operations. However, existing works often use a whole-program fine-grained monitoring method with incredibly high overhead. To this end, we propose a learning-aided approach to identify unnecessary memory operations intelligently with low overhead. By applying several prevalent graph neural network models to extract program semantics with respect to program structure, execution order and dynamic states, we present a novel, hybrid program embedding approach so that to derive unnecessary memory operations through the embedding. We train our model with tens of thousands of samples acquired from a set of real-world benchmarks. Results show that our model achieves 90% of accuracy and incurs only around a half of time overhead of the state-of-art tool.'
27+
28+
# Summary. An optional shortened abstract.
29+
summary: ''
30+
31+
tags: []
32+
33+
# Display this page in the Featured widget?
34+
featured: true
35+
36+
# Custom links (uncomment lines below)
37+
# links:
38+
# - name: Custom Link
39+
# url: http://example.org
40+
41+
url_pdf: 'https://arxiv.org/pdf/2011.09501'
42+
url_code: ''
43+
url_dataset: ''
44+
url_poster: ''
45+
url_project: ''
46+
url_slides: ''
47+
url_source: ''
48+
url_video: ''
49+
---
Lines changed: 18 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,18 @@
1+
@article{DBLP:journals/jcst/ShaHLWW20,
2+
author = {Sai Sha and
3+
Jingyuan Hu and
4+
Yingwei Luo and
5+
Xiaolin Wang and
6+
Zhenlin Wang},
7+
title = {Huge Page Friendly Virtualized Memory Management},
8+
journal = {J. Comput. Sci. Technol.},
9+
volume = {35},
10+
number = {2},
11+
pages = {433--452},
12+
year = {2020},
13+
url = {https://doi.org/10.1007/s11390-020-9693-0},
14+
doi = {10.1007/S11390-020-9693-0},
15+
timestamp = {Tue, 21 Mar 2023 21:05:42 +0100},
16+
biburl = {https://dblp.org/rec/journals/jcst/ShaHLWW20.bib},
17+
bibsource = {dblp computer science bibliography, https://dblp.org}
18+
}
Lines changed: 49 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,49 @@
1+
---
2+
title: "Huge Page Friendly Virtualized Memory Management"
3+
4+
authors:
5+
- Sai Sha
6+
- Jingyuan Hu
7+
- Yingwei Luo
8+
- Xiaolin Wang
9+
- Zhenlin Wang
10+
11+
date: "2020-01-01T00:00:00Z"
12+
doi: ""
13+
14+
# Schedule page publish date (NOT publication's date).
15+
publishDate: "2020-01-01T00:00:00Z"
16+
17+
# Publication type.
18+
# Accepts a single type but formatted as a YAML list (for Hugo requirements).
19+
# Enter a publication type from the CSL standard.
20+
publication_types: ["article-journal"]
21+
22+
# Publication name and optional abbreviated publication name.
23+
publication: "*Journal of Computer Science and Technology*"
24+
publication_short: ""
25+
26+
abstract: 'With the rapid increase of memory consumption by applications running on cloud data centers, we need more efficient memory management in a virtualized environment. Exploiting huge pages becomes more critical for a virtual machine’s performance when it runs large working set size programs. Programs with large working set sizes are more sensitive to memory allocation, which requires us to quickly adjust the virtual machine’s memory to accommodate memory phase changes. It would be much more efficient if we could adjust virtual machines’ memory at the granularity of huge pages. However, existing virtual machine memory reallocation techniques, such as ballooning, do not support huge pages. In addition, in order to drive effective memory reallocation, we need to predict the actual memory demand of a virtual machine. We find that traditional memory demand estimation methods designed for regular pages cannot be simply ported to a system adopting huge pages. How to adjust the memory of virtual machines timely and effectively according to the periodic change of memory demand is another challenge we face. This paper proposes a dynamic huge page based memory balancing system (HPMBS) for efficient memory management in a virtualized environment. We first rebuild the ballooning mechanism in order to dispatch memory in the granularity of huge pages. We then design and implement a huge page working set size estimation mechanism which can accurately estimate a virtual machine’s memory demand in huge pages environments. Combining these two mechanisms, we finally use an algorithm based on dynamic programming to achieve dynamic memory balancing. Experiments show that our system saves memory and improves overall system performance with low overhead.'
27+
28+
# Summary. An optional shortened abstract.
29+
summary: ''
30+
31+
tags: []
32+
33+
# Display this page in the Featured widget?
34+
featured: true
35+
36+
# Custom links (uncomment lines below)
37+
# links:
38+
# - name: Custom Link
39+
# url: http://example.org
40+
41+
url_pdf: 'https://link.springer.com/article/10.1007/s11390-020-9693-0'
42+
url_code: ''
43+
url_dataset: ''
44+
url_poster: ''
45+
url_project: ''
46+
url_slides: ''
47+
url_source: ''
48+
url_video: ''
49+
---
Lines changed: 18 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,18 @@
1+
@article{DBLP:journals/tos/PanWLW21,
2+
author = {Cheng Pan and
3+
Xiaolin Wang and
4+
Yingwei Luo and
5+
Zhenlin Wang},
6+
title = {Penalty- and Locality-aware Memory Allocation in Redis Using Enhanced
7+
{AET}},
8+
journal = {{ACM} Trans. Storage},
9+
volume = {17},
10+
number = {2},
11+
pages = {15:1--15:45},
12+
year = {2021},
13+
url = {https://doi.org/10.1145/3447573},
14+
doi = {10.1145/3447573},
15+
timestamp = {Sun, 19 Jan 2025 14:52:13 +0100},
16+
biburl = {https://dblp.org/rec/journals/tos/PanWLW21.bib},
17+
bibsource = {dblp computer science bibliography, https://dblp.org}
18+
}
Lines changed: 49 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,49 @@
1+
---
2+
title: "Penalty- and Locality-aware Memory Allocation in Redis Using Enhanced AET"
3+
4+
authors:
5+
- Cheng Pan
6+
- Xiaolin Wang
7+
- Yingwei Luo
8+
- Zhenlin Wang
9+
10+
date: "2021-02-01T00:00:00Z"
11+
doi: ""
12+
13+
# Schedule page publish date (NOT publication's date).
14+
publishDate: "2021-02-01T00:00:00Z"
15+
16+
# Publication type.
17+
# Accepts a single type but formatted as a YAML list (for Hugo requirements).
18+
# Enter a publication type from the CSL standard.
19+
publication_types: ["article-journal"]
20+
21+
# Publication name and optional abbreviated publication name.
22+
publication: "*ACM Transactions on Storage*"
23+
publication_short: ""
24+
25+
abstract: 'Due to large data volume and low latency requirements of modern web services, the use of an in-memory key-value (KV) cache often becomes an inevitable choice (e.g., Redis and Memcached). The in-memory cache holds hot data, reduces request latency, and alleviates the load on background databases. Inheriting from the traditional hardware cache design, many existing KV cache systems still use recency-based cache replacement algorithms, e.g., least recently used or its approximations. However, the diversity of miss penalty distinguishes a KV cache from a hardware cache. Inadequate consideration of penalty can substantially compromise space utilization and request service time. KV accesses also demonstrate locality, which needs to be coordinated with miss penalty to guide cache management.
26+
In this article, we first discuss how to enhance the existing cache model, the Average Eviction Time model, so that it can adapt to modeling a KV cache. After that, we apply the model to Redis and propose pRedis, Penalty- and Locality-aware Memory Allocation in Redis, which synthesizes data locality and miss penalty, in a quantitative manner, to guide memory allocation and replacement in Redis. At the same time, we also explore the diurnal behavior of a KV store and exploit long-term reuse. We replace the original passive eviction mechanism with an automatic dump/load mechanism, to smooth the transition between access peaks and valleys. Our evaluation shows that pRedis effectively reduces the average and tail access latency with minimal time and space overhead. For both real-world and synthetic workloads, our approach delivers an average of 14.0%∼52.3% latency reduction over a state-of-the-art penalty-aware cache management scheme, Hyperbolic Caching (HC), and shows more quantitative predictability of performance. Moreover, we can obtain even lower average latency (1.1%∼5.5%) when dynamically switching policies between pRedis and HC.'
27+
28+
# Summary. An optional shortened abstract.
29+
summary: ''
30+
31+
tags: []
32+
33+
# Display this page in the Featured widget?
34+
featured: true
35+
36+
# Custom links (uncomment lines below)
37+
# links:
38+
# - name: Custom Link
39+
# url: http://example.org
40+
41+
url_pdf: 'https://dl.acm.org/doi/pdf/10.1145/3447573'
42+
url_code: ''
43+
url_dataset: ''
44+
url_poster: ''
45+
url_project: ''
46+
url_slides: ''
47+
url_source: ''
48+
url_video: ''
49+
---
Lines changed: 18 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,18 @@
1+
@article{DBLP:journals/corr/abs-2209-13111,
2+
author = {Sai Sha and
3+
Chuandong Li and
4+
Yingwei Luo and
5+
Xiaolin Wang and
6+
Zhenlin Wang},
7+
title = {{HMM-V:} Heterogeneous Memory Management for Virtualization},
8+
journal = {CoRR},
9+
volume = {abs/2209.13111},
10+
year = {2022},
11+
url = {https://doi.org/10.48550/arXiv.2209.13111},
12+
doi = {10.48550/ARXIV.2209.13111},
13+
eprinttype = {arXiv},
14+
eprint = {2209.13111},
15+
timestamp = {Wed, 04 Oct 2023 19:04:03 +0200},
16+
biburl = {https://dblp.org/rec/journals/corr/abs-2209-13111.bib},
17+
bibsource = {dblp computer science bibliography, https://dblp.org}
18+
}
Lines changed: 50 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,50 @@
1+
---
2+
title: "HMM-V: Heterogeneous Memory Management for Virtualization"
3+
4+
authors:
5+
- Sai Sha
6+
- Chuandong Li
7+
- Yingwei Luo
8+
- Xiaolin Wang
9+
- Zhenlin Wang
10+
11+
date: "2022-09-01T00:00:00Z"
12+
doi: ""
13+
14+
# Schedule page publish date (NOT publication's date).
15+
publishDate: "2022-09-01T00:00:00Z"
16+
17+
# Publication type.
18+
# Accepts a single type but formatted as a YAML list (for Hugo requirements).
19+
# Enter a publication type from the CSL standard.
20+
publication_types: ["article-journal"]
21+
22+
# Publication name and optional abbreviated publication name.
23+
publication: "*Computing Research Repository*"
24+
publication_short: ""
25+
26+
abstract: 'The memory demand of virtual machines (VMs) is increasing, while DRAM has limited capacity and high power consumption. Non-volatile memory (NVM) is an alternative to DRAM, but it has high latency and low bandwidth. We observe that the VM with heterogeneous memory may incur up to a 1.5× slowdown compared to a DRAM VM, if not managed well. However, none of the state-of-the-art heterogeneous memory management designs are customized for virtualization on a real system.
27+
In this paper, we propose HMM-V, a Heterogeneous Memory Management system for Virtualization. HMM-V automatically determines page hotness and migrates pages between DRAM and NVM to achieve performance close to the DRAM system. First, HMM-V tracks memory accesses through page table manipulation, but reduces the cost by leveraging Intel page-modification logging (PML) and a multi-level queue. Second, HMM-V quantifies the ``temperature'' of page and determines the hot set with bucket-sorting. HMM-V then efficiently migrates pages with minimal access pause and handles dirty pages with the assistance of PML. Finally, HMM-V provides pooling management to balance precious DRAM across multiple VMs to maximize utilization and overall performance. HMM-V is implemented on a real system with Intel Optane DC persistent memory. The four-VM co-running results show that HMM-V outperforms NUMA balancing and hardware management (Intel Optane memory mode) by 51% and 31%, respectively.'
28+
29+
# Summary. An optional shortened abstract.
30+
summary: ''
31+
32+
tags: []
33+
34+
# Display this page in the Featured widget?
35+
featured: true
36+
37+
# Custom links (uncomment lines below)
38+
# links:
39+
# - name: Custom Link
40+
# url: http://example.org
41+
42+
url_pdf: 'https://arxiv.org/pdf/2209.13111'
43+
url_code: ''
44+
url_dataset: ''
45+
url_poster: ''
46+
url_project: ''
47+
url_slides: ''
48+
url_source: ''
49+
url_video: ''
50+
---
Lines changed: 19 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,19 @@
1+
@article{DBLP:journals/tc/ShaZLWW22,
2+
author = {Sai Sha and
3+
Yi Zhang and
4+
Yingwei Luo and
5+
Xiaolin Wang and
6+
Zhenlin Wang},
7+
title = {Accelerating Address Translation for Virtualization by Leveraging
8+
Hardware Mode},
9+
journal = {{IEEE} Trans. Computers},
10+
volume = {71},
11+
number = {11},
12+
pages = {3047--3060},
13+
year = {2022},
14+
url = {https://doi.org/10.1109/TC.2022.3145671},
15+
doi = {10.1109/TC.2022.3145671},
16+
timestamp = {Sun, 13 Nov 2022 17:52:42 +0100},
17+
biburl = {https://dblp.org/rec/journals/tc/ShaZLWW22.bib},
18+
bibsource = {dblp computer science bibliography, https://dblp.org}
19+
}
Lines changed: 49 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,49 @@
1+
---
2+
title: "Accelerating Address Translation for Virtualization by Leveraging Hardware Mode"
3+
4+
authors:
5+
- Sai Sha
6+
- Chuandong Li
7+
- Yingwei Luo
8+
- Xiaolin Wang
9+
- Zhenlin Wang
10+
11+
date: "2022-01-01T00:00:00Z"
12+
doi: ""
13+
14+
# Schedule page publish date (NOT publication's date).
15+
publishDate: "2022-01-01T00:00:00Z"
16+
17+
# Publication type.
18+
# Accepts a single type but formatted as a YAML list (for Hugo requirements).
19+
# Enter a publication type from the CSL standard.
20+
publication_types: ["article-journal"]
21+
22+
# Publication name and optional abbreviated publication name.
23+
publication: "*IEEE Transactions on Computers*"
24+
publication_short: ""
25+
26+
abstract: 'The overhead of memory virtualization remains nontrivial. The traditional shadow paging (TSP) resorts to a shadow page table (SPT) to achieve the native page walk speed, but page table updates require hypervisor interventions. Alternatively, nested paging enables low-overhead page table updates, but utilizes the hardware MMU to perform a long-latency two-dimensional page walk. This paper proposes new memory virtualization solutions based on hardware (machine) mode—the highest CPU privilege level in some architectures like Sunway and RISC-V. A programming interface, running in hardware mode, enables software-implementation of hardware support functions. We first propose Software-based Nested Paging (SNP), which extends the software MMU to perform a two-dimensional page walk in hardware mode. Second, we present Swift Shadow Paging (SSP), which accomplishes page table synchronization by intercepting TLB flushing in hardware mode. Finally we propose Accelerated Shadow Paging (ASP) combining SSP and SNP. ASP handles the last-level SPT page faults by walking two-dimensional page tables in hardware mode, which eliminates most hypervisor interventions. This paper systematically compares multiple memory virtualization models by analyzing their designs and evaluating their performance both on a real system and a simulator. The experiments show that the virtualization overhead of ASP is less than 4.5% for all workloads.'
27+
28+
# Summary. An optional shortened abstract.
29+
summary: ''
30+
31+
tags: []
32+
33+
# Display this page in the Featured widget?
34+
featured: true
35+
36+
# Custom links (uncomment lines below)
37+
# links:
38+
# - name: Custom Link
39+
# url: http://example.org
40+
41+
url_pdf: 'https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=9693114'
42+
url_code: ''
43+
url_dataset: ''
44+
url_poster: ''
45+
url_project: ''
46+
url_slides: ''
47+
url_source: ''
48+
url_video: ''
49+
---

0 commit comments

Comments
 (0)