Detail View
MTAT: Adaptive Fast Memory Management for Co-located Latency-Critical Workloads in Tiered Memory System
Citations
WEB OF SCIENCE
Citations
SCOPUS
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Kim, Minho | - |
| dc.contributor.author | Han, Seonggyu | - |
| dc.contributor.author | Park, Gyeongseo | - |
| dc.contributor.author | Kim, Daehoon | - |
| dc.date.accessioned | 2026-02-09T23:40:10Z | - |
| dc.date.available | 2026-02-09T23:40:10Z | - |
| dc.date.created | 2026-01-15 | - |
| dc.date.issued | 2025-12-19 | - |
| dc.identifier.isbn | 9798400715549 | - |
| dc.identifier.uri | https://scholar.dgist.ac.kr/handle/20.500.11750/59998 | - |
| dc.description.abstract | Modern data centers increasingly employ multi-tenant deployment models in which multiple applications or virtual machines share a single physical server. However, existing tiered memory management schemes classify pages solely by access frequency to govern promotions and demotions across memory tiers without accounting for the distinct access patterns of latency-critical (LC) and best-effort (BE) workloads. LC workloads demand low-latency service yet lack sustained high-frequency access; consequently, frequency-based tiering demotes LC data to slower memory (SMem), degrading responsiveness and violating service-level objectives (SLOs).To address these challenges, we propose MTAT, an adaptive tiered memory management framework that guarantees the SLO of LC workloads while maintaining overall system performance for BE workloads. Rather than relying solely on hotness-based page placement, MTAT employs distinct policies for LC and BE workloads by isolating them into dedicated fast memory (FMem) partitions. Specifically, MTAT employs reinforcement learning to identify the minimal FMem capacity necessary to satisfy stringent SLOs, supporting rapid response to sudden demand surges, and uses a simulated annealing algorithm to allocate the remaining FMem fairly among co-located BE workloads. Compared to state-of-the-art tiered memory page-placement solutions, MTAT improves the maximum throughput of LC workloads by up to 1.7× and enhances BE workloads' fairness by up to 3.3×, all while incurring only a 19% throughput penalty at worst. | - |
| dc.language | English | - |
| dc.publisher | Association for Computing Machinery | - |
| dc.relation.ispartof | Middleware '25: Proceedings of the 26th International Middleware Conference | - |
| dc.title | MTAT: Adaptive Fast Memory Management for Co-located Latency-Critical Workloads in Tiered Memory System | - |
| dc.type | Conference Paper | - |
| dc.identifier.doi | 10.1145/3721462.3770767 | - |
| dc.identifier.scopusid | 2-s2.0-105026673174 | - |
| dc.identifier.bibliographicCitation | ACM/IFIP/USENIX International Middleware Conference, pp.86 - 98 | - |
| dc.identifier.url | https://middleware-conf.github.io/2025/program/full-program/ | - |
| dc.citation.conferenceDate | 2025-12-15 | - |
| dc.citation.conferencePlace | US | - |
| dc.citation.conferencePlace | Nashville | - |
| dc.citation.endPage | 98 | - |
| dc.citation.startPage | 86 | - |
| dc.citation.title | ACM/IFIP/USENIX International Middleware Conference | - |
File Downloads
공유
Total Views & Downloads
???jsp.display-item.statistics.view???: , ???jsp.display-item.statistics.download???:
