Cited time in webofscience Cited time in scopus

Full metadata record

DC Field Value Language
dc.contributor.author Alian, Mohammad -
dc.contributor.author Agarwal, Siddharth -
dc.contributor.author Shin, Jongmin -
dc.contributor.author Patel, Neel -
dc.contributor.author Yuan, Yifan -
dc.contributor.author Kim, Daehoon -
dc.contributor.author Wang, Ren -
dc.contributor.author Kim, Nam Sung -
dc.date.accessioned 2023-12-26T18:12:47Z -
dc.date.available 2023-12-26T18:12:47Z -
dc.date.created 2022-12-30 -
dc.date.issued 2022-10-03 -
dc.identifier.isbn 9781665462723 -
dc.identifier.issn 1072-4451 -
dc.identifier.uri http://hdl.handle.net/20.500.11750/46812 -
dc.description.abstract High-bandwidth network interface cards (NICs), each capable of transferring 100s of Gigabits per second, are making inroads into the servers of next-generation datacenters. Such unprecedented data delivery rates impose immense pressure, especially on the server's memory subsystem, as NICs first transfer network data to DRAM before processing. To alleviate the pressure, the cache hierarchy has evolved, supporting a direct data I/O (DDIO) technology to directly place network data in the last-level cache (LLC). Subsequently, various policies have been explored to manage such LLC and have proven to effectively reduce service latency and memory bandwidth consumption of network applications. However, the more recent evolution of the cache hierarchy decreased the size of LLC per core but significantly increased that of midlevel cache (MLC) with a non-inclusive policy. This calls for a re-examination of the aforementioned DDIO technology and management policies. In this paper, first, we identify three shortcomings of the current static data placement policy placing network data to LLC first and the non-inclusive policy with a commercial server system: (1) ineffectively using large MLC, (2) suffering from high rates of writebacks from MLC to LLC, and (3) breaking the isolation between application and network data enforced by limiting cache ways for DDIO. Second, to tackle the three shortcomings, we propose an intelligent direct I/O (IDIO) technology that extends DDIO to MLC and provides three synergistic mechanisms: (1) self-invalidating I/O buffer, (2) network-driven MLC prefetching, and (3) selective direct DRAM access. Our detailed experiments using a full-system simulator - capable of running modern DPDK userspace network functions while sustaining 100Gbps + network bandwidth - show that IDIO significantly reduces data movement (up to 84% MLC and LLC writeback reduction), provides LLC isolation (up to 22% performance improvement), and improves tail latency (up to 38% reduction in 99th latency) for receive-intensive network applications. © 2022 IEEE. -
dc.language English -
dc.publisher IEEE Computer Society -
dc.title IDIO: Network-Driven, Inbound Network Data Orchestration on Server Processors -
dc.type Conference Paper -
dc.identifier.doi 10.1109/MICRO56248.2022.00042 -
dc.identifier.scopusid 2-s2.0-85141705887 -
dc.identifier.bibliographicCitation IEEE/ACM International Symposium on Microarchitecture, pp.480 - 493 -
dc.identifier.url https://www.microarch.org/micro55/program/index.php#day1 -
dc.citation.conferencePlace US -
dc.citation.conferencePlace Chicago -
dc.citation.endPage 493 -
dc.citation.startPage 480 -
dc.citation.title IEEE/ACM International Symposium on Microarchitecture -
Files in This Item:

There are no files associated with this item.

Appears in Collections:
Department of Electrical Engineering and Computer Science Computer Architecture and Systems Lab 2. Conference Papers

qrcode

  • twitter
  • facebook
  • mendeley

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.

BROWSE