Latency‐aware adaptive micro‐batching techniques for streamed data compression on graphics processing units. (4th May 2020)
- Record Type:
- Journal Article
- Title:
- Latency‐aware adaptive micro‐batching techniques for streamed data compression on graphics processing units. (4th May 2020)
- Main Title:
- Latency‐aware adaptive micro‐batching techniques for streamed data compression on graphics processing units
- Authors:
- Stein, Charles M.
Rockenbach, Dinei A.
Griebler, Dalvan
Torquati, Massimo
Mencagli, Gabriele
Danelutto, Marco
Fernandes, Luiz G. - Other Names:
- Wang Zhibo guestEditor.
Jiang Lin guestEditor.
Suman Bilial guestEditor.
Wyrzykowski Roman guestEditor.
Szymanski Boleslaw K. guestEditor.
Bentes Cristiana Barbosa guestEditor.
França Felipe M.G. guestEditor.
Marzulo Leandro Augusto Justen guestEditor.
Mencagli Gabriele guestEditor.
Pilla Mauricio Lima guestEditor. - Abstract:
- Summary: Stream processing is a parallel paradigm used in many application domains. With the advance of graphics processing units (GPUs), their usage in stream processing applications has increased as well. The efficient utilization of GPU accelerators in streaming scenarios requires to batch input elements in microbatches, whose computation is offloaded on the GPU leveraging data parallelism within the same batch of data. Since data elements are continuously received based on the input speed, the bigger the microbatch size the higher the latency to completely buffer it and to start the processing on the device. Unfortunately, stream processing applications often have strict latency requirements that need to find the best size of the microbatches and to adapt it dynamically based on the workload conditions as well as according to the characteristics of the underlying device and network. In this work, we aim at implementing latency‐aware adaptive microbatching techniques and algorithms for streaming compression applications targeting GPUs. The evaluation is conducted using the Lempel‐Ziv‐Storer‐Szymanski compression application considering different input workloads. As a general result of our work, we noticed that algorithms with elastic adaptation factors respond better for stable workloads, while algorithms with narrower targets respond better for highly unbalanced workloads.
- Is Part Of:
- Concurrency and computation. Volume 33:Number 11(2021)
- Journal:
- Concurrency and computation
- Issue:
- Volume 33:Number 11(2021)
- Issue Display:
- Volume 33, Issue 11 (2021)
- Year:
- 2021
- Volume:
- 33
- Issue:
- 11
- Issue Sort Value:
- 2021-0033-0011-0000
- Page Start:
- n/a
- Page End:
- n/a
- Publication Date:
- 2020-05-04
- Subjects:
- data compression algorithms -- dynamic reconfiguration -- parallel programming -- service level objective -- stream parallelism -- stream processing
Parallel processing (Electronic computers) -- Periodicals
Parallel computers -- Periodicals
004.35 - Journal URLs:
- http://onlinelibrary.wiley.com/ ↗
- DOI:
- 10.1002/cpe.5786 ↗
- Languages:
- English
- ISSNs:
- 1532-0626
- Deposit Type:
- Legaldeposit
- View Content:
- Available online (eLD content is only available in our Reading Rooms) ↗
- Physical Locations:
- British Library DSC - 3405.622000
British Library DSC - BLDSS-3PM
British Library STI - ELD Digital store - Ingest File:
- 16900.xml