Quantifying the latency benefits of near-edge and in-network FPGA acceleration

Ryan A. Cooke, Suhaib A. Fahmy

Research output: Chapter in Book/Report/Conference proceedingConference contribution

2 Scopus citations

Abstract

Transmitting data to cloud datacenters in distributed IoT applications introduces significant communication latency, but is often the only feasible solution when source nodes are computationally limited. To address latency concerns, Cloudlets, in-network computing, and more capable edge nodes are all being explored as a way of moving processing capability towards the edge of the network. Hardware acceleration using Field programmable gate arrays (FPGAs) is also seeing increased interest due to reduced computation time and improved efficiency. This paper evaluates the the implications of these offloading approaches using a case study neural network based image classification application, quantifying both the computation and communication latency resulting from different platform choices. We demonstrate that emerging in-network accelerator approaches offer much improved and predictable performance as well as better scaling to support multiple data sources.
Original languageEnglish (US)
Title of host publicationEdgeSys 2020 - Proceedings of the 3rd ACM International Workshop on Edge Systems, Analytics and Networking, Part of EuroSys 2020
PublisherAssociation for Computing Machinery
Pages7-12
Number of pages6
ISBN (Print)9781450371322
DOIs
StatePublished - Apr 27 2020
Externally publishedYes

Fingerprint

Dive into the research topics of 'Quantifying the latency benefits of near-edge and in-network FPGA acceleration'. Together they form a unique fingerprint.

Cite this