EdgeServe: Efficient Deep Learning Model Caching at the Edge

Tian Guo, Robert J. Walls, Samuel S. Ogden

Research output: Contribution to journalArticlepeer-review

Abstract

In this work, we look at how to effectively manage and utilize deep learning models at each edge location, to provide performance guarantees to inference requests. We identify challenges to use these deep learning models at resource-constrained edge locations, and propose to adapt existing cache algorithms to effectively manage these deep learning models.
Original languageAmerican English
JournalSEC '19: Proceedings of the 4th ACM/IEEE Symposium on Edge Computing
DOIs
StatePublished - 2019
Externally publishedYes

Keywords

  • Caching Algorithm
  • Deep Learning Inference
  • Edge Computing
  • Performance Optimization

Disciplines

  • Computer Sciences

Cite this