RM-SSD: In-Storage Computing for Large-Scale Recommendation Inference
Journal
Proceedings - International Symposium on High-Performance Computer Architecture
Journal Volume
2022-April
Pages
1056-1070
Date Issued
2022
Author(s)
Abstract
To meet the strict service level agreement requirements of recommendation systems, the entire set of embeddings in recommendation systems needs to be loaded into the memory. However, as the model and dataset for production-scale recommendation systems scale up, the size of the embeddings is approaching the limit of memory capacity. Limited physical memory constrains the algorithms that can be trained and deployed, posing a severe challenge for deploying advanced recommendation systems. Recent studies offload the embedding lookups into SSDs, which targets the embedding-dominated recommendation models. This paper takes it one step further and proposes to offload the entire recommendation system into SSD with in-storage computing capability. The proposed SSD-side FPGA solution leverages a low-end FPGA to speed up both the embedding-dominated and MLP-dominated models with high resource efficiency. We evaluate the performance of the proposed solution with a prototype SSD. Results show that we can achieve 20-100× throughput improvement compared with the baseline SSD and 1.5-15× improvement compared with the state-of-art. © 2022 IEEE.
Subjects
n/a
Other Subjects
Embeddings; Field programmable gate arrays (FPGA); Computing capability; Embeddings; Large-scales; Lookups; Memory capacity; N/a; Physical memory; Production scale; Scale-up; Servicelevel agreement (SLA); Recommender systems
Type
conference paper