2020 IEEE International Symposium on Workload Characterization (IISWC) 2020
DOI: 10.1109/iiswc50251.2020.00024
|View full text |Cite
|
Sign up to set email alerts
|

Cross-Stack Workload Characterization of Deep Recommendation Systems

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
13
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
8
1

Relationship

5
4

Authors

Journals

citations
Cited by 22 publications
(13 citation statements)
references
References 38 publications
0
13
0
Order By: Relevance
“…For commensurate analysis, final results are presented based on the highest quality achieved for each model and dataset: NDCG of 92. 25 for Criteo (see Section 2). • Tail-latency: To maintain user-experience, recommendations must meet SLAs and be served under strict taillatency targets [16], measured as 99 th percentile (p99).…”
Section: Experimental Methodologymentioning
confidence: 99%
See 1 more Smart Citation
“…For commensurate analysis, final results are presented based on the highest quality achieved for each model and dataset: NDCG of 92. 25 for Criteo (see Section 2). • Tail-latency: To maintain user-experience, recommendations must meet SLAs and be served under strict taillatency targets [16], measured as 99 th percentile (p99).…”
Section: Experimental Methodologymentioning
confidence: 99%
“…Lots of research effort has been devoted to design specialized hardware for deep learning-especially MLPs, CNNs, and RNNs [4,6,7,8,17,19,30,42,43,44,47,54,55]. However, recommendation systems pose distinct challenges owing to their network architectures and use cases [18,25,41]. Given its importance, hardware proposals for accelerating recommendation models have begun to emerge [1,3,9,10,16,26,29,33,34,36,37,40,50].…”
Section: Related Workmentioning
confidence: 99%
“…It has also been adopted by the industry-wide MLPerf benchmark suites [43,44,49,56]. Recent works [23,24,29,48] presented in-depth system performance characterization studies and shared the architectural implications for additional optimization opportunities. Our work here demonstrates a hardware-software co-design in action on deploying production-scale recommender systems to lowprecision hardware architectures.…”
Section: Related Workmentioning
confidence: 99%
“…In particular, efficient training for the large embedding tables with varying memory access patterns imposes significant system design and optimization challenges. In addition, recent studies have started analyzing the system-and architecturelevel implications of neural recommendation inference [20], [21], [24]. Recent works also examine near memory processing architectures, such as RecNMP [29], TensorDIMM [30].…”
Section: Related Workmentioning
confidence: 99%