2016 IEEE International Symposium on Workload Characterization (IISWC) 2016
DOI: 10.1109/iiswc.2016.7581275
|View full text |Cite
|
Sign up to set email alerts
|

Fathom: reference workloads for modern deep learning methods

Abstract: Deep learning has been popularized by its recent successes on challenging artificial intelligence problems. One of the reasons for its dominance is also an ongoing challenge: the need for immense amounts of computational power. Hardware architects have responded by proposing a wide array of promising ideas, but to date, the majority of the work has focused on specific algorithms in somewhat narrow application domains. While their specificity does not diminish these approaches, there is a clear need for more fl… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
97
1

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 136 publications
(98 citation statements)
references
References 31 publications
0
97
1
Order By: Relevance
“…In our cluster, commodity embedding, search and recommendation workloads have large training datasets and may exploit hundreds to thousands of workers to achieve high throughput on the huge training dataset. Notably, such extra large-scale workloads always have significant commercial impact on the company's business; however, they are often not included in DL workload benchmarks [16], [22]. We find that they are non-negligible: only 0.7% of all workloads have more than 128 cNodes; however, they consume more than 16% computation resource on our cluster.…”
Section: A Overview Of the Workloadsmentioning
confidence: 95%
See 2 more Smart Citations
“…In our cluster, commodity embedding, search and recommendation workloads have large training datasets and may exploit hundreds to thousands of workers to achieve high throughput on the huge training dataset. Notably, such extra large-scale workloads always have significant commercial impact on the company's business; however, they are often not included in DL workload benchmarks [16], [22]. We find that they are non-negligible: only 0.7% of all workloads have more than 128 cNodes; however, they consume more than 16% computation resource on our cluster.…”
Section: A Overview Of the Workloadsmentioning
confidence: 95%
“…Some other work aim to establish the performance benchmark [16], [18], [22], [50]. Fathom [16] establishes a set of reference implementation for eight archetypal DL jobs.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…DNN performance analysis and optimization Current publicly available benchmarks [3,8,19,55] for DNNs fo-cus on neural networks with FC, CNN, and RNN layers only.…”
Section: Related Workmentioning
confidence: 99%
“…Most of the existing AI benchmarks [8,27,9,7,26,10] are based on commercial scenarios. Deep500 [28] is a benchmarking framework aiming to evaluate high-performance deep learning.…”
Section: Introductionmentioning
confidence: 99%