2021
DOI: 10.48550/arxiv.2110.12894
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

The Efficiency Misnomer

Abstract: Model efficiency is a critical aspect of developing and deploying machine learning models. Inference time and latency directly affect the user experience, and some applications have hard requirements. In addition to inference costs, model training also have direct financial and environmental impacts. Although there are numerous well-established metrics (cost indicators) for measuring model efficiency, researchers and practitioners often assume that these metrics are correlated with each other and report only f… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
22
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
2

Relationship

1
8

Authors

Journals

citations
Cited by 16 publications
(22 citation statements)
references
References 49 publications
0
22
0
Order By: Relevance
“…Limitations: There is no perfect metric for measuring the overall performance of a given neural network architecture [14]. We have provided 4 different metrics but there are probably some aspects that are not considered.…”
Section: Discussionmentioning
confidence: 99%
“…Limitations: There is no perfect metric for measuring the overall performance of a given neural network architecture [14]. We have provided 4 different metrics but there are probably some aspects that are not considered.…”
Section: Discussionmentioning
confidence: 99%
“…Lastly, as one exception to general trends, the training latency of GAT on HPO-Metab is higher than that of SubGNN. Note that latency ignores the parallelism from large batch sizes (Dehghani et al, 2021). Our model can show relatively high latency since it requires full batch computation.…”
Section: Models For S2n Translated Graphsmentioning
confidence: 99%
“…We conduct experiments with four real-world datasets to evaluate the performance and efficiency of S2N translation. We measure the number of parameters, throughput (samples per second), and latency (seconds per forward pass) for efficiency (Dehghani et al, 2021). We demonstrate that models with S2N translation are more efficient than the existing approach without a significant performance drop.…”
Section: Introductionmentioning
confidence: 99%
“…This section provides a complementary study that compares ARTEMIS' complexity with the complexity of several approaches. Since no single measure is enough to assess the complexity of a model (Dehghani et al, 2021), we compare the different models on three metrics:…”
Section: Complexity and Efficiency Studymentioning
confidence: 99%