2013 IEEE 24th International Conference on Application-Specific Systems, Architectures and Processors 2013
DOI: 10.1109/asap.2013.6567600
|View full text |Cite
|
Sign up to set email alerts
|

Correctly rounded architectures for Floating-Point multi-operand addition and dot-product computation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
9
0

Year Published

2016
2016
2024
2024

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 14 publications
(10 citation statements)
references
References 24 publications
1
9
0
Order By: Relevance
“…7.3), (Tenca, 2009), and is most likely done in hardware. This is in line with the literature on hardware dot products (Kim & Kim, 2009;Tao et al, 2013;Sohn & Swartzlander, 2016;Kaul et al, 2019), where either sorting or a search for the maximum exponent is performed. Furthermore, this experiment demonstrates that none of the additions are performed before aligning the significands relative to the largest exponent: if evaluated before the arguments are shifted right relative to the largest magnitude arguments' exponent (by having multiple alignment stages), any other sum would evaluate to 2 −24 + 2 −24 = 2 −23 , a value that then could be added exactly to the total sum as the least significand bits would not be lost in the alignment.…”
Section: Accuracy Of the Dot Productssupporting
confidence: 80%
See 1 more Smart Citation
“…7.3), (Tenca, 2009), and is most likely done in hardware. This is in line with the literature on hardware dot products (Kim & Kim, 2009;Tao et al, 2013;Sohn & Swartzlander, 2016;Kaul et al, 2019), where either sorting or a search for the maximum exponent is performed. Furthermore, this experiment demonstrates that none of the additions are performed before aligning the significands relative to the largest exponent: if evaluated before the arguments are shifted right relative to the largest magnitude arguments' exponent (by having multiple alignment stages), any other sum would evaluate to 2 −24 + 2 −24 = 2 −23 , a value that then could be added exactly to the total sum as the least significand bits would not be lost in the alignment.…”
Section: Accuracy Of the Dot Productssupporting
confidence: 80%
“…We can show that the lack of normalization causes the dot product in tensor cores-and most likely in any other similar architectures in which partial sums are not normalized (Kim & Kim, 2009;Tao et al, 2013;Sohn & Swartzlander, 2016;Kaul et al, 2019)-to behave non-monotonically. Let us consider (3) and set all the elements in the first column of B to 2 −24 and then c 11 to 1 − 2 −24 and 1 in turn.…”
Section: Monotonicity Of Dot Productmentioning
confidence: 94%
“…Redistribution subject to CCBY license use in interval arithmetic (see, for example, [24], [29]). Evaluating according to (2.1) is expensive, and although proposals have been made for implementing it in hardware (e.g., [5], [33]), it is not, to our knowledge, supported in commercial processors because of the hardware costs. However, manufacturers could implement something between (2.1) and (2.2) by using a little extra precision, perhaps the extended precisions defined in the IEEE standard [22] or the 80-bit registers on Intel processors.…”
Section: Introductionmentioning
confidence: 99%
“…Fused sums or sums of products have been studied before [5], [6], [7], [8], [9], [10], [11], [1]. However, only [6], [8], [1] are exact.…”
Section: B Related Work and Previous Implementationsmentioning
confidence: 99%
“…Fused sums or sums of products have been studied before [5], [6], [7], [8], [9], [10], [11], [1]. However, only [6], [8], [1] are exact. All the others either truncate the term summation or compress the smaller magnitude terms into sticky bits, which leads to inexact results in some cases of cancellation.…”
Section: B Related Work and Previous Implementationsmentioning
confidence: 99%