2018
DOI: 10.1145/3145573
|View full text |Cite
|
Sign up to set email alerts
|

Simultaneous Conversions with the Residue Number System Using Linear Algebra

Abstract: We present an algorithm for simultaneous conversion between a given set of integers and their Residue Number System representations based on linear algebra. We provide a highly optimized implementation of the algorithm that exploits the computational features of modern processors. The main application of our algorithm is matrix multiplication over integers. Our speed-up of the conversions to and from the Residue Number System significantly improves the overall running time of matrix multiplication.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
3
1
1

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(5 citation statements)
references
References 37 publications
0
5
0
Order By: Relevance
“…In addition to the ideas for algorithmic improvements already noted in this paper, we point out that Arb would benefit from faster integer matrix multiplication in FLINT. More than a factor two can be gained with better residue conversion code and use of BLAS [3], [6]. BLAS could also be used for the radius matrix multiplications in Arb (we currently use simple C code since the FLINT multiplications are the bottleneck).…”
Section: Discussionmentioning
confidence: 99%
“…In addition to the ideas for algorithmic improvements already noted in this paper, we point out that Arb would benefit from faster integer matrix multiplication in FLINT. More than a factor two can be gained with better residue conversion code and use of BLAS [3], [6]. BLAS could also be used for the radius matrix multiplications in Arb (we currently use simple C code since the FLINT multiplications are the bottleneck).…”
Section: Discussionmentioning
confidence: 99%
“…There exist numerous important schemes: fast Fourier transforms [25], Karatsuba and Toom-Cook transforms [83,116], Chinese remaindering [26,61], Nussbaumer (or Schönhage-Strassen) polynomial transforms [99,113], truncated Fourier transforms [55], Chudnovsky 2 evaluations and interpolations [21], and so on.…”
Section: Algebra Complexitymentioning
confidence: 99%
“…• O(n ω d + n 2 d log(d)) if K supports FFT in degree 2d. We also mention a polynomial analogue of an integer matrix multiplication algorithm from [10] which uses evaluation/interpolation, done plainly via multiplication by (inverse) Vandermonde matrices. Then, the corresponding part of the cost (e.g.…”
Section: Polynomial Matrix Multiplicationmentioning
confidence: 99%
“…For FFT primes, we use evaluation/interpolation at roots of unity. For general primes, we use either evaluation/interpolation at geometric progressions (if such points exist in F p ), or our adaptation of the algorithm of [10], or 3-primes multiplication (as for polynomials, we lift the product from F p [x] to Z[x], where it is done modulo up to 3 FFT primes). No single variant outperformed or underperformed all others for all sizes and degrees, so thresholds were experimentally determined to switch between these options, with different values for small (less than 23 bits) and for large primes.…”
Section: Polynomial Matrix Multiplicationmentioning
confidence: 99%
See 1 more Smart Citation