2023
DOI: 10.1007/s10915-023-02125-5
|View full text |Cite
|
Sign up to set email alerts
|

DCT-Former: Efficient Self-Attention with Discrete Cosine Transform

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
11
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 11 publications
(11 citation statements)
references
References 25 publications
0
11
0
Order By: Relevance
“…These include the Linformer (Wang et al, 2020 ), which approximates the self-attention matrix with a low-rank matrix, and the Reformer (Kitaev et al, 2020 ), which introduces locality-sensitive hashing to accelerate self-attention computation. DCT-Former (Scribano et al, 2023 ) achieves efficient self attention computation by introducing discrete cosine transform as a frequency domain based conversion method. By calculating attention weights in the frequency domain, DCT-Former can significantly reduce computational complexity while maintaining high performance, improving the efficiency and scalability of the model.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations
“…These include the Linformer (Wang et al, 2020 ), which approximates the self-attention matrix with a low-rank matrix, and the Reformer (Kitaev et al, 2020 ), which introduces locality-sensitive hashing to accelerate self-attention computation. DCT-Former (Scribano et al, 2023 ) achieves efficient self attention computation by introducing discrete cosine transform as a frequency domain based conversion method. By calculating attention weights in the frequency domain, DCT-Former can significantly reduce computational complexity while maintaining high performance, improving the efficiency and scalability of the model.…”
Section: Related Workmentioning
confidence: 99%
“…First, we performed the Kerformer model and the remaining five models [Performer (Choromanski et al, 2020 ), Reformer (Kitaev et al, 2020 ), and Liner Trans (Katharopoulos et al, 2020 ), Longformer (Beltagy et al, 2020 ), RFA (Peng et al, 2021 ), and Dct-former (Scribano et al, 2023 )] were compared in terms of accuracy. This was achieved by conducting comparative fine-tuning experiments on five datasets, including GLUE (QQP, SST-2, MNLI) (Wang et al, 2018 ), IMDB (Maas et al, 2011 ), and Amazon (Ni et al, 2019 ).…”
Section: Nlp Taskmentioning
confidence: 99%
See 2 more Smart Citations
“…While these methods are easy to implement, they are less adaptable and stable, and the denoised images are not sharp enough. Transform domain denoising methods [7] can effectively avoid image distortion by transforming the image into the frequency domain for processing, including Fourier transform [8], discrete cosine transform [9], and wavelet transform [10], but such methods usually have a high complexity and uncertainty. Moreover, deep-learning-based denoising methods enhance the denoising effect to some extent due to their strong feature representation capabilities, but they face higher training-set requirements and more time-consuming computational cost [11].…”
Section: Introductionmentioning
confidence: 99%