2022
DOI: 10.1007/978-3-031-17433-9_16
|View full text |Cite
|
Sign up to set email alerts
|

TransNet: Shift Invariant Transformer Network for Side Channel Analysis

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
17
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(17 citation statements)
references
References 25 publications
0
17
0
Order By: Relevance
“…sensitive information) given a trace T. A discriminative model estimates a φ-parametric probability conditional distribution Pr[Y |T, φ] that is as similar as possible to the true unknown joint probability distribution Pr [Y |T]. This approach is beneficial for directly solving a classification problem without modeling unnecessary information and thus, mitigating the impact of some countermeasures such as the desynchronization effect [CDP17a,ZBHV19,Mag19,HSAM22].This reason leads the side-channel community to investigate the DL approaches to improve the profiled SCA [CDP17a, KPH + 19, ZBHV19, BPS + 20] relying on discriminative approach. While [WPP22] combines a DL dimensionality reduction method with template attacks as an alternative to the Principal Component Analysis [APSQ06], the Linear Discriminant Analysis [BGH + 15] or the Kernel Discriminant Analysis [CDP17b], all the end-to-end DLSCA models proposed in the state-of-the-art are based on the discriminative approach (e.g.…”
Section: Related Work and Limitations Of Discriminative Modelsmentioning
confidence: 99%
See 2 more Smart Citations
“…sensitive information) given a trace T. A discriminative model estimates a φ-parametric probability conditional distribution Pr[Y |T, φ] that is as similar as possible to the true unknown joint probability distribution Pr [Y |T]. This approach is beneficial for directly solving a classification problem without modeling unnecessary information and thus, mitigating the impact of some countermeasures such as the desynchronization effect [CDP17a,ZBHV19,Mag19,HSAM22].This reason leads the side-channel community to investigate the DL approaches to improve the profiled SCA [CDP17a, KPH + 19, ZBHV19, BPS + 20] relying on discriminative approach. While [WPP22] combines a DL dimensionality reduction method with template attacks as an alternative to the Principal Component Analysis [APSQ06], the Linear Discriminant Analysis [BGH + 15] or the Kernel Discriminant Analysis [CDP17b], all the end-to-end DLSCA models proposed in the state-of-the-art are based on the discriminative approach (e.g.…”
Section: Related Work and Limitations Of Discriminative Modelsmentioning
confidence: 99%
“…While [WPP22] combines a DL dimensionality reduction method with template attacks as an alternative to the Principal Component Analysis [APSQ06], the Linear Discriminant Analysis [BGH + 15] or the Kernel Discriminant Analysis [CDP17b], all the end-to-end DLSCA models proposed in the state-of-the-art are based on the discriminative approach (e.g. fully-connected neural networks [MZ13, MHM14, Wei20], ResNets [ZS19, JZHY20, GJS20, MS21], RNNs [LLY + 20], transformer neural network [HSAM22], attention mechanisms [LZC + 21]). However, due to the lack of theoretical results, the discriminative models can be seen as black-box tools, and the design of models can be a real challenge even against unprotected cryptographic implementations.…”
Section: Related Work and Limitations Of Discriminative Modelsmentioning
confidence: 99%
See 1 more Smart Citation
“…Commonly used network architectures including Multilayer Perceptron (MLP) [4], [6], [7], [28]- [30], Convolutional Neural Network (CNN) [6], [7], [10], [11], [20], [31]- [33], and the recently employed Transformer [34]. CNNs used in DL-PSCA typically consist of two components: a feature extractor comprising convolutional layers, and a classifier consisting of several fully-connected layers that integrate the extracted features and perform classification predictions.…”
Section: B Deep Learning-based Profiled Side-channel Analysismentioning
confidence: 99%
“…In 2017, the attention [8] mechanism was used in machine translation tasks and achieved good results, becoming a new generation of deep learning architecture. In 2022, TransNet [9] introduced the long-term attention mechanism [10] to side-channel attacks. TransNet can reduce the guessing entropy to below 1 using only 400 records in multiple desynchronized datasets and has achieved significant results on several datasets.…”
Section: Introductionmentioning
confidence: 99%