2023
DOI: 10.3390/ai4010004
|View full text |Cite
|
Sign up to set email alerts
|

End-to-End Transformer-Based Models in Textual-Based NLP

Abstract: Transformer architectures are highly expressive because they use self-attention mechanisms to encode long-range dependencies in the input sequences. In this paper, we present a literature review on Transformer-based (TB) models, providing a detailed overview of each model in comparison to the Transformer’s standard architecture. This survey focuses on TB models used in the field of Natural Language Processing (NLP) for textual-based tasks. We begin with an overview of the fundamental concepts at the heart of t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
19
0
1

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 51 publications
(20 citation statements)
references
References 238 publications
0
19
0
1
Order By: Relevance
“…In practical applications, encoding and decoding 3 models are commonly used for processing. The encoder encodes the input text into a vector representation, and the decoder converts this vector into an output title 4 , which show in Figure 1.…”
Section: Overview Of Text Title Generation and Abstract Extraction Mo...mentioning
confidence: 99%
“…In practical applications, encoding and decoding 3 models are commonly used for processing. The encoder encodes the input text into a vector representation, and the decoder converts this vector into an output title 4 , which show in Figure 1.…”
Section: Overview Of Text Title Generation and Abstract Extraction Mo...mentioning
confidence: 99%
“…В контексте обнаружения ВПО для этого используются такие методы, как Word2Vec [10], HMM2Vec [11], BERT [12] и ELMo [13]. BERT показал наибольшую эффективность в системах MalBERT [14] и её усовершенствованной версии MalBERTv2 [15]. Однако, в указанных выше системах используется подход статического анализа.…”
Section: Introductionunclassified
“…Transformers, a neural network architecture, have garnered considerable interest in both Natural Language Processing (NLP) and time series analysis. Their capacity to manage long-range dependencies and parallel processing has propelled their popularity in these domains (Rahali and Akhloufi, 2023). Utilizing self-attention or scaled dot-product attention, Transformers can assess the significance of inputs within a sequence, enabling the capture of intricate data patterns.…”
Section: ) Introductionmentioning
confidence: 99%