2023
DOI: 10.48550/arxiv.2302.02041
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

REaLTabFormer: Generating Realistic Relational and Tabular Data using Transformers

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
6
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(6 citation statements)
references
References 0 publications
0
6
0
Order By: Relevance
“…Moreover, accurately learning and reproducing nonstandard distribution patterns is di cult and may yield generated samples that cannot appropriately represent the complexities inherent in the original data. Recent advancements in deep learning, particularly those centered on Transformer architectures, have demonstrated promising applications in handling tabular datasets [22]- [24]. A notable development involves the implementation of a transformer-based GAN for the generation of synthetic data in the text and sequence areas [25], [26].…”
Section: Discussionmentioning
confidence: 99%
“…Moreover, accurately learning and reproducing nonstandard distribution patterns is di cult and may yield generated samples that cannot appropriately represent the complexities inherent in the original data. Recent advancements in deep learning, particularly those centered on Transformer architectures, have demonstrated promising applications in handling tabular datasets [22]- [24]. A notable development involves the implementation of a transformer-based GAN for the generation of synthetic data in the text and sequence areas [25], [26].…”
Section: Discussionmentioning
confidence: 99%
“…More recent works use deep learning models to generate synthetic tabular data. Solatorio et al [32] proposed RE-alTabFormer, a transformer-based framework for data generation. REalTabFormer uses an autoregressive model based on the GPT-2 [51] to generate non-relational tabular data, as shown in the proposed architecture in Figure 3.…”
Section: Figure 1: General Data Augmentation Workflow For Imbalanced ...mentioning
confidence: 99%
“…Concerning REalTabFormer, we established the training parameters and set the batch size to 64 and 10 epochs. However, the training was interrupted before this number of epochs due to the Early Stopping with Q ′ δ condition, as described in the [32]. In the case of GReaT, we chose to train the model for a single epoch driven by computational constraints.…”
Section: A Experimental Environmentmentioning
confidence: 99%
See 2 more Smart Citations