2023
DOI: 10.1109/tmm.2022.3147664
|View full text |Cite
|
Sign up to set email alerts
|

Tube-Embedded Transformer for Pixel Prediction

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(4 citation statements)
references
References 49 publications
0
4
0
Order By: Relevance
“…[53] a spatial information-guided convolution is proposed to enhance the spatial adaptability, which incorporates geometric information into the feature learning process by generating spatially adaptive convolutional weights. Zhang et al [71] propose a coarse-to-fine decoder to generate the feature based on the transformer mechanism, which achieves pixel-level multi-task prediction.…”
Section: B Rgb-d Semantic Segmentationmentioning
confidence: 99%
“…[53] a spatial information-guided convolution is proposed to enhance the spatial adaptability, which incorporates geometric information into the feature learning process by generating spatially adaptive convolutional weights. Zhang et al [71] propose a coarse-to-fine decoder to generate the feature based on the transformer mechanism, which achieves pixel-level multi-task prediction.…”
Section: B Rgb-d Semantic Segmentationmentioning
confidence: 99%
“…CaiT [15] introduces a class-attention layer which can separate the contradictory objectives of guiding attention. In addition, there are many works [16,17,18,19,20] focusing on the application of vision transformers.…”
Section: A Vision Transformersmentioning
confidence: 99%
“…17 There are many scenarios for multitask learning applications, such as pixel prediction, sentiment analysis, hotspot detection, and audio pattern recognition. 18–22…”
Section: Introductionmentioning
confidence: 99%
“…17 There are many scenarios for multitask learning applications, such as pixel prediction, sentiment analysis, hotspot detection, and audio pattern recognition. [18][19][20][21][22] This study implements a multiforward reaction prediction transformer (MFRPT) and retro-forward reaction prediction transformer (RFRPT) to demonstrate the feasibility of multitasking in chemical reaction prediction. Several classical lowresource datasets involving Baeyer-Villiger, Heck, and Chan-Lam reactions are used.…”
Section: Introductionmentioning
confidence: 99%