2022
DOI: 10.1016/j.cmpb.2022.107129
|View full text |Cite
|
Sign up to set email alerts
|

ISA-Net: Improved spatial attention network for PET-CT tumor segmentation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2023
2023
2025
2025

Publication Types

Select...
8
1

Relationship

0
9

Authors

Journals

citations
Cited by 15 publications
(7 citation statements)
references
References 15 publications
0
7
0
Order By: Relevance
“…Finally, data volume augmentation is used to improve the model’s robustness and avoid overfitting. Data augmentation methods including flip, rotations, random noise, scaling, elastic deformations, and the mixed data (mix-up) method ( 46 , 47 ) are commonly used in PET/CT tumor segmentation tasks. The mix-up ( 48 ) technique can increase data diversity, which generates randomly weighted combined image pairs based on training data pairs.…”
Section: Discussionmentioning
confidence: 99%
“…Finally, data volume augmentation is used to improve the model’s robustness and avoid overfitting. Data augmentation methods including flip, rotations, random noise, scaling, elastic deformations, and the mixed data (mix-up) method ( 46 , 47 ) are commonly used in PET/CT tumor segmentation tasks. The mix-up ( 48 ) technique can increase data diversity, which generates randomly weighted combined image pairs based on training data pairs.…”
Section: Discussionmentioning
confidence: 99%
“…In this section, a series of ablation studies on the proposed YOLOS and extensive comparative experiments (including RetinaNet [24], Faster R-CNN [26], Cascade R-CNN [27], YOLOv3 [46], YOLOx [47], Mask-CNN [34], Cascade Mask-R-CNN [48], YOLACT [35], Seg-Former [32], ISA-Net [33], PID-Net [36], YOLOv5 [18], YOLOv7 [19], YOLOv8 [20] and YOLOv9 [21]) for station building surface defect detection are conducted to evaluate the performance of the YOLOS with a true railroad station images captured by UAV. Besides, the UAV imagery datasets acquisition, training details, and evaluation criteria are all presented in this section.…”
Section: Methodsmentioning
confidence: 99%
“…Furthermore, U 2 -Net [29], one of the many U-Net structure-based variants [30,31], achieves competitive effects in multiple segmented datasets through capturing more contextual information from different scales. To improve the semantic segmentation effect and reduce the amount of calculation, SegFormer [32] and ISANet [33] designed a lightweight multi-layer perceptron and Self-Attention mechanism respectively, which significantly increased the receptive field of the encoder. Besides, Mask R-CNN [34] that extends Faster R-CNN [27] to incorporate an additional branch for predicting segmentation masks, realizes high-precision instance-level segmentation, but only has an inference speed of 5 FPS.…”
Section: Related Workmentioning
confidence: 99%
“…Drawing on the efficacy of self-attention, scholars have suggested incorporating spatial-wise attention alongside channel-wise attention within transformer blocks to capture interactions [29] , [30] , [31] . These dual attention schemes based on self-attention have demonstrated improved performance across medical image segmentation tasks [32] , [33] , [34] , [35] , [36] . Taking inspiration from the aforementioned research, we propose a novel transformer block that leverages spatial-wise and channel-wise attention mechanisms to enhance the capture of both spatial and channel information in 3D medical images.…”
Section: Introductionmentioning
confidence: 99%