2021
DOI: 10.1109/access.2021.3078371
|View full text |Cite
|
Sign up to set email alerts
|

AMSASeg: An Attention-Based Multi-Scale Atrous Convolutional Neural Network for Real-Time Object Segmentation From 3D Point Cloud

Abstract: Extracting meaningful information on objects varying scale and shape is a challenging task while obtaining distinctive features on small to large size objects to enhance overall object segmentation accuracy from 3D point cloud. To handle this challenge, we propose an attention-based multi-scale atrous convolutional neural network (AMSASeg) for object segmentation from 3D point cloud. Specifically, a backbone network consists of three modules: distinctive atrous spatial pyramid pooling (DASPP), FireModule, and … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5

Relationship

1
4

Authors

Journals

citations
Cited by 8 publications
(2 citation statements)
references
References 14 publications
0
2
0
Order By: Relevance
“…Further, to verify the hypothesis of effectiveness in modeling long-range dependencies, as the majority of CNNs adopt pyramid structure to compute multi-scale feature efficiently, we compared the proposed model (ConvMixer) with CNN-like pyramid architectures. Inspired by recent work in processing multi-scale information [52,53], in the tested CNNs conventional convolutions were replaced with either atrous convolutions (AtrousCNN) or Inception module [54] (InceptionV2CNN, InceptionACNN, Incep-tionBCNN). Specifically, an Inception-v2 version [26] was adopted for InceptionV2CNN, while an inspired Inception-ResNet version [46] was used in both InceptionACNN and InceptionBCNN.…”
Section: Discussionmentioning
confidence: 99%
“…Further, to verify the hypothesis of effectiveness in modeling long-range dependencies, as the majority of CNNs adopt pyramid structure to compute multi-scale feature efficiently, we compared the proposed model (ConvMixer) with CNN-like pyramid architectures. Inspired by recent work in processing multi-scale information [52,53], in the tested CNNs conventional convolutions were replaced with either atrous convolutions (AtrousCNN) or Inception module [54] (InceptionV2CNN, InceptionACNN, Incep-tionBCNN). Specifically, an Inception-v2 version [26] was adopted for InceptionV2CNN, while an inspired Inception-ResNet version [46] was used in both InceptionACNN and InceptionBCNN.…”
Section: Discussionmentioning
confidence: 99%
“…However, these methods typically use the original ASPP structure, which utilizes multiple branches to obtain rich feature information but also carries a significant amount of redundant information. The AMSASeg 16 method adds a regular convolution operation before each dilated convolution operation in each ASPP branch. The SAR‐U‐Net 17 method employs a three‐branch ASPP structure, added to the original five‐layer deep U‐Net network, while the SENetCount 18 method uses ASPP for skip connections to integrate low‐level information.…”
Section: Related Workmentioning
confidence: 99%