Abstract:Autoencoder reconstructions are widely used for the task of unsupervised anomaly localization. Indeed, an autoencoder trained on normal data is expected to only be able to reconstruct normal features of the data, allowing the segmentation of anomalous pixels in an image via a simple comparison between the image and its autoencoder reconstruction. In practice however, local defects added to a normal image can deteriorate the whole reconstruction, making this segmentation challenging. To tackle the issue, we pro… Show more
“…To get a segmentation map, we can directly impose a threshold on the reconstruction errors. Table 2 shows our defect segmentation performance in comparison with several recent works on defect segmentation [3,5,6,12,19,22,23,34]. We can observe that our method consistently performs competitively across all classes, achieving the best mean AUC among all the baselines.…”
Section: Comparison With Baselinesmentioning
confidence: 87%
“…Techniques for defect detection can be broadly grouped into: classification-based [11,24,32,35,36], detection-based [10,37], segmentation-based [7,10,17,20,28,[38][39][40], and reconstruction-based [3,5,6,23,25,34,41,47].…”
Section: Related Workmentioning
confidence: 99%
“…We adopted most of the baseline performance values from Bergmann et al [2], Dehaene et al [6], Huang et al [16], and Liu et al [22].…”
Section: B Baselines Implementation Detailsmentioning
confidence: 99%
“…The baselines AE-SSIM [3] and AE-L2 [2] follows the setting of Dehaene et al [6]. They used the network architecture of Bergmann et al [3] with a latent dimension of 100 and trained for 300 epochs with a learning rate of 1e −4 .…”
Section: B Baselines Implementation Detailsmentioning
In this paper, we propose a framework called Trust-MAE to address the problem of product defect classification. Instead of relying on defective images that are difficult to collect and laborious to label, our framework can accept datasets with unlabeled images. Moreover, unlike most anomaly detection methods, our approach is robust against noises, or defective images, in the training dataset. Our framework uses a memory-augmented auto-encoder with a sparse memory addressing scheme to avoid overgeneralizing the auto-encoder, and a novel trust-region memory updating scheme to keep the noises away from the memory slots. The result is a framework that can reconstruct defect-free images and identify the defective regions using a perceptual distance network. When compared against various state-of-the-art baselines, our approach performs competitively under noise-free MVTec datasets. More importantly, it remains effective at a noise level up to 40% while significantly outperforming other baselines.
“…To get a segmentation map, we can directly impose a threshold on the reconstruction errors. Table 2 shows our defect segmentation performance in comparison with several recent works on defect segmentation [3,5,6,12,19,22,23,34]. We can observe that our method consistently performs competitively across all classes, achieving the best mean AUC among all the baselines.…”
Section: Comparison With Baselinesmentioning
confidence: 87%
“…Techniques for defect detection can be broadly grouped into: classification-based [11,24,32,35,36], detection-based [10,37], segmentation-based [7,10,17,20,28,[38][39][40], and reconstruction-based [3,5,6,23,25,34,41,47].…”
Section: Related Workmentioning
confidence: 99%
“…We adopted most of the baseline performance values from Bergmann et al [2], Dehaene et al [6], Huang et al [16], and Liu et al [22].…”
Section: B Baselines Implementation Detailsmentioning
confidence: 99%
“…The baselines AE-SSIM [3] and AE-L2 [2] follows the setting of Dehaene et al [6]. They used the network architecture of Bergmann et al [3] with a latent dimension of 100 and trained for 300 epochs with a learning rate of 1e −4 .…”
Section: B Baselines Implementation Detailsmentioning
In this paper, we propose a framework called Trust-MAE to address the problem of product defect classification. Instead of relying on defective images that are difficult to collect and laborious to label, our framework can accept datasets with unlabeled images. Moreover, unlike most anomaly detection methods, our approach is robust against noises, or defective images, in the training dataset. Our framework uses a memory-augmented auto-encoder with a sparse memory addressing scheme to avoid overgeneralizing the auto-encoder, and a novel trust-region memory updating scheme to keep the noises away from the memory slots. The result is a framework that can reconstruct defect-free images and identify the defective regions using a perceptual distance network. When compared against various state-of-the-art baselines, our approach performs competitively under noise-free MVTec datasets. More importantly, it remains effective at a noise level up to 40% while significantly outperforming other baselines.
“…4 shows anomaly localisation results on MVTec AD images, where red regions in the heatmap indicate higher anomaly probability. From this results, we can see that our approach can localise anomalous regions of different sizes and structures from different object Metric Method Mean Accuracy AVID (Sabokrou et al 2018) 0.730 AESSIM (Bergmann et al 2018) 0.630 DAE (Hadsell et al 2006) 0.710 AnoGAN (Schlegl et al 2017) 0.550 λ-VAEu (Dehaene et al 2020) 0.770 LSA (Abati et al 2019) 0.730 CAVGA-Du (Venkataramanan et al 2019) (Bergmann et al 2018) 0.87 AVID (Sabokrou et al 2018) 0.78 SCADN (Yan et al 2021) 0.75 LSA (Abati et al 2019) 0.79 λ-VAEu (Dehaene et al 2020) 0.86 AnoGAN (Schlegl et al 2017) 0.74 ADVAE (Liu et al 2020) 0.86 CAVGA-Du (Venkataramanan et al 2019) 0.85 CAVGA-Ru (Venkataramanan et al 2019) 0.89 Ours -ImageNet 0.91 Ours -SSL 0.93…”
One-class classification (OCC) aims to learn an effective data description to enclose all normal training samples and detect anomalies based on the deviation from the data description. Current state-of-the-art OCC models learn a compact normality description by hyper-sphere minimisation, but they often suffer from overfitting the training data, especially when the training set is small or contaminated with anomalous samples. To address this issue, we introduce the interpolated Gaussian descriptor (IGD) method, a novel OCC model that learns a one-class Gaussian anomaly classifier trained with adversarially interpolated training samples. The Gaussian anomaly classifier differentiates the training samples based on their distance to the Gaussian centre and the standard deviation of these distances, offering the model a discriminability w.r.t. the given samples during training. The adversarial interpolation is enforced to consistently learn a smooth Gaussian descriptor, even when the training data is small or contaminated with anomalous samples. This enables our model to learn the data description based on the representative normal samples rather than fringe or anomalous samples, resulting in significantly improved normality description. In extensive experiments on diverse popular benchmarks, including MNIST, Fashion MNIST, CIFAR10, MVTec AD and two medical datasets, IGD achieves better detection accuracy than current state-of-the-art models. IGD also shows better robustness in problems with small or contaminated training sets.
Aiming at the problem that traditional surface detection is easily affected by complex industrial environments and cannot extract effective features, a deep learning-based knowledge distillation anomaly detection model is proposed. Firstly, a pre-trained teacher network was used to transfer knowledge of normal samples to the student network in the training phase. In the testing phase, defect detection was achieved based on the feature differences in the output of the teacher-student network for abnormal samples. Secondly, the attention mechanism module and the feature fusion module were added to the teacher network, which enhanced the detection ability of various defects of different sizes and increased the difference in output features between the teacher network and the student network. Finally, the defect image was located at the pixel level. Compared with the advanced knowledge distillation methods, the experimental results of the proposed model on the detection on MVTecAD and the anomaly localization on MVTecAD showed that the method in this paper improved to 90.3% and 90.1% on AUROC respectively, which verified the effectiveness of the method.
INDEX TERMSDefect detection; Knowledge distillation; Attention mechanism; Feature fusion QUNYING ZHOU received the B.S. degree from Huaiyin Normal University of mathematics in 2021. She is currently pursuing the M.E degree with ChangZhou University.Her main research interests include computer vision and defect detection.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.