The classification of various document image classes is considered an important step towards building a modern digital library or office automation system. Convolutional Neural Network (CNN) classifiers trained with backpropagation are considered to be the current state of the art model for this task. However, there are two major drawbacks for these classifiers: the huge computational power demand for training, and their very large number of weights. Previous successful attempts at learning document image features have been based on training very large CNNs. SqueezeNet is a CNN architecture that achieves accuracies comparable to other state of the art CNNs while containing up to 50 times less weights, but never before experimented on document image classification tasks. In this research we have taken a novel approach towards learning these document image features by training on a very small CNN network such as SqueezeNet. We show that an ImageNet pretrained SqueezeNet achieves an accuracy of approximately 75 percent over 10 classes on the Tobacco-3482 dataset, which is comparable to other state of the art CNN. We then visualize saliency maps of the gradient of our trained SqueezeNet's output to input, which shows that the network is able to learn meaningful features that are useful for document classification. Previous works in this field have made no emphasis on visualizing the learned document features. The importance of features such as the existence of handwritten text, document titles, text alignment and tabular structures in the extracted saliency maps, proves that the network does not overfit to redundant representations of the rather small Tobacco-3482 dataset, which contains only 3482 document images over 10 classes.
Early diagnosis of Alzheimer's is crucial to slow the progression of the disease. In this regard, there are many attempts to detect this disease at the early stages using AI techniques such as deep learning. We have proposed an explainable method to solve the early‐stage detection of Alzheimer's using transfer learning as a well‐known approach when there is not enough data. The employed transfer learning method is a combination of fine‐tuned ResNet‐50 and Inception‐V3 with Soft‐max and SVM classifiers using averaging. Moreover, local interpretable model‐agnostic explanations (LIME) are used to show the explainability of the proposed method. The AUC, accuracy, specificity, and sensitivity on structural MRI data of 100 MCI patients were 0.94, 87%, 92%, and 82%, respectively. Also, the LIME results were subjectively evaluated. The results showed the proposed method outperformed some related works. In addition, LIME technique make model more reliable to identify the parts involved in the patient's brain.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.