2018
DOI: 10.1049/iet-cvi.2018.5249
|View full text |Cite
|
Sign up to set email alerts
|

ILGNet: inception modules with connected local and global features for efficient image aesthetic quality classification using domain adaptation

Abstract: In this paper, we address a challenging problem of aesthetic image classification, which is to label an input image as high or low aesthetic quality. We take both the local and global features of images into consideration. A novel deep convolutional neural network named ILGNet is proposed, which combines both the Inception modules and an connected layer of both Local and Global features. The ILGnet is based on GoogLeNet. Thus, it is easy to use a pre-trained GoogLeNet for large-scale image classification probl… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
48
0
5

Year Published

2018
2018
2024
2024

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 55 publications
(53 citation statements)
references
References 50 publications
0
48
0
5
Order By: Relevance
“…Another important model used in both binary and multiclass aesthetic classification are neural networks. A fully connected network can be used with softmax activation in the last layer for classification algorithms [Ma et al 2017, Lu et al 2015, Talebi and Milanfar 2018, Jin et al 2016b].…”
Section: Learning Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Another important model used in both binary and multiclass aesthetic classification are neural networks. A fully connected network can be used with softmax activation in the last layer for classification algorithms [Ma et al 2017, Lu et al 2015, Talebi and Milanfar 2018, Jin et al 2016b].…”
Section: Learning Methodsmentioning
confidence: 99%
“…It is also possible to obtain generic features by using transfer learning on DCNNs that have been previously trained to classify objects, like Inception, VGG and MobileNet [Talebi and Milanfar 2018]. These transfer learned descriptors can also be used with additional convolutional layers [Jin et al 2016b].…”
Section: Aesthetic Descriptorsmentioning
confidence: 99%
“…Compared with previous methods, the classification accuracy is greatly improved. The accuracy we achieve on the AVA dataset is 81.68%, and the accuracy is up to 82.66% by using the Inception V4 module [1,2].…”
mentioning
confidence: 89%
“…Many well-known deep learning models, such as VGGNet [37], GoogLeNet [38], and Microsoft ResNet [39], have achieved excellent results in international classification recognition competitions, and they offer many ways to solve problems. In this study, we compare the proposed 2-CLSTM Model with VGGNet, GoogLeNet, and ResNet50 using the same input data set and Adam [40] optimizer.…”
Section: E Baseline Modelmentioning
confidence: 99%
“…The Flatten structure is used to connect the largest pool layer to the fully connected layer.Compared to VGGNet, GoogLeNet has a lighter structural design. Inception [38] is the core structure of GoogLeNet. This structure stacks convolution and pooling operations in CNN, which allows the model to increase both network width and network adaptability to scale.…”
Section: E Baseline Modelmentioning
confidence: 99%