Abstract | ||
---|---|---|
Convolutional Neural Network (CNN) is a powerful model for image classification, but it is insufficient to deal with the spatial variance of the input. This paper presents a Maxout-based and input-normalized transformation invariant neural network (MINTIN), which aims at addressing the nuisance variation of images and accumulating transformation invariance. We introduce an innovative module, the Normalization, and combine it with the Maxout operator. While the former focuses on each image itself, the latter pays attention to augmented versions of input, resulting in fully-utilized information. This combination, which can be inserted into existing CNN architectures, enables the network to learn invariance to rotation and scaling. While the authors of TI-POOLING acclaimed that they reached state-of-the-art results, ours reach a maximum decrease of 0.71%, 0.23% and 0.51% in error rate on MNIST-rot-12k, half-rotated MNIST and scaling MNIST, respectively. The size of the network is also significantly reduced, leading to high computational efficiency. |
Year | Venue | Keywords |
---|---|---|
2018 | 2018 25TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP) | transformation invariant, Normalization, Maxout |
Field | DocType | ISSN |
MNIST database,Normalization (statistics),Pattern recognition,Invariant (physics),Convolutional neural network,Computer science,Network topology,Invariant (mathematics),Artificial intelligence,Artificial neural network,Contextual image classification | Conference | 1522-4880 |
Citations | PageRank | References |
0 | 0.34 | 0 |
Authors | ||
7 |
Name | Order | Citations | PageRank |
---|---|---|---|
Jingyang Zhang | 1 | 0 | 1.35 |
Kaige Jia | 2 | 5 | 1.89 |
Pengshuai Yang | 3 | 0 | 0.34 |
Fei Qiao | 4 | 94 | 35.38 |
Qi Wei | 5 | 49 | 20.68 |
Xin-Jun Liu | 6 | 35 | 10.04 |
Huazhong Yang | 7 | 2239 | 214.90 |