Title
Mintin: Maxout-Based And Input-Normalized Transformation Invariant Neural Network
Abstract
Convolutional Neural Network (CNN) is a powerful model for image classification, but it is insufficient to deal with the spatial variance of the input. This paper presents a Maxout-based and input-normalized transformation invariant neural network (MINTIN), which aims at addressing the nuisance variation of images and accumulating transformation invariance. We introduce an innovative module, the Normalization, and combine it with the Maxout operator. While the former focuses on each image itself, the latter pays attention to augmented versions of input, resulting in fully-utilized information. This combination, which can be inserted into existing CNN architectures, enables the network to learn invariance to rotation and scaling. While the authors of TI-POOLING acclaimed that they reached state-of-the-art results, ours reach a maximum decrease of 0.71%, 0.23% and 0.51% in error rate on MNIST-rot-12k, half-rotated MNIST and scaling MNIST, respectively. The size of the network is also significantly reduced, leading to high computational efficiency.
Year
Venue
Keywords
2018
2018 25TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP)
transformation invariant, Normalization, Maxout
Field
DocType
ISSN
MNIST database,Normalization (statistics),Pattern recognition,Invariant (physics),Convolutional neural network,Computer science,Network topology,Invariant (mathematics),Artificial intelligence,Artificial neural network,Contextual image classification
Conference
1522-4880
Citations 
PageRank 
References 
0
0.34
0
Authors
7
Name
Order
Citations
PageRank
Jingyang Zhang101.35
Kaige Jia251.89
Pengshuai Yang300.34
Fei Qiao49435.38
Qi Wei54920.68
Xin-Jun Liu63510.04
Huazhong Yang72239214.90