goingDeeperWithConvolutions: Difference between revisions
Line 4: | Line 4: | ||
= Related work = | = Related work = | ||
In | In 2013 [http://arxiv.org/pdf/1312.4400v3.pdf [2]] Lin et al.<ref> | ||
Min Lin, Qiang Chen and Shuicheng Yan. Network in Network | |||
</ref> pointed out that the convolution filter in CNN is a generalized linear model (GLM) for the underlying data patch and the level of abstraction is low with GLM. They suggested replacing GLM with a ”micro network” structure which is a general nonlinear function approximator. | |||
[[File:nin.tiff]] |
Revision as of 15:08, 20 October 2015
Introduction
In the last three years, due to the advances of deep learning and more concretely convolutional networks. [an introduction of CNN] , the quality of image recognition has increased dramatically. The error rates for ILSVRC competition dropped significantly year by year.[LSVRC] This paper proposed a new deep convolutional neural network architecture codenamed Inception. With the inception module and carefully crafted design researchers build a 22 layers deep network called Google Lenet, which uses 12X fewer parameters while being significantly more accurate than the winners of ILSVRC 2012. [1]
Related work
In 2013 [2] Lin et al.<ref> Min Lin, Qiang Chen and Shuicheng Yan. Network in Network </ref> pointed out that the convolution filter in CNN is a generalized linear model (GLM) for the underlying data patch and the level of abstraction is low with GLM. They suggested replacing GLM with a ”micro network” structure which is a general nonlinear function approximator. File:nin.tiff