文章阅读: network in network

The main idea is to abstract more information from the receptive field.

As in CNN, filters from higher layers map to larger regions in original input. It generates a higher level concept by combining the lower level concepts from the layer below. Therefore, we argue that it would be beneficial to do a better abstraction on each local patch,before combining them into higher level concepts."

The method is to replace the GLM (generalized linear model) with a MLP (multilayer perceptron). Previously, CNN prefers to use linear filters followed by a nonlinear activation function. But this structure may impose some prior on the input data. In this article, the author use multilayer perceptron to extract feature on the local receptive field.

Maxout Network imposes the prior that instances of a latent concept lie within a convex set in the input space.

Given no priors about the distributions of the latent concepts, it is desirable to use a universal function approximator for feature extraction of the local patches, as it is capable of approximating more abstract representations of the latent concepts.

Another trick to prevent overfitting is, in this article, using global average pooling instead of fully connected layer.

One advantage of global average pooling over the fully connected layers is that itis more native to the convolution structure by enforcing correspondences between feature maps and categories. Thus the feature maps can be easily interpreted as categories confidence maps. Another advantages is that there is no parameter to optimize in the global average pooling thus overfitting is avoided at this layer. Futhermore, global average pooling sums out the spatial information, thus it is more robust to spatial translations of the input.
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值