[Read Paper] Maxout Networks

版权声明:本文为博主原创文章,未经博主允许不得转载。 https://blog.csdn.net/Linoi/article/details/50506405

Title: Maxout Networks
Author: Ian J. Goodfellow, David Warde-Farley Mehdi Mirza et al.

摘要:
We consider the problem of designing models to leverage a recently introduced approximate model averaging technique called dropout. We define a simple new model called maxout (so named because its output is the max of a set of inputs, and because it is a natural companion to dropout) designed to both facilitate optimization by dropout and improve the accuracy of dropout’s fast approximate model averaging technique. We empirically verify that the model successfully accomplishes both of these tasks. We use maxout and dropout to demonstrate state of the art classification performance on four benchmark datasets: MNIST, CIFAR-10, CIFAR-100, and SVHN.

全文链接:http://arxiv.org/abs/1302.4389

Note:

The maxout model is simply a feed-forward achitecture, such as a multilayer perceptron or deep convolutional neural network, that uses a new type of activation function: the maxout unit.

Given an input x , a maxout hidden layer implements the function
这里写图片描述

猜你喜欢

转载自blog.csdn.net/Linoi/article/details/50506405