index.html 36.5 KB
Newer Older
1

Y
Yu Yang 已提交
2 3 4 5
<html>
<head>
  <script type="text/x-mathjax-config">
  MathJax.Hub.Config({
Y
Yu Yang 已提交
6
    extensions: ["tex2jax.js", "TeX/AMSsymbols.js", "TeX/AMSmath.js"],
Y
Yu Yang 已提交
7 8
    jax: ["input/TeX", "output/HTML-CSS"],
    tex2jax: {
9 10
      inlineMath: [ ['$','$'] ],
      displayMath: [ ['$$','$$'] ],
Y
Yu Yang 已提交
11 12 13 14
      processEscapes: true
    },
    "HTML-CSS": { availableFonts: ["TeX"] }
  });
Y
Yi Wang 已提交
15 16
  </script>
  <script src="https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.0/MathJax.js" async></script>
Y
Yu Yang 已提交
17
  <script type="text/javascript" src="../.tools/theme/marked.js">
Y
Yu Yang 已提交
18 19
  </script>
  <link href="http://cdn.bootcss.com/highlight.js/9.9.0/styles/darcula.min.css" rel="stylesheet">
Y
Yi Wang 已提交
20
  <script src="http://cdn.bootcss.com/highlight.js/9.9.0/highlight.min.js"></script>
Y
Yu Yang 已提交
21
  <link href="http://cdn.bootcss.com/bootstrap/4.0.0-alpha.6/css/bootstrap.min.css" rel="stylesheet">
Y
Yu Yang 已提交
22
  <link href="https://cdn.jsdelivr.net/perfect-scrollbar/0.6.14/css/perfect-scrollbar.min.css" rel="stylesheet">
Y
Yu Yang 已提交
23
  <link href="../.tools/theme/github-markdown.css" rel='stylesheet'>
Y
Yu Yang 已提交
24 25
</head>
<style type="text/css" >
Y
Yu Yang 已提交
26 27 28 29 30 31
.markdown-body {
    box-sizing: border-box;
    min-width: 200px;
    max-width: 980px;
    margin: 0 auto;
    padding: 45px;
Y
Yu Yang 已提交
32 33 34 35
}
</style>


Y
Yu Yang 已提交
36
<body>
Y
Yu Yang 已提交
37

Y
Yu Yang 已提交
38
<div id="context" class="container-fluid markdown-body">
Y
Yu Yang 已提交
39 40 41 42
</div>

<!-- This block will be replaced by each markdown file content. Please do not change lines below.-->
<div id="markdown" style='display:none'>
C
choijulie 已提交
43
# Recognize Digits
Y
Yu Yang 已提交
44

H
Hao Wang 已提交
45
The source code of this tutorial is in [book/recognize_digits](https://github.com/PaddlePaddle/book/tree/develop/02.recognize_digits). For new users, please refer to [Running This Book](https://github.com/PaddlePaddle/book/blob/develop/README.md#running-the-book) .
Y
Yu Yang 已提交
46

H
Hao Wang 已提交
47 48
## Background
When we learning programming, the first program we write is generally to implement printing “Hello World”. But the tutorial of machine learning or deep learning for the beginner is usually handwriting recognition on the [MNIST](http://yann.lecun.com/exdb/mnist/) database. Because handwriting recognition is a typical classification problem, relatively simple and the MNIST dataset is complete. MNIST dataset as a simple computer vision dataset contains a series of pictures and corresponding labels of handwriting digits. The picture is a 28x28 pixel matrix, and the label corresponds to 10 numbers from 0 to 9. Each picture has been normalized in size and centered in the position.
Y
Yu Yang 已提交
49 50

<p align="center">
H
Hao Wang 已提交
51 52
<img src="https://github.com/PaddlePaddle/book/blob/develop/02.recognize_digits/image/mnist_example_image.png?raw=true" width="400"><br/>
Figure 1. Example of a MNIST picture
Y
Yu Yang 已提交
53 54
</p>

H
Hao Wang 已提交
55
MNIST dataset is created from [NIST](https://www.nist.gov/srd/nist-special-database-19) Special Database 3(SD-3) and Special Database 1(SD-1). Because SD-3 is labeled by stuff of US Census Bureau and SD-1 is labeled by US high school students, so SD-3 is clearer and easier to be recognized than SD-1. Yann LeCun et al. pick half of SD-1 and half of SD-3 as train dataset (60000 data) and test dataset (10000 data).250 annotators labeled the training set, thus guaranteed that there wasn't a complete overlap of annotators of training set and test set.
Y
Yu Yang 已提交
56

H
Hao Wang 已提交
57
MNIST attracts scholars to train model based on the dataset. In 1998, LeCun conducted experiments respectively using Single layer classifier, Multilayer Perceptron and Multilayer convolutional neural network LeNet, constantly decreasing the error on test dataset ( from 12% to 0.7%)\[[1](#References)\]。 In the process of research, LeCun, the pioneer in the field of deep learning, came up with Convolutional Neural Network, largely improving the performance of handwriting recognition. After that, researchers take a large number of experiments based on K-Nearest Neighbors algorithm\[[2](#References)\], SVM\[[3](#References)\], Neural Network\[[4-7](#References)\] and Boosting method\[[8](#References)\] and so on, with multiple pre-processing methods(like distortion removal, noise removal, and blurring) to upgrade accuracy of recognition.
Y
Yu Yang 已提交
58

H
Hao Wang 已提交
59
Convolutional Neural Network plays an important role in the field of deep learning now. From simple LeNet proposed by Yann LeCun in early days to model VGGNet, GoogleNet, ResNet and so on in the ImageNet competition (please refer to [Image Classification](https://github.com/PaddlePaddle/book/tree/develop/03.image_classification) tutorial ), we have gain a serious of great achievements with convolutional neural network in the field of image classification.
Y
Yu Yang 已提交
60 61 62



H
Hao Wang 已提交
63 64 65 66 67 68 69 70 71 72 73
In this tutorial, starting from simple Softmax regression model, we help you learn about handwriting recognition and introduce you how to upgrade model and how to use MLP and CNN to optimize recognition result.


## Exploration of Models

To train a classifier based on MNIST dataset, before the introduction of three basic image classification networks used in this tutorial, we first give some definitions:

- $X$ is the input: the MNIST image is a two-dimensional image of $28\times28$. For the calculation, we transform it into a $784$ dimensional vector, ie $X=\left ( x_0, x_1, \dots, x_{783} \right )$.

- $Y$ is the output: the output of the classifier is number (0-9), ie $Y=\left ( y_0, y_1, \dots, y_9 \right )$, and each dimension $y_i$ represents the probability of image classification as $i$th number.

X
xiaoting 已提交
74
- $Label$ is the actual label of the picture: $Label=\left ( l_0, l_1, \dots, l_9 \right ) $ is also 10 dimensions, but only one dimension represents 1, and the rest is 0. For example, if the number on an image is 2, its label is $(0,0,1,0, \dots, 0)$
Y
Yu Yang 已提交
75

C
choijulie 已提交
76
### Softmax Regression
Y
Yu Yang 已提交
77

H
Hao Wang 已提交
78
The simplest Softmax regression model is to get features with input layer passing through a fully connected layer and then compute and ouput probabilities of multiple classifications directly via Softmax function \[[9](#references)\].
Y
Yu Yang 已提交
79

H
Hao Wang 已提交
80
The data of the input layer $X$ is passed to the output layer. The input $X$ is multiplied by weights $W$ and then added to the bias $b$ to generate activations:
Y
Yu Yang 已提交
81

L
Luo Tao 已提交
82
$$ y_i = \text{softmax}(\sum_j W_{i,j}x_j + b_i) $$
Y
Yu Yang 已提交
83

C
choijulie 已提交
84
where $ \text{softmax}(x_i) = \frac{e^{x_i}}{\sum_j e^{x_j}} $
Y
Yu Yang 已提交
85

H
Hao Wang 已提交
86
Figure 2 is a network of softmax regression, in which weights are represented by blue lines, bias are represented by red lines, and +1 indicates that the bias is $1$.
Y
Yu Yang 已提交
87

H
Hao Wang 已提交
88 89 90 91
<p align="center">
<img src="https://github.com/PaddlePaddle/book/blob/develop/02.recognize_digits/image/softmax_regression.png?raw=true" width=400><br/>
Figure 2. Softmax regression network structure <br/>
</p>
Y
Yu Yang 已提交
92

H
Hao Wang 已提交
93 94 95 96 97
For an $N$-class classification problem with $N$ output nodes, Softmax normalizes the resulting $N$ dimensional vector so that each of its entries falls in the range $[0,1]\in {R}$, representing the probability that the sample belongs to a certain category. Here $y_i$ denotes the predicted probability that an image is of number $i$.

In the classification problem, we usually use cross-entropy loss, the formula is as follows:

$$  L_{cross-entropy}(label, y) = -\sum_i label_ilog(y_i) $$
Y
Yu Yang 已提交
98

Y
Yu Yang 已提交
99

Y
Yu Yang 已提交
100

C
choijulie 已提交
101
### Multilayer Perceptron
Y
Yu Yang 已提交
102

H
Hao Wang 已提交
103
The Softmax regression model uses the simplest two-layer neural network, which contains only the input layer and the output layer, so its performance is limited. In order to achieve better recognition, we consider adding several hidden layers \[[10](#references)\] between the input and output layer.
Y
Yu Yang 已提交
104

H
Hao Wang 已提交
105 106 107
1.In the first hidden layer, you can get $ H_1 = \phi(W_1X + b_1) $, where $\phi$ represents the activation function. And common functions are [sigmoid, tanh or ReLU](#common activation functions).
2.In the second hidden layer, you can get $ H_2 = \phi(W_2H_1 + b_2) $.
3.Finally, in the output layer, you can get $Y=\text{softmax}(W_3H_2 + b_3)$, that is the final classification result vector.
108

Y
Yu Yang 已提交
109

H
Hao Wang 已提交
110
Figure 3 is a network structure of a multi-layer perceptron, in which weights are represented by blue lines, bias are represented by red lines, and +1 indicates that the bias is $1$.
Y
Yu Yang 已提交
111

H
Hao Wang 已提交
112 113 114
<p align="center">
<img src="https://github.com/PaddlePaddle/book/blob/develop/02.recognize_digits/image/mlp.png?raw=true" width=500><br/>
Figure 3. Multilayer perceptron network structure <br/>
L
Luo Tao 已提交
115 116
</p>

C
choijulie 已提交
117
### Convolutional Neural Network
Y
Yu Yang 已提交
118

H
Hao Wang 已提交
119
In the multi-layer perceptron model, an image is expanded into a one-dimensional vector and input into the network, ignoring its position and structure information. And the convolutional neural network can better utilize the structure information of the image. [LeNet-5](http://yann.lecun.com/exdb/lenet/) is a relatively simple convolutional neural network. Figure 4 shows the structure: the input two-dimensional image, first through the two convolutional layers to the pooling layer, then through the fully connected layer, and finally using the softmax as the output layer. Below we mainly introduce the convolutional layer and the pooling layer.
L
Luo Tao 已提交
120

Y
Yu Yang 已提交
121
<p align="center">
H
Hao Wang 已提交
122 123
<img src="https://github.com/PaddlePaddle/book/blob/develop/02.recognize_digits/image/cnn.png?raw=true" width="600"><br/>
Figure 4. LeNet-5 convolutional neural network structure<br/>
Y
Yu Yang 已提交
124
</p>
Y
Yu Yang 已提交
125

H
Hao Wang 已提交
126
#### Convolutional Layer
L
Luo Tao 已提交
127

H
Hao Wang 已提交
128
Convolutional Layer is the core of convolutional neural network. The convolution we mentioned in image recognition is a two-dimensional convolution, that is, a discrete two-dimensional filter (also called a convolutional kernel) and a two-dimensional image for convoluting. In short, the two-dimensional filter slides to all positions on two-dimensional images and dot product is taken for this pixel and its domain pixel at each position. Convolution operations are widely used in the field of image processing. Different convolutional kernels can extract different features, such as edges, lines, and angles. In deep convolutional neural networks, low-level to complex image features can be extracted by convolution operation.
Y
Yu Yang 已提交
129

C
choijulie 已提交
130
<p align="center">
H
Hao Wang 已提交
131 132
<img src="https://github.com/PaddlePaddle/book/blob/develop/02.recognize_digits/image/conv_layer.png?raw=true" width='750'><br/>
Figure 5. Convolutional Layer Picture <br/>
C
choijulie 已提交
133
</p>
L
Luo Tao 已提交
134

H
Hao Wang 已提交
135 136 137 138 139 140 141 142
Figure 5 shows an example of the process of computing convolution with input image in size of $H=5, W=5, D=3$, ie $5 \times 5$ size of 3 channel (RGB, also known as depth) color image.

This example contains two (denoted by $K$) groups of convolutional kernels, i.e. $Filter W_0$ and $Filter W_1$ in the figure. In convolution calculation, different convolutional kernels are usually used for different input channels. In the example, each set of convolutional kernels contains ($D=3$) $3\times 3$ (indicated by $F \times F$) convolutional kernel. In addition, the stride of convolutional kernel in horizontal and vertical direction of image is 2 (indicated by $S$); Pad 1 (represented by $P$) 0 in the four directions of input image, that is, the input layer raw data in the figure is the blue part, and the gray part is expanded with 0 in the size of 1. The convolution operation yields a feature map of the size of $3 \times 3 \times 2$ (represented by $H_{o} \times W_{o} \times K$), which is a 2-channel feature map in size of $3 \times 3$, where $H_o$ is calculated as: $H_o = (H - F + 2 \times P)/S + 1$, so is $W_o$. And each pixel in the output feature map is the summation of the inner product of each set of filters and each feature of the input image, plus the bias $b_o$, the bias is usually shared by each output feature map. The last $-2$ in the output feature map $o[:,:,0]$ is calculated as shown in the lower right corner of Figure 5.

The convolutional kernel is a learnable parameter in the convolution operation. As shown in the example above, the parameter of each layer of convolution is $D \times F \times F \times K$. In the multi-layer perceptron model, neurons are usually fully connected therefore with a large number of parameters. There are fewer parameters in the convolutional layer, which is also determined by main features of the convolutional layer, namely local connections and shared weights.

- Local connection: Each neuron is connected to only one region of the input neuron, which is called Receptive Field. In the image convolution operation, that is, the neurons are locally connected in the spatial dimension (the plane in which the above examples H and W are located), but are fully connected in depth. For the two-dimensional image itself, the local pixels are strongly related. This local connection ensures that the learned filter makes the strongest response to local input features. The idea of local connection is also inspired by the structure of visual system in biology. The neurons in the visual cortex receive information locally.

X
xiaoting 已提交
143
- Weight sharing: The filters used to calculate neurons in the same deep slice are shared. For example, in Figure 5, the filter for each neuron calculated by $o[:,:,0]$ is the same, both are $W_0$, which can greatly reduce the parameters. The sharing weight is meaningful to a certain extent, for example, the bottom edge feature of the image is independent of the specific location of the feature in the graph. However, it is unintentional in some cases. For example, the input picture is a face, eyes and hair are in different positions. And to learn different features in different positions, please (refer to [Stanford University Open Class](http://cs231n.Github.io/convolutional-networks/)). Note that the weights are only shared for the neurons of the same depth slice. In the convolutional layer, multiple sets of convolutional kernels are usually used to extract different features, that is, the weights of neurons with different depth slices are not shared by the features with different depth slices. In addition, bias are shared by all neurons with the same depth.
Y
Yu Yang 已提交
144

H
Hao Wang 已提交
145 146 147 148 149
By introducing the calculation process of convolution and its features, convolution could be seen as a linear operation with shift-invariant, which is the same operation performed at each position of the image. The local connection and weight sharing of the convolutional layer greatly reduce the parameters that need to be learned, which helps with training larger convolutional neural networks.

For more information about convolution, please refer to [Reference Reading](http://ufldl.stanford.edu/wiki/index.php/Feature_extraction_using_convolution#Convolutions)。

### Pooling Layer
Y
Yu Yang 已提交
150

Y
Yu Yang 已提交
151
<p align="center">
H
Hao Wang 已提交
152 153
<img src="https://github.com/PaddlePaddle/book/blob/develop/02.recognize_digits/image/max_pooling.png?raw=true" width="400px"><br/>
Figure 6. Picture in pooling layer</br>
Y
Yu Yang 已提交
154
</p>
Y
Yu Yang 已提交
155

H
Hao Wang 已提交
156
Pooling is a form of nonlinear downsampling. The main functionality of this layer is to reduce computation by reducing the network parameters and to control the overfitting to some extent. Normally a pooling layer is added after the convolutional layer. Pooling includes maximum pooling, average pooling and so on. The largest pooling is to divide the input layer into different areas by non-overlapping rectangular boxes, and the maximum value of each rectangular box is taken as the output layer, as shown in Figure. 6.
C
choijulie 已提交
157

H
Hao Wang 已提交
158
For details about convolutional neural network, please refer to the tutorial of [Standford Online Course]( http://cs231n.github.io/convolutional-networks/ ), [Ufldl](http://ufldl.stanford.edu/wiki/index.php/Pooling) and [Image Classification]( https://github.com/PaddlePaddle/book/tree/develop/03.image_classification ).
C
choijulie 已提交
159

H
Hao Wang 已提交
160 161
<a name="common activation functions"></a>
### Common activation functions
Y
Yu Yang 已提交
162

C
choijulie 已提交
163
- Sigmoid activation function: $ f(x) = sigmoid(x) = \frac{1}{1+e^{-x}} $
Y
Yu Yang 已提交
164

C
choijulie 已提交
165
- Tanh activation function: $ f(x) = tanh(x) = \frac{e^x-e^{-x}}{e^x+e^{-x}} $
Y
Yu Yang 已提交
166

H
Hao Wang 已提交
167
In fact, the tanh function is only a sigmoid function with change of scale. The value of the sigmoid function is doubled and then shifted down by 1 unit: tanh(x) = 2sigmoid(2x) - 1 .
Y
Yu Yang 已提交
168

C
choijulie 已提交
169
- ReLU activation function: $ f(x) = max(0, x) $
Y
Yu Yang 已提交
170

H
Hao Wang 已提交
171
For details, please refer to [activation function in Wikipedia](https://en.wikipedia.org/wiki/Activation_function).
Y
Yu Yang 已提交
172

H
Hao Wang 已提交
173
## Dataset Preparation
Y
Yu Yang 已提交
174

H
Hao Wang 已提交
175
PaddlePaddle provides a module `paddle.dataset.mnist` that automatically loads [MNIST] (http://yann.lecun.com/exdb/mnist/) data in the API. The loaded data is located under `/home/username/.cache/paddle/dataset/mnist`:
Y
Yu Yang 已提交
176 177


H
Hao Wang 已提交
178 179 180 181 182 183
|    filename          |       note              |
|----------------------|-------------------------|
|train-images-idx3-ubyte|  train data picture, 60,000 data |
|train-labels-idx1-ubyte|  train data label, 60,000 data    |
|t10k-images-idx3-ubyte |  test data picture, 10,000 data  |
|t10k-labels-idx1-ubyte |  test data label, 10,000 data     |
Y
Yu Yang 已提交
184

W
Wang,Jeff 已提交
185 186
## Fluid API Overview

H
Hao Wang 已提交
187 188 189 190 191 192 193 194 195 196
The demo will use the latest [Fluid API](http://paddlepaddle.org/documentation/docs/en/1.2/api_cn/index_cn.html). Fluid API is the latest PaddlePaddle API. It simplifies model configuration without sacrificing performance.
We recommend using the Fluid API, which is easy to learn and use to help you complete your machine learning tasks quickly.

Here is an overview of several important concepts in the Fluid API:

1. `inference_program`: specifies how to get the inference function from the data input.
This is where the network flow is defined.

2. `train_program`: specifies how to get the `loss` function from `inference_program` and `tag value`.
This is where the loss calculation is specified.
W
Wang,Jeff 已提交
197

H
Hao Wang 已提交
198
3. `optimizer_func`: Specifies the function of the optimizer configuration. The optimizer is responsible for reducing losses and driving training. Paddle supports a number of different optimizers.
W
Wang,Jeff 已提交
199

H
Hao Wang 已提交
200
In the code examples below, we'll take a closer look at them.
W
Wang,Jeff 已提交
201

H
Hao Wang 已提交
202
## Configuration Instructions
Y
Yu Yang 已提交
203

H
Hao Wang 已提交
204
Load the Fluid API package for PaddlePaddle.
Y
Yu Yang 已提交
205 206

```python
207
from __future__ import print_function #load print of python3 into current version
H
Hao Wang 已提交
208 209 210 211 212
import os
from PIL import Image # load module of image processing
import matplotlib.pyplot as plt
import numpy
import paddle # load paddle module
213
import paddle.fluid as fluid
Y
Yu Yang 已提交
214 215
```

W
Wang,Jeff 已提交
216 217
### Program Functions Configuration

H
Hao Wang 已提交
218
We need to configure `inference_program` function. We want to use this program to show three different classifiers, each of which is defined as a Python function.
219
We need to input the image data into the classifier. Paddle provides a special layer `layer.data` for reading data.
H
Hao Wang 已提交
220
Let's create a data layer to read the image and connect it to the network of classification.
C
choijulie 已提交
221

H
Hao Wang 已提交
222
-Softmax regression: The results of the classification can be obtained only through a simple layer of simple fully connected layer with softmax as the activation function.
Y
Yu Yang 已提交
223 224

```python
225
def softmax_regression():
H
Hao Wang 已提交
226 227 228 229 230 231 232
    """
    Define softmax classifier:
    A fully connected layer with activation function  softmax
    Return:
    predict_image -- result of classification
    """
    # input original image data in size of 28*28*1
233
    img = fluid.layers.data(name='img', shape=[1, 28, 28], dtype='float32')
H
Hao Wang 已提交
234
    # With softmax as the fully connected layer of the activation function, the size of the output layer must be 10
W
Wang,Jeff 已提交
235
    predict = fluid.layers.fc(
H
Hao Wang 已提交
236
    input=img, size=10, act='softmax')
Y
Yu Yang 已提交
237 238
    return predict
```
C
choijulie 已提交
239

H
Hao Wang 已提交
240
-Multilayer Perceptron: The following code implements a multilayer perceptron with two hidden layers (that is, fully connected layers). The activation functions of the two hidden layers are all ReLU, and the activation function of the output layer is Softmax.
Y
Yu Yang 已提交
241 242

```python
243
def multilayer_perceptron():
H
Hao Wang 已提交
244 245 246 247 248 249 250 251 252
    """
    Define multilayer perceptron classifier:
    Multilayer perceptron with two hidden layers (fully connected layers)
    The activation function of the first two hidden layers uses ReLU, and the activation function of the output layer uses Softmax.

    Return:
    predict_image -- result of classification
    """
    # input raw image data in size of 28*28*1
253
    img = fluid.layers.data(name='img', shape=[1, 28, 28], dtype='float32')
H
Hao Wang 已提交
254
    # the first fully connected layer, whose activation function is ReLU
255
    hidden = fluid.layers.fc(input=img, size=200, act='relu')
H
Hao Wang 已提交
256
    # the second fully connected layer, whose activation function is ReLU
257
    hidden = fluid.layers.fc(input=hidden, size=200, act='relu')
H
Hao Wang 已提交
258
    # With softmax as the fully connected output layer of the activation function, the size of the output layer must be 10
259 260
    prediction = fluid.layers.fc(input=hidden, size=10, act='softmax')
    return prediction
Y
Yu Yang 已提交
261
```
C
choijulie 已提交
262

H
Hao Wang 已提交
263
-Convolutional neural network LeNet-5: The input two-dimensional image first passes through two convolutional layers to the pooling layer, then passes through the fully connected layer, and finally fully connection layer with softmax as activation function is used as output layer.
Y
Yu Yang 已提交
264 265

```python
266
def convolutional_neural_network():
H
Hao Wang 已提交
267 268 269 270 271 272 273 274
    """
    Define convolutional neural network classifier:
    The input 2D image passes through two convolution-pooling layers, using the fully connected layer with softmax as the output layer

    Return:
    predict -- result of classification
    """
    # input raw image data in size of 28*28*1
275
    img = fluid.layers.data(name='img', shape=[1, 28, 28], dtype='float32')
H
Hao Wang 已提交
276 277
    # the first convolution-pooling layer
    # Use 20 5*5 filters, the pooling size is 2, the pooling step is 2, and the activation function is Relu.
278
    conv_pool_1 = fluid.nets.simple_img_conv_pool(
Y
Yu Yang 已提交
279 280 281 282 283
        input=img,
        filter_size=5,
        num_filters=20,
        pool_size=2,
        pool_stride=2,
284 285
        act="relu")
    conv_pool_1 = fluid.layers.batch_norm(conv_pool_1)
H
Hao Wang 已提交
286 287
    # the second convolution-pooling layer
    # Use 20 5*5 filters, the pooling size is 2, the pooling step is 2, and the activation function is Relu.
288
    conv_pool_2 = fluid.nets.simple_img_conv_pool(
Y
Yu Yang 已提交
289 290 291 292 293
        input=conv_pool_1,
        filter_size=5,
        num_filters=50,
        pool_size=2,
        pool_stride=2,
294
        act="relu")
H
Hao Wang 已提交
295
    # With softmax as the fully connected output layer of the activation function, the size of the output layer must be 10
296 297
    prediction = fluid.layers.fc(input=conv_pool_2, size=10, act='softmax')
    return prediction
Y
Yu Yang 已提交
298 299
```

W
Wang,Jeff 已提交
300
#### Train Program Configuration
H
Hao Wang 已提交
301 302
Then we need to set train program `train_program` It firstly infers from classifier.
During the training, it will compute `avg_cost`.
303

H
Hao Wang 已提交
304
** Note:** train program should return an array. The first parameter returned must be `avg_cost`. The trainer uses it to compute gradient.
305

H
Hao Wang 已提交
306
Please write your code and then test results of different classifiers of  `softmax_regression`, `MLP` and `convolutional neural network`.
Y
Yu Yang 已提交
307

Y
Yi Wang 已提交
308
```python
309
def train_program():
H
Hao Wang 已提交
310 311 312 313 314 315 316 317 318 319
    """
    Configure train_program

    Return:
    predict -- result of classification
    avg_cost -- mean loss
    acc -- accuracy of classification

    """
    # label layer, called label, correspondent with label category of input picture
320
    label = fluid.layers.data(name='label', shape=[1], dtype='int64')
321

H
Hao Wang 已提交
322 323 324
    # predict = softmax_regression() # cancel note and run Softmax regression
    # predict = multilayer_perceptron() # cancel note and run multiple perceptron
    predict = convolutional_neural_network() # cancel note and run LeNet5 convolutional neural network
W
Wang,Jeff 已提交
325

H
Hao Wang 已提交
326
    # use class cross-entropy function to compute loss function between predict and label
327
    cost = fluid.layers.cross_entropy(input=predict, label=label)
H
Hao Wang 已提交
328
    # compute mean loss
329
    avg_cost = fluid.layers.mean(cost)
H
Hao Wang 已提交
330
    # compute accuracy of classification
331
    acc = fluid.layers.accuracy(input=predict, label=label)
H
Hao Wang 已提交
332
    return predict, [avg_cost, acc]
333

Y
Yu Yang 已提交
334 335
```

336 337
#### Optimizer Function Configuration

H
Hao Wang 已提交
338
`Adam optimizer`,`learning_rate` below are learning rate. Their size is associated with speed of network train convergence.
339 340 341 342 343 344

```python
def optimizer_program():
    return fluid.optimizer.Adam(learning_rate=0.001)
```

H
Hao Wang 已提交
345
### Data Feeders for dataset Configuration
Y
Yu Yang 已提交
346

H
Hao Wang 已提交
347
Next We start the training process. `Paddle.dataset.mnist.train()` and `paddle.dataset.mnist.test()` are respectively as train dataset and test dataset. These two functions respectively return a reader-- reader in PaddlePaddle is a Python function, which returns a Python yield generator when calling the reader.
Y
Yu Yang 已提交
348

H
Hao Wang 已提交
349
`Shuffle` below is a reader decorator, which receives a reader A and returns another reader B. Reader B read `buffer_size` train data into a buffer and then the data is disordered randomly and is output one by one.
Y
Yu Yang 已提交
350

H
Hao Wang 已提交
351
`Batch` is a special decorator. Its input is a reader and output is a batched reader. In PaddlePaddle, a reader yield a piece of data every time while batched reader yield a minibatch every time.
Y
Yu Yang 已提交
352

Q
qiaolongfei 已提交
353
```python
H
Hao Wang 已提交
354 355 356 357
# there are 64 data in a minibatch
BATCH_SIZE = 64

# read 500 data in train dataset, randomly disorder them and then transfer it into batched reader which yield 64 data each time.
358 359 360
train_reader = paddle.batch(
        paddle.reader.shuffle(
            paddle.dataset.mnist.train(), buf_size=500),
H
Hao Wang 已提交
361 362
        batch_size=BATCH_SIZE)
# read data in test dataset and yield 64 data every time
363
test_reader = paddle.batch(
H
Hao Wang 已提交
364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396
            paddle.dataset.mnist.test(), batch_size=BATCH_SIZE)
```

### create training process

Now we need to create a training process. We will use `train_program`, `place` and `optimizer` defined before, conclude test loss in the period of training iteration and training verification and save parameters of model for prediction.


#### Event Handler Configuration

We can call a handler function to supervise training process during training.
We display two `event_handler` programs here. Please freely update Jupyter Notebook and find the changes.

`Event_handler` is used to output training result during the train

```python
def event_handler(pass_id, batch_id, cost):
    # print the intermediate results of training, like
    # training iterations, number of batch, and loss function
    print("Pass %d, Batch %d, Cost %f" % (pass_id,batch_id, cost))
```

```python
from paddle.utils.plot import Ploter

train_prompt = "Train cost"
test_prompt = "Test cost"
cost_ploter = Ploter(train_prompt, test_prompt)

# visualize training process
def event_handler_plot(ploter_title, step, cost):
    cost_ploter.append(ploter_title, step, cost)
    cost_ploter.plot()
Q
qiaolongfei 已提交
397 398
```

H
Hao Wang 已提交
399 400 401 402
`event_handler_plot` can be visualized as follows:

![png](./image/train_and_test.png)

W
Wang,Jeff 已提交
403

H
Hao Wang 已提交
404 405 406 407 408 409 410 411
### Start training

Aftering adding `event_handler` and `data reader` we configured, we can start to train the model.
Set parameters for operation to configure data description.
`Feed_order` is used to map data directory to `train_program`
Create a `train_test` reflecting the loss during our training.

Define network structure:
C
choijulie 已提交
412

Y
Yi Wang 已提交
413
```python
H
Hao Wang 已提交
414 415
# the model is run on single CPU
use_cuda = False # If you want to use GPU, please set it True
416
place = fluid.CUDAPlace(0) if use_cuda else fluid.CPUPlace()
W
Wang,Jeff 已提交
417

H
Hao Wang 已提交
418 419
# call train_program to get prediction value and loss value,
prediction, [avg_loss, acc] = train_program()
L
Luo Tao 已提交
420

H
Hao Wang 已提交
421 422
# input original image data in size of 28*28*1
# label layer, called label, correspondent with label category of input picture.
X
xiaoting 已提交
423

H
Hao Wang 已提交
424
# It is informed that data in network consists of two parts. One is img value, the other is label value.
X
xiaoting 已提交
425
feeder = fluid.DataFeeder(feed_list=['img', 'label'], place=place)
W
Wang,Jeff 已提交
426

H
Hao Wang 已提交
427
# choose Adam optimizer
X
xiaoting 已提交
428
optimizer = optimizer_program()
H
Hao Wang 已提交
429 430
optimizer.minimize(avg_loss)
```
431

H
Hao Wang 已提交
432
Configure hyper parameter during the training:
W
Wang,Jeff 已提交
433 434

```python
H
Hao Wang 已提交
435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478

PASS_NUM = 5 #train 5 iterations
epochs = [epoch_id for epoch_id in range(PASS_NUM)]

# save parameters of model into save_dirname file
save_dirname = "recognize_digits.inference.model"
```


```python
def train_test(train_test_program,
                   train_test_feed, train_test_reader):

    # save classification accuracy into acc_set
    acc_set = []
    # save mean loss in avg_loss_set
    avg_loss_set = []
    # transfer each data which is the output of testing reader_yield into network to train
    for test_data in train_test_reader():
        acc_np, avg_loss_np = exe.run(
            program=train_test_program,
            feed=train_test_feed.feed(test_data),
            fetch_list=[acc, avg_loss])
        acc_set.append(float(acc_np))
        avg_loss_set.append(float(avg_loss_np))
    # get accuracy and loss value on the test data
    acc_val_mean = numpy.array(acc_set).mean()
    avg_loss_val_mean = numpy.array(avg_loss_set).mean()
    # return mean loss value and mean accuracy
    return avg_loss_val_mean, acc_val_mean
```

Create executor

```python
exe = fluid.Executor(place)
exe.run(fluid.default_startup_program())
```

Set up main_program and test_program:

```python
main_program = fluid.default_main_program()
test_program = fluid.default_main_program().clone(for_test=True)
479 480
```

H
Hao Wang 已提交
481 482
Start training:

L
Luo Tao 已提交
483

484 485


Q
qiaolongfei 已提交
486
```python
H
Hao Wang 已提交
487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519
lists = []
step = 0
for epoch_id in epochs:
    for step_id, data in enumerate(train_reader()):
        metrics = exe.run(main_program,
                          feed=feeder.feed(data),
                          fetch_list=[avg_loss, acc])
        if step % 100 == 0: # print a log for every 100 times of training
            print("Pass %d, Batch %d, Cost %f" % (step, epoch_id, metrics[0]))
            event_handler_plot(train_prompt, step, metrics[0])
        step += 1

    # test classification result of each epoch
    avg_loss_val, acc_val = train_test(train_test_program=test_program,
                                       train_test_reader=test_reader,
                                       train_test_feed=feeder)

    print("Test with Epoch %d, avg_cost: %s, acc: %s" %(epoch_id, avg_loss_val, acc_val))
    event_handler_plot(test_prompt, step, metrics[0])

    lists.append((epoch_id, avg_loss_val, acc_val))

    # save parameters of trained model for prediction
    if save_dirname is not None:
        fluid.io.save_inference_model(save_dirname,
                                      ["img"], [prediction], exe,
                                      model_filename=None,
                                      params_filename=None)

# Choose the best pass
best = sorted(lists, key=lambda list: float(list[1]))[0]
print('Best pass is %s, testing Avgcost is %s' % (best[0], best[1]))
print('The classification accuracy is %.2f%%' % (float(best[2]) * 100))
Y
Yu Yang 已提交
520 521
```

H
Hao Wang 已提交
522 523 524 525 526 527

The training process is completely automatic. The log printed in event_handler is like as follows.

Pass represents iterations of train. Batch represents times to train all data. cost represents loss value of current pass.

Compute the mean loss and accuracy of classification after an epoch.
Y
Yu Yang 已提交
528

529 530 531 532 533 534 535 536 537 538 539 540 541
```
Pass 0, Batch 0, Cost 0.125650
Pass 100, Batch 0, Cost 0.161387
Pass 200, Batch 0, Cost 0.040036
Pass 300, Batch 0, Cost 0.023391
Pass 400, Batch 0, Cost 0.005856
Pass 500, Batch 0, Cost 0.003315
Pass 600, Batch 0, Cost 0.009977
Pass 700, Batch 0, Cost 0.020959
Pass 800, Batch 0, Cost 0.105560
Pass 900, Batch 0, Cost 0.239809
Test with Epoch 0, avg_cost: 0.053097883707459624, acc: 0.9822850318471338
```
Y
Yu Yang 已提交
542

H
Hao Wang 已提交
543
Check prediction accuracy of the model after training. In the train with MNIST, generally classification accuracy of softmax regression model is about 92.34%, while that of multilayer perceptron is 97.66% and that of convolutional neural network is 99.20%.
Y
Yu Yang 已提交
544

C
choijulie 已提交
545

H
Hao Wang 已提交
546
## Deploy the Model
L
liaogang 已提交
547

H
Hao Wang 已提交
548
You can use trained model to classify handwriting pictures of digits. The program below shows how to use well-trained model to predict.
549

H
Hao Wang 已提交
550
### Generate input data to be inferred
551

H
Hao Wang 已提交
552
`infer_3.png` is an example picture of number 3. Transform it into a numpy to match feed data format
553 554


L
liaogang 已提交
555
```python
556
def load_image(file):
X
xiaoting 已提交
557
    # open the image file and covert to grayscale
558
    im = Image.open(file).convert('L')
X
xiaoting 已提交
559
    # adjust the input image to a 28*28 high quality image
560
    im = im.resize((28, 28), Image.ANTIALIAS)
X
xiaoting 已提交
561
    # convert img to numpy
H
Hao Wang 已提交
562
    im = numpy.array(im).reshape(1, 1, 28, 28).astype(numpy.float32)
X
xiaoting 已提交
563
    # normalize
564 565 566
    im = im / 255.0 * 2.0 - 1.0
    return im

567
cur_dir = os.getcwd()
H
Hao Wang 已提交
568
tensor_img = load_image(cur_dir + '/image/infer_3.png')
569
```
570

571
### Inference
572

H
Hao Wang 已提交
573 574 575
By configuring network and training parameters via `load_inference_model`, We can simply insert classifier defined before.


576

577
```python
H
Hao Wang 已提交
578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597
inference_scope = fluid.core.Scope()
with fluid.scope_guard(inference_scope):
    # use fluid.io.load_inference_model to get inference program desc,
    # feed_target_names is used to define variable name needed to be passed into network
    # fetch_targets define variable name to be fetched from network
    [inference_program, feed_target_names,
     fetch_targets] = fluid.io.load_inference_model(
     save_dirname, exe, None, None)

    # Make feed a dictionary {feed_target_name: feed_target_data}
    # The result will contain a data list corresponding to fetch_targets
    results = exe.run(inference_program,
                            feed={feed_target_names[0]: tensor_img},
                            fetch_list=fetch_targets)
    lab = numpy.argsort(results)

    # Print prediction result of  infer_3.png
    img=Image.open('image/infer_3.png')
    plt.imshow(img)
    print("Inference result of image/infer_3.png is: %d" % lab[0][0][-1])
L
liaogang 已提交
598 599
```

Y
Yu Yang 已提交
600

C
choijulie 已提交
601 602


H
Hao Wang 已提交
603 604 605 606
### Result

If successful, the inference result input is as follows:
`Inference result of image/infer_3.png is: 3` , which indicates that out network successfully recognize the picture!
C
choijulie 已提交
607

H
Hao Wang 已提交
608
## Summary
C
choijulie 已提交
609

H
Hao Wang 已提交
610
Softmax regression, multilayer perceptron and convolutional neural network are the most basic deep learning model, from which complex neural networks are all derivative, so these models are helpful for later learning. At the same time, we found that from simple softmax regression transform to slightly complex convolutional neural network, the accuracy of recognition on MNIST dataset largely increased, resulting from that convolution layer is featured with local connection and sharing weight. When study of new models later, hope you make a deep understand of the key upgrade of new model compared with original model. In addition, this tutorial also talks about the basic steps to build PaddlePadle model, from the code of dataprovider, build of network to training and prediction. Familiar with the work flow, you can use your own data, define your own network model and finish your training and prediction tasks.
Y
Yu Yang 已提交
611

H
Hao Wang 已提交
612
<a name="References"></a>
C
choijulie 已提交
613
## References
Y
Yu Yang 已提交
614 615

1. LeCun, Yann, Léon Bottou, Yoshua Bengio, and Patrick Haffner. ["Gradient-based learning applied to document recognition."](http://ieeexplore.ieee.org/abstract/document/726791/) Proceedings of the IEEE 86, no. 11 (1998): 2278-2324.
H
Hao Wang 已提交
616
2. Wejéus, Samuel. ["A Neural Network Approach to Arbitrary SymbolRecognition on Modern Smartphones."](http://www.diva-portal.org/smash/record.jsf?pid=diva2%3A753279&dswid=-434) (2014).
Y
Yu Yang 已提交
617 618 619 620 621 622 623
3. Decoste, Dennis, and Bernhard Schölkopf. ["Training invariant support vector machines."](http://link.springer.com/article/10.1023/A:1012454411458) Machine learning 46, no. 1-3 (2002): 161-190.
4. Simard, Patrice Y., David Steinkraus, and John C. Platt. ["Best Practices for Convolutional Neural Networks Applied to Visual Document Analysis."](http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.160.8494&rep=rep1&type=pdf) In ICDAR, vol. 3, pp. 958-962. 2003.
5. Salakhutdinov, Ruslan, and Geoffrey E. Hinton. ["Learning a Nonlinear Embedding by Preserving Class Neighbourhood Structure."](http://www.jmlr.org/proceedings/papers/v2/salakhutdinov07a/salakhutdinov07a.pdf) In AISTATS, vol. 11. 2007.
6. Cireşan, Dan Claudiu, Ueli Meier, Luca Maria Gambardella, and Jürgen Schmidhuber. ["Deep, big, simple neural nets for handwritten digit recognition."](http://www.mitpressjournals.org/doi/abs/10.1162/NECO_a_00052) Neural computation 22, no. 12 (2010): 3207-3220.
7. Deng, Li, Michael L. Seltzer, Dong Yu, Alex Acero, Abdel-rahman Mohamed, and Geoffrey E. Hinton. ["Binary coding of speech spectrograms using a deep auto-encoder."](http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.185.1908&rep=rep1&type=pdf) In Interspeech, pp. 1692-1695. 2010.
8. Kégl, Balázs, and Róbert Busa-Fekete. ["Boosting products of base classifiers."](http://dl.acm.org/citation.cfm?id=1553439) In Proceedings of the 26th Annual International Conference on Machine Learning, pp. 497-504. ACM, 2009.
9. Rosenblatt, Frank. ["The perceptron: A probabilistic model for information storage and organization in the brain."](http://psycnet.apa.org/journals/rev/65/6/386/) Psychological review 65, no. 6 (1958): 386.
624
10. Bishop, Christopher M. ["Pattern recognition."](http://users.isr.ist.utl.pt/~wurmd/Livros/school/Bishop%20-%20Pattern%20Recognition%20And%20Machine%20Learning%20-%20Springer%20%202006.pdf) Machine Learning 128 (2006): 1-58.
Y
Yu Yang 已提交
625 626

<br/>
X
xiaoting 已提交
627
<a rel="license" href="http://creativecommons.org/licenses/by-sa/4.0/"><img alt="知识共享许可协议" style="border-width:0" src="https://paddlepaddleimage.cdn.bcebos.com/bookimage/camo.png" /></a><br /><span xmlns:dct="http://purl.org/dc/terms/" href="http://purl.org/dc/dcmitype/Text" property="dct:title" rel="dct:type">This tutorial</span> is contributed by <a xmlns:cc="http://creativecommons.org/ns#" href="http://book.paddlepaddle.org" property="cc:attributionName" rel="cc:attributionURL">PaddlePaddle</a>, and licensed under a <a rel="license" href="http://creativecommons.org/licenses/by-sa/4.0/">Creative Commons Attribution-ShareAlike 4.0 International License</a>.
628

Y
Yu Yang 已提交
629 630 631 632 633 634 635
</div>
<!-- You can change the lines below now. -->

<script type="text/javascript">
marked.setOptions({
  renderer: new marked.Renderer(),
  gfm: true,
Y
Yu Yang 已提交
636 637 638
  breaks: false,
  smartypants: true,
  highlight: function(code, lang) {
Y
Yu Yang 已提交
639
    code = code.replace(/&amp;/g, "&")
Y
Yu Yang 已提交
640 641
    code = code.replace(/&gt;/g, ">")
    code = code.replace(/&lt;/g, "<")
642
    code = code.replace(/&nbsp;/g, " ")
Y
Yu Yang 已提交
643
    return hljs.highlightAuto(code, [lang]).value;
Y
Yu Yang 已提交
644 645 646
  }
});
document.getElementById("context").innerHTML = marked(
647
        document.getElementById("markdown").innerHTML)
Y
Yu Yang 已提交
648 649
</script>
</body>