Deep Learning Overview-NQU
Deep Learning Overview-NQU
Nguyen Quang Uy
4/15/16
Outline
This presentation provides an introduction to the machine
learning and deep learning.
Concept of Machine Learning
Deep learning
3/11/16
3/11/16
New Data
Features
Extraction/Selection
Features
Extraction/Selection
Learning
Learnt Model
Learnt Model
Decision
a) Training phase
3/11/16
b) Testing/Deploying
5
Datasets
Often in the form of tables
Iris flower
3/11/16
sepal
width
petal
length
petal Class
width
5.1
3.5
1.4
0.2
Iris-setosa
7.0
3.2
4.7
1.4
Irisversicolor
6.3
3.3
6.0
2.5
Iris-virginica
..
3/11/16
Classification Problem
3/11/16
F1
5.1
7.0
6.3
2.1
1.0
F1
2.2
3.5
F2
3.5
3.2
3.3
1.3
1.8
F2
1.8
2.9
F3
1.4
4.7
6.0
6.2
2.5
F3
3.7
4.8
F4
0.2
1.4
2.5
5.7
2.6
Class
1
0
1
1
0
F4 Class
4.6 ?
5.2 ?
8
3/11/16
Calculate the distance between the new sample and all examples
in the training set
Select K-nearest examples to the new sample in the training set
Assign the new sample to the most common class among its Knearest neighbors
3/11/16
10
X1 (Acid durability)
(seconds)
X2 (Strengh)
(Kg/m2 )
Y=Classification
Bad
Bad
Good
Good
Please test the new paper tissue with X1=3, X2=7 with
K=3
3/11/16
11
X1 (Acid
durability)
(seconds)
X2 (Strengh)
(Kg/m2 )
Distance to the
new sample
Y=Classification
16
Bad
25
Bad
Good
13
Good
Please test the new paper tissue with X1=3, X2=7 with
K=3
Since K=3 we have two out of three closet samples are
good then the testing sample is good.
3/11/16
12
3/11/16
13
Performance Measure
Accuracy =
3/11/16
14
3/11/16
15
3/11/16
16
Robotics on Mars
3/11/16
17
Outline
This presentation provides an introduction to the machine
learning and deep learning.
Concept of Machine Learning
Deep learning
3/11/16
18
3/12/16
19
3/12/16
20
Activation function
The activation function defines the output of that node
given a set of inputs.
3/12/16
21
3/12/16
22
Nummerical Example
Neth1=0.15*0.05+0.2*0.1+0.35=0.3775
Outh1=1/(1+e-Neth1)=1/(1+e-0.3775)=0.596
3/12/16
23
Nummerical Example
Neth1=0.15*0.05+0.2*0.1+0.35=0.3775
Outh1=1/(1+e-Neth1)=1/(1+e-0.3775)=0.596
3/12/16
24
Multilayer Perceptron
3/12/16
25
Training MLP
3/12/16
26
Cost function
Cost function is the objective function for the model that we want to
find the parameters to optimize.
One popular cost function for neural network is cross-entroy cost
function:
J()=-[ylogh(x)+(1y)log(1h(x))]
where y is the objective value with the input is x
h(x) is the output of the model given x
3/12/16
27
Parameters Estimation
We select that:
J()=[ylogh(x)+(1y)log(1h(x))]
is minimal
3/12/16
28
J ( xk )
1.Compute a search direction pk =
2.Update xk+1 = xk - pk
3.Check for convergence (stopping criteria) e.g. df/dx = 0
4. k=k+1, repeat step 1 to 4.
3/12/16
29
3/12/16
30
3/12/16
31
Stopping criteria?
3/12/16
32
Initialization MLP
For biases:
For weights
6
a=
H k + H k1
3/12/16
33
How do we pick ?
t =
t =1
<
t=1
3/12/16
34
How do we pick ?
Decreased strategies:
1+ t
t =
1+t
t =
35
3/12/16
36
3/12/16
37
3/12/16
38
Outline
This presentation provides an introduction to the machine
learning and deep learning.
Concept of Machine Learning
Deep learning
3/11/16
39
3/12/16
40
3/12/16
41
Leading researchers
Lecun at Facebook
3/12/16
42
Successful application
3/12/16
43
3/12/16
44
A network with 1000 inputs two hidden layers with 500 nodes
and 10 output have more than 1000*500*500*10=2,500,000,000
parameters
Computationally expensive
Gradient decay quickly.
3/12/16
45
45
What is novelty of DL
1. What exactly is deep learning?
2. Why is it generally better than other methods on
image, speech and certain other types of data?
3/12/16
46
Network structure
Training algorithms
3/12/16
47
3/12/16
48
48
3/12/16
49
49
3/12/16
50
50
3/12/16
51
51
3/12/16
52
Numpy
Theano
3/12/16
53
Thank you!
3/11/16
54