NNDL
NNDL
Neural
aud
Deep Learning
.
nates)
(sem
-z
uram
Models of a nevror .
model :
Deterministic
D Synapses/connecting links .
2) Addes .
lowered or increased
through bias .
of
Typesactivation functions
---
function
function) Heaside
1) Threshold
.
(unity amplificationn
a
linear function
1) Piecewise
function
3) sigmoid
.
parameter)
Fac [a
I = slope
6 () =
a
Memory
-
.
connected
memory and learning intricately
->
are
-> Thus
memory
ransforms an
activity pattern in the
into
input space another
activity pattern in the
output space .
[It is through matrix multiplication) Y =RX.
.
memory
.
Correlation
-
Matrimony .
M
M = YpX
= · Ya si
k=1
call normal
M ↑ onsider Ij to be
y =
Mxj
m vector (1) 11 jll=1
j
=
y =
I
&j]Y1
7
+Ej) yj
2L+x
3
=[c)Y;
=
+
Ej j
Failure in learning .
->
Training algo may
not find solution parameters
.
wrong function
->
due to
Training algo moose
may
over
fitting
.
[MM
-
->
Captures Spatial features from an
image .
Spatial features
identify object and it's
->
help the location
more
accurately
.
PN
->
intermediate results fed back to the
are
predict
of
outcome the .
layer
-> Information from
previous time-step is remembered by
a function
memory
.
--
-> with
> more to
more
capacity capture complex patterns
.
But
-> , more with may lead to
overfitting .
Depth
of a Wal etwork
-> No of,
layers in a MM
.
->
may face and exploding gradients problems
vanishing ,
Activation Functions
I ,
--
-
-
Used for
=
generative modelling and unsupervised learning
.
->
Statistical in .
nature
used
->
in supervised learning .
->
They have fixed weight .
-> Howevers ,
the weights are to be set
->
Arms capable of dense representations of input data ,
-> These dense representations are called 'latent representation's
or
'codings',
-> It works unsupervised learn to copy inputs to .
outputs
,
than the
->
Codings typically have a lower
dimensionality
,
input
->
They also act as feature detectors
Can be
-> used for unsupervised pretaining of nearal
networks
.
->
Morse can be added or Site of latent representations can ba
limited.
->
This forces the model to learn efficient of representing
ways
the data ,
-> It
always
has two ports: (i) Encoder (recognition network)
of op No of input
neurons in layer layers.
=
- No ,
, neurons in
under complete
-> Thus a full layer of nevrons using the same filter
,
the
Convolution will
During training layer automatically
-> learn
the most useful filters for its took and the layers above it
,
to
will know combine them into complex patterns
-> Each canvolation layer has multiple milters and outputs
-> Thus
,
a convolational layer applies multiple mainable filters to
-
> Valid Padding : No zero padding
. Each neuson's receptive
-
Same padding - Inputs are padded with enough zeroes on
Size even
for some pudding .
: Controls outpat gate
which decides
[Iv> Outputs O(P) ,
state
- long-term ,
too as long as
.
needed
(ii) Preserve it
GRU cells's
-
-> A simplified version of LSTM cell .
·
No olpgate .
Full state vector is outputted at
time step
.
every
However good LSTMs and RGRU cells are , they still can't