SlideShare uma empresa Scribd logo
1 de 21
Baixar para ler offline
Parveen Malik
Assistant Professor
KIIT University
Neural Networks
Backpropagation
Background
โ€ข Perceptron Learning Algorithm , Hebbian Learning can classify input pattern if input
patterns are linearly separable.
โ€ข We need an algorithm which can train multilayer of perceptron or classify patterns
which are not linearly separable.
โ€ข Algorithm should also be able to use non-linear activation function.
๐’™๐Ÿ
๐’™๐Ÿ
๐‘ช๐’๐’‚๐’”๐’” ๐Ÿ
๐‘ช๐’๐’‚๐’”๐’” ๐Ÿ
๐‘ณ๐’Š๐’๐’†๐’‚๐’“๐’๐’š ๐‘บ๐’†๐’‘๐’†๐’“๐’‚๐’ƒ๐’๐’†
๐’™๐Ÿ
๐’™๐Ÿ
๐‘ช๐’๐’‚๐’”๐’” ๐Ÿ
๐‘ช๐’๐’‚๐’”๐’” ๐Ÿ
๐‘ณ๐’Š๐’๐’†๐’‚๐’“๐’๐’š ๐‘ต๐’๐’ โˆ’ ๐’”๐’†๐’‘๐’†๐’“๐’‚๐’ƒ๐’๐’†
โ€ข Need non-linear boundaries
โ€ข Perceptron algorithm can't be used
โ€ข Variation of GD rule is used.
โ€ข More layers are required
โ€ข Non-linear activation function required
Perceptron Algorithm โˆ’
๐‘พ๐’Š+๐Ÿ
๐‘ต๐’†๐’˜
= ๐‘พ๐’Š
๐‘ต๐’†๐’˜
+ (๐’• โˆ’ ๐’‚)๐’™๐’Š
Gradient Descent Algorithm-
๐‘พ๐’Š+๐Ÿ
๐‘ต๐’†๐’˜
= ๐‘พ๐’Š
๐’๐’๐’…
โˆ’ ๐œผ
๐๐‘ณ
๐๐’˜๐’Š
๐‘ณ๐’๐’”๐’” ๐’‡๐’–๐’๐’„๐’•๐’Š๐’๐’, ๐‘ณ =
๐Ÿ
๐Ÿ
๐’• โˆ’ ๐’‚ ๐Ÿ
Background- Back Propagation
โ€ข The perceptron learning rule of Frank Rosenblatt and the LMS algorithm of Bernard Widrow and
Marcian Hoff were designed to train single-layer perceptron-like networks.
โ€ข Single-layer networks suffer from the disadvantage that they are only able to solve linearly separable
classification problems. Both Rosenblatt and Widrow were aware of these limitations and proposed
multilayer networks that could overcome them, but they were not able to generalize their algorithms
to train these more powerful networks.
โ€ข First description of an algorithm to train multilayer networks was contained in the thesis of Paul
Werbos in 1974 .This thesis presented the algorithm in the context of general networks, with neural
networks as a special case, and was not disseminated in the neural network community.
โ€ข It was not until the mid 1980s that the backpropagation algorithm was rediscovered and widely
publicized. It was rediscovered independently by David Rumelhart, Geoffrey Hinton and Ronald
Williams 1986, David Parker 1985 and Yann Le Cun 1985.
โ€ข The algorithm was popularized by its inclusion in the book Parallel Distributed Processing [RuMc86],
which described the work of the Parallel Distributed Processing Group led by psychologists David
Rumelhart and James Mc-Clelland
โ€ข The multilayer perceptron, trained by the backpropagation algorithm, is currently the most widely
used neural network.
Network Design
Problem : Whether you watch a movie or not ?
Step 1 : Design โ€“ Output can be Yes (1) or No (0). Therefore one neuron or perceptron is
sufficient.
Step -2 : Choose suitable activation function in the output along with a rule to update the
weights. (Hard Limit function for perceptron learning algorithm, sigmoid for the
Widrow-Hoff rule or delta rule.)
๐‘พ๐’Š+๐Ÿ
๐‘ต๐’†๐’˜
= ๐‘พ๐’Š
๐’๐’๐’…
โˆ’ ๐œผ
๐๐‘ณ
๐๐’˜๐’Š
๐‘ณ =
๐Ÿ
๐Ÿ
๐’š โˆ’ เท
๐’š ๐Ÿ
=
๐Ÿ
๐Ÿ
๐’š โˆ’ ๐’‡ ๐’˜๐’™ + ๐’ƒ
๐Ÿ
๐๐‘ณ
๐๐’˜
= ๐Ÿ โˆ—
๐Ÿ
๐Ÿ
๐’š โˆ’ ๐’‡ ๐’˜๐’™ + ๐’ƒ
๐๐’‡ ๐’˜๐’™ + ๐’ƒ
๐๐’˜
= โˆ’ ๐’š โˆ’ เท
๐’š ๐’‡โ€ฒ ๐’˜๐’™ + ๐’ƒ ๐’™
๐‘ฅ เท ๐‘“
Director
or
Actor
or
Genre
or
IMDB
w
Yes (1)
or
No (0)
๐‘ค๐‘ฅ + ๐‘
เท
๐’š = ๐’‡ ๐’˜๐’™ + ๐’ƒ =
๐Ÿ
๐Ÿ + ๐’†โˆ’๐’˜๐’™+๐’ƒ
๐’‡ ๐’˜๐’™ + ๐’ƒ
๐‘ค0 = ๐‘
1
Network Design
Problem : Sort the students in the 4 house based on their three qualities like lineage,
choice and ethics ?
Step 1 : Design โ€“ Here, the input vector is 3-D i.e for each students, ๐‘†๐‘ก๐‘ข๐‘‘๐‘’๐‘›๐‘ก 1 =
๐ฟ1
๐ถ1
๐ธ1
,
๐‘†๐‘ก๐‘ข๐‘‘๐‘’๐‘›๐‘ก 2 =
๐ฟ2
๐ถ2
๐ธ2
๐’™๐Ÿ
๐’™๐Ÿ
๐’™๐Ÿ‘
N
๐’™๐Ÿ
๐’™๐Ÿ
๐’™๐Ÿ‘
๐’™๐ŸŽ=1
๐’˜๐Ÿ
๐’˜๐Ÿ
๐’˜๐Ÿ‘
๐’˜๐ŸŽ = ๐’ƒ
Yes (1)
or
No (0)
๐‘1
๐‘2
เท
๐’š๐Ÿ = ๐’‡ ๐’˜๐Ÿ๐Ÿ๐’™๐Ÿ + ๐’˜๐Ÿ๐Ÿ๐’™๐Ÿ + ๐’˜๐Ÿ๐Ÿ‘๐’™๐Ÿ‘ + ๐’ƒ๐Ÿ
๐’™๐Ÿ
๐’™๐Ÿ
๐’™๐Ÿ‘
๐’˜๐Ÿ๐Ÿ
๐’˜๐Ÿ๐Ÿ
๐’˜๐Ÿ๐Ÿ‘
๐’˜๐Ÿ๐Ÿ
๐’˜๐Ÿ๐Ÿ
๐’˜๐Ÿ๐Ÿ‘
เท
๐’š๐Ÿ = ๐’‡ ๐’˜๐Ÿ๐Ÿ๐’™๐Ÿ + ๐’˜๐Ÿ๐Ÿ๐’™๐Ÿ + ๐’˜๐Ÿ๐Ÿ‘๐’™๐Ÿ‘ + ๐’ƒ๐Ÿ
๐’ƒ๐Ÿ
๐’ƒ๐Ÿ
เทœ
๐‘ฆ1
เทœ
๐‘ฆ2
0
1
1
0
1
1
0
0
A B C D
Houses
๐‘ฆ1
๐‘ฆ2
Actual Output
Target Output
Network Design
Step 2 : Choosing the activation function and rule to update weights
Loss function, ๐ฟ =
1
2
๐‘ฆ โˆ’ เทœ
๐‘ฆ 2
1
1
๐‘1
๐‘2
เท
๐’š๐Ÿ = ๐’‡ ๐’˜๐Ÿ๐Ÿ๐’™๐Ÿ + ๐’˜๐Ÿ๐Ÿ๐’™๐Ÿ + ๐’˜๐Ÿ๐Ÿ‘๐’™๐Ÿ‘ + ๐’ƒ๐Ÿ
๐’™๐Ÿ
๐’™๐Ÿ
๐’™๐Ÿ‘
๐’˜๐Ÿ๐Ÿ
๐’˜๐Ÿ๐Ÿ
๐’˜๐Ÿ๐Ÿ‘
๐’˜๐Ÿ๐Ÿ
๐’˜๐Ÿ๐Ÿ
๐’˜๐Ÿ๐Ÿ‘
เท
๐’š๐Ÿ = ๐’‡ ๐’˜๐Ÿ๐Ÿ๐’™๐Ÿ + ๐’˜๐Ÿ๐Ÿ๐’™๐Ÿ + ๐’˜๐Ÿ๐Ÿ‘๐’™๐Ÿ‘ + ๐’ƒ๐Ÿ
๐’ƒ๐Ÿ
๐’ƒ๐Ÿ
เทœ
๐‘ฆ1
เทœ
๐‘ฆ2
0
1
1
0
0
0
A B C D
Houses
๐‘ฆ1
๐‘ฆ2
Actual Output
Target Output
๐‘พ๐’Š๐’‹ ๐’• + ๐Ÿ = ๐‘พ๐’Š๐’‹ ๐’• โˆ’ ๐œผ
๐๐‘ณ
๐๐’˜๐’Š๐’‹
๐๐‘ณ
๐๐’˜๐Ÿ๐Ÿ
= ๐’š โˆ’ เท
๐’š ๐’‡โ€ฒ ๐’˜๐Ÿ๐Ÿ๐’™๐Ÿ + +๐’˜๐Ÿ๐Ÿ๐’™๐Ÿ + ๐’˜๐Ÿ๐Ÿ‘๐’™๐Ÿ‘ + ๐’ƒ๐Ÿ ๐’™๐Ÿ
Network Architectures (Complex)
๐’™๐Ÿ
๐’‰๐Ÿ
๐’™๐’
๐’™๐Ÿ
๐’™๐’Š
๐’‰๐Ÿ
๐’‰๐’Ž
๐’‰๐’‹
๐’š๐Ÿ
๐’š๐Ÿ
๐’š๐’
๐’š๐’Œ
โ‹ฎ
Input Layer Hidden Layer Output Layer
๐‘พ(๐Ÿ)
๐‘พ(๐Ÿ)
โ‹ฎ
โ‹ฎ
โ‹ฎ
โ‹ฎ
โ‹ฎ
Network Architectures (More Complex)
๐‘ฅ1 ๐‘ฅ2 ๐‘ฅ2 ๐‘ฅ2
โ„Ž2
(1)
โ„Ž1
(1)
โ„Ž3
(1)
โ„Ž1
(2)
โ„Ž2
(2)
โ„Ž3
(2)
๐‘ฆ1 ๐‘ฆ2
๐‘พ(๐Ÿ)
๐‘พ(๐Ÿ)
๐‘พ(๐Ÿ‘)
Input
๐๐‘ณ
๐๐‘พ๐’Š๐’
= ๐œน๐’Š๐’๐’
๐œน๐’Š = ๐ˆโ€ฒ
๐’‚๐’Š เท
๐‘ฑ
๐œน๐’‹๐‘พ๐’‹๐’Š
๐‘Ž๐‘™ ๐‘ง๐‘™
๐‘Ž๐‘– ๐‘ง๐‘–
๐‘Ž๐‘— ๐‘ง๐‘—
โ‹ฎ
โ‹ฎ
โ‹ฎ
โ‹ฎ
โ‹ฎ
โ‹ฎ
๐‘พ๐’Š๐’
๐‘พ๐’‹๐’Š
๐œน๐’Š =
๐๐‘ณ
๐๐’‚๐’Š
๐œน๐’‹ =
๐๐‘ณ
๐๐’‚๐’‹
Cost Function
๐‘ณ =
๐Ÿ
๐Ÿ
(๐’š โˆ’ เท
๐’š) ๐Ÿ
Error to
input layer
๐ˆ ๐’‚๐’Š ๐Ÿ โˆ’ ๐ˆ ๐’‚๐’Š
Back-propagation Algorithm (Generalized Expression)
๐œน๐’‹ =
๐๐‘ณ
๐๐’‚๐’‹
= โˆ’ ๐’š โˆ’ เท
๐’š
๐เท
๐’š
๐๐’‚๐’‹
๐’‚๐’Š = เท
๐’
๐‘พ๐’Š๐’๐’๐’
๐œน๐’Š =
๐๐‘ณ
๐๐’‚๐’Š
= เท
๐‘ฑ
๐๐‘ณ
๐๐’‚๐’‹
๐๐’‚๐’‹
๐๐’‚๐’Š
๐๐’‚๐’‹
๐๐’‚๐’Š
=
๐๐’‚๐’‹
๐๐’๐’Š
๐๐’๐’Š
๐๐’‚๐’Š
= ๐‘พ๐’‹๐’Š๐ˆโ€ฒ ๐’‚๐’Š
Back-propagation Algorithm
๐’™๐Ÿ
๐’™๐Ÿ
เท
๐’š
๐‘Ž1 ๐œŽ ๐‘Ž1
๐‘Ž2 ๐œŽ ๐‘Ž2
๐‘1 ๐œŽ ๐‘1
1
1
1
0.6
0.4
-0.1
0.5
-0.3
0.3
0.4
0.1 -0.2
Back-propagation Algorithm
๐’™๐Ÿ
๐’™๐Ÿ
เท
๐’š
๐‘Ž1 ๐œŽ ๐‘Ž1
๐‘Ž2 ๐œŽ ๐‘Ž2
๐‘1 ๐œŽ ๐‘1
๐‘ป๐’‚๐’“๐’ˆ๐’†๐’•
๐’š = ๐Ÿ
1
1
1
0
1
0.6
0.4
-0.1
0.5
-0.3
0.3
0.4
0.1 -0.2
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
Back-propagation Algorithm
๐’™๐Ÿ
๐’™๐Ÿ
เท
๐’š
๐‘Ž1 ๐œŽ ๐‘Ž1
๐‘Ž2 ๐œŽ ๐‘Ž2
๐‘1 ๐œŽ ๐‘1
๐‘ป๐’‚๐’“๐’ˆ๐’†๐’•
๐’š = ๐Ÿ
1
1
1
0
1
0.6
0.4
-0.1
0.5
-0.3
0.3
0.4
0.1 -0.2
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘ณ =
๐Ÿ
๐Ÿ
๐’š โˆ’ เท
๐’š ๐Ÿ
Loss function
Back-propagation Algorithm
๐’™๐Ÿ
๐’™๐Ÿ
เท
๐’š
๐‘Ž1 ๐œŽ ๐‘Ž1
๐‘Ž2 ๐œŽ ๐‘Ž2
๐‘1 ๐œŽ ๐‘1
๐‘ป๐’‚๐’“๐’ˆ๐’†๐’•
๐’š = ๐Ÿ
1
1
1
0
1
0.6
0.4
-0.1
0.5
-0.3
0.3
0.4
0.1 -0.2
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘ณ =
๐Ÿ
๐Ÿ
๐’š โˆ’ เท
๐’š ๐Ÿ
Loss function
๐‘พ๐‘ต๐’†๐’˜
= ๐‘พ๐’๐’๐’…
โˆ’ ๐œผ
๐๐‘ณ
๐๐‘พ๐’๐’๐’…
Back-propagation Algorithm
Step 1 : Forward pass
๐’™๐Ÿ
๐’™๐Ÿ
๐‘Ž1 ๐œŽ ๐‘Ž1
๐’‚๐Ÿ=0.2
๐‘Ž2 ๐œŽ ๐‘Ž2
๐‘1 ๐œŽ ๐‘1
๐‘ป๐’‚๐’“๐’ˆ๐’†๐’•
๐’š = ๐Ÿ
1
1
1
0
1
0.6
0.4
-0.1
0.5
-0.3
0.3
0.4
0.1 -0.2
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
๐‘Š
12
(1)
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘ณ =
๐Ÿ
๐Ÿ
๐’š โˆ’ เท
๐’š ๐Ÿ
Loss function
๐ˆ ๐’‚๐Ÿ =
๐Ÿ
๐Ÿ + ๐’†โˆ’๐ŸŽ.๐Ÿ
= ๐ŸŽ. ๐Ÿ“๐Ÿ’๐Ÿ—๐Ÿ–
๐‘พ๐‘ต๐’†๐’˜ = ๐‘พ๐’๐’๐’… โˆ’ ๐œผ
๐๐‘ณ
๐๐‘พ๐’๐’๐’…
Back-propagation Algorithm
Step 1 : Forward pass
๐’™๐Ÿ
๐’™๐Ÿ
๐‘Ž1 ๐œŽ ๐‘Ž1
๐’‚๐Ÿ=0.2
๐’‚๐Ÿ=0.9
๐‘Ž2 ๐œŽ ๐‘Ž2
๐‘1 ๐œŽ ๐‘1
๐‘ป๐’‚๐’“๐’ˆ๐’†๐’•
๐’š = ๐Ÿ
1
1
1
0
1
0.6
0.4
-0.1
0.5
-0.3
0.3
0.4
0.1 -0.2
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘ณ =
๐Ÿ
๐Ÿ
๐’š โˆ’ เท
๐’š ๐Ÿ
Loss function
๐ˆ ๐’‚๐Ÿ =
๐Ÿ
๐Ÿ + ๐’†โˆ’๐ŸŽ.๐Ÿ
= ๐ŸŽ. ๐Ÿ“๐Ÿ’๐Ÿ—๐Ÿ–
๐ˆ ๐’‚๐Ÿ =
๐Ÿ
๐Ÿ + ๐’†โˆ’๐ŸŽ.๐Ÿ—
= ๐ŸŽ. ๐Ÿ•๐Ÿ๐ŸŽ๐Ÿ—
เท
๐’š
๐‘พ๐‘ต๐’†๐’˜
= ๐‘พ๐’๐’๐’…
โˆ’ ๐œผ
๐๐‘ณ
๐๐‘พ๐’๐’๐’…
Back-propagation Algorithm
Step 1 : Forward pass
๐’™๐Ÿ
๐’™๐Ÿ
๐‘Ž1 ๐œŽ ๐‘Ž1
๐’‚๐Ÿ=0.2
๐’‚๐Ÿ=0.9
๐’ƒ๐Ÿ=0.09101
๐‘Ž2 ๐œŽ ๐‘Ž2
๐‘1 ๐œŽ ๐‘1
เท
๐’š = ๐ˆ ๐’ƒ๐Ÿ =0.5227
๐‘ป๐’‚๐’“๐’ˆ๐’†๐’•
๐’š = ๐Ÿ
1
1
1
0
1
0.6
0.4
-0.1
0.5
-0.3
0.3
0.4
0.1 -0.2
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘ณ =
๐Ÿ
๐Ÿ
๐’š โˆ’ เท
๐’š ๐Ÿ
Loss function
๐ˆ ๐’‚๐Ÿ =
๐Ÿ
๐Ÿ + ๐’†โˆ’๐ŸŽ.๐Ÿ
= ๐ŸŽ. ๐Ÿ“๐Ÿ’๐Ÿ—๐Ÿ–
๐ˆ ๐’‚๐Ÿ =
๐Ÿ
๐Ÿ + ๐’†โˆ’๐ŸŽ.๐Ÿ—
= ๐ŸŽ. ๐Ÿ•๐Ÿ๐ŸŽ๐Ÿ—
Back-propagation Algorithm
Step 2 : Backpropagation of error
๐’™๐Ÿ
๐’™๐Ÿ
๐‘Ž1 ๐œŽ ๐‘Ž1
๐’‚๐Ÿ=0.2
๐’‚๐Ÿ=0.9
๐’ƒ๐Ÿ=0.09101
๐‘Ž2 ๐œŽ ๐‘Ž2
๐‘1 ๐œŽ ๐‘1
เท
๐’š = ๐ˆ ๐’ƒ๐Ÿ =0.5227
๐‘ป๐’‚๐’“๐’ˆ๐’†๐’•
๐’š = ๐Ÿ
1
1
1
0
1
0.6
0.4
-0.1
0.5
-0.3
0.3
0.4
0.1 -0.2
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘ณ =
๐Ÿ
๐Ÿ
๐’š โˆ’ เท
๐’š ๐Ÿ
Loss function
๐ˆ ๐’‚๐Ÿ =
๐Ÿ
๐Ÿ + ๐’†โˆ’๐ŸŽ.๐Ÿ
= ๐ŸŽ. ๐Ÿ“๐Ÿ’๐Ÿ—๐Ÿ–
๐ˆ ๐’‚๐Ÿ =
๐Ÿ
๐Ÿ + ๐’†โˆ’๐ŸŽ.๐Ÿ—
= ๐ŸŽ. ๐Ÿ•๐Ÿ๐ŸŽ๐Ÿ—
๐‘Ž๐‘™ ๐‘ง๐‘™
๐‘Ž๐‘– ๐‘ง๐‘–
๐‘Ž๐‘— ๐‘ง๐‘—
โ‹ฎ
โ‹ฎ
โ‹ฎ
โ‹ฎ
โ‹ฎ
โ‹ฎ
๐๐‘ณ
๐๐‘พ๐’Š๐’
= ๐œน๐’Š๐’๐’
๐œน๐’Š = ๐ˆโ€ฒ
๐’‚๐’Š เท
๐‘ฑ
๐œน๐’‹๐‘พ๐’‹๐’Š
๐‘พ๐’Š๐’
๐‘พ๐’‹๐’Š
Imagine
Back-propagation Algorithm
๐๐‘ณ
๐๐‘พ๐’Š๐’
= ๐œน๐’Š๐’๐’
๐œน๐’Š = ๐ˆโ€ฒ
๐’‚๐’Š เท
๐‘ฑ
๐œน๐’‹๐‘พ๐’‹๐’Š
๐‘Ž๐‘™ ๐‘ง๐‘™
๐‘Ž๐‘– ๐‘ง๐‘–
๐‘Ž๐‘— ๐‘ง๐‘—
โ‹ฎ
โ‹ฎ
โ‹ฎ
โ‹ฎ
โ‹ฎ
โ‹ฎ
๐‘พ๐’Š๐’ ๐‘พ๐’‹๐’Š
๐œน๐’Š=
๐๐‘ณ
๐๐’‚๐’Š
๐œน๐’‹=
๐๐‘ณ
๐๐’‚๐’‹
๐ˆโ€ฒ
๐’‚๐’Š = ๐ˆ ๐’‚๐’Š ๐Ÿ โˆ’ ๐ˆ ๐’‚๐’Š
Cost/Error Function
๐‘ณ =
๐Ÿ
๐Ÿ
(๐’š โˆ’ เท
๐’š) ๐Ÿ
๐’›๐’Š = ๐ˆ ๐’‚๐’Š
๐œน๐Ÿ = ๐ˆโ€ฒ
๐’‚๐Ÿ ๐œน๐’‹๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
= ๐ˆ ๐’‚๐Ÿ ๐Ÿ โˆ’ ๐ˆ ๐’‚๐Ÿ ๐œน๐’๐’–๐’•๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
= โˆ’๐ˆ ๐’‚๐Ÿ ๐Ÿ โˆ’ ๐ˆ ๐’‚๐Ÿ ๐’š โˆ’ ๐’›๐’๐’–๐’• ๐ˆ ๐’ƒ๐’๐’–๐’• ๐Ÿ โˆ’ ๐ˆ ๐’ƒ๐’๐’–๐’• ๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐œน๐’‹=
๐๐‘ณ
๐๐’‚๐’‹
๐œน๐’๐’–๐’• =
๐๐‘ณ
๐๐’ƒ๐’๐’–๐’•
= โˆ’ ๐’š โˆ’ เท
๐’š
๐เท
๐’š
๐๐’ƒ๐’๐’–๐’•
= โˆ’ ๐’š โˆ’ ๐’›๐’๐’–๐’•
๐๐’›๐’๐’–๐’•
๐๐’ƒ๐’๐’–๐’•
= โˆ’ ๐’š โˆ’ ๐’›๐’๐’–๐’• ๐ˆโ€ฒ ๐’ƒ๐’๐’–๐’•
= โˆ’ ๐’š โˆ’ ๐’›๐’๐’–๐’• ๐ˆ ๐’ƒ๐’๐’–๐’• ๐Ÿ โˆ’ ๐ˆ ๐’ƒ๐’๐’–๐’•
๐’™๐Ÿ
๐‘Ž1 ๐‘ง1
๐’ƒ๐’๐’–๐’• ๐’›๐’๐’–๐’•
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐’›๐Ÿ = ๐ˆ ๐’‚๐Ÿ
๐’›๐’๐’–๐’• = ๐ˆ ๐’ƒ๐’๐’–๐’• = เท
๐’š
๐’™๐Ÿ
๐‘Ž2 ๐‘ง2
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐’›๐Ÿ = ๐ˆ ๐’‚๐Ÿ
๐œน๐Ÿ = ๐ˆโ€ฒ ๐’‚๐Ÿ ๐œน๐’‹๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
= ๐ˆ ๐’‚๐Ÿ ๐Ÿ โˆ’ ๐ˆ ๐’‚๐Ÿ ๐œน๐’๐’–๐’•๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
= โˆ’๐ˆ ๐’‚๐Ÿ ๐Ÿ โˆ’ ๐ˆ ๐’‚๐Ÿ ๐’š โˆ’ ๐’›๐’๐’–๐’• ๐ˆ ๐’ƒ๐’๐’–๐’• ๐Ÿ โˆ’ ๐ˆ ๐’ƒ๐’๐’–๐’• ๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
Back-propagation Algorithm - error propagation (Update of Layer 1 weights)
0.5227
0.09101
๐ŸŽ. ๐Ÿ•๐Ÿ๐ŸŽ๐Ÿ—
0.9
๐ŸŽ. ๐Ÿ“๐Ÿ’๐Ÿ—๐Ÿ–
0.2
๐œน๐Ÿ = ๐ˆโ€ฒ
๐’‚๐Ÿ ๐œน๐’‹๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
= ๐ˆ ๐’‚๐Ÿ ๐Ÿ โˆ’ ๐ˆ ๐’‚๐Ÿ ๐œน๐’๐’–๐’•๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
= โˆ’๐ˆ ๐’‚๐Ÿ ๐Ÿ โˆ’ ๐ˆ ๐’‚๐Ÿ ๐’š โˆ’ ๐’›๐’๐’–๐’• ๐ˆ ๐’ƒ๐’๐’–๐’• ๐Ÿ โˆ’ ๐ˆ ๐’ƒ๐’๐’–๐’• ๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐œน๐’‹=
๐๐‘ณ
๐๐’‚๐’‹
๐œน๐’๐’–๐’• =
๐๐‘ณ
๐๐’ƒ๐’๐’–๐’•
= โˆ’ ๐’š โˆ’ เท
๐’š
๐เท
๐’š
๐๐’ƒ๐’๐’–๐’•
= โˆ’ ๐’š โˆ’ ๐’›๐’๐’–๐’•
๐๐’›๐’๐’–๐’•
๐๐’ƒ๐’๐’–๐’•
= โˆ’ ๐’š โˆ’ ๐’›๐’๐’–๐’• ๐ˆโ€ฒ
๐’ƒ๐’๐’–๐’•
= โˆ’ ๐’š โˆ’ ๐’›๐’๐’–๐’• ๐ˆ ๐’ƒ๐’๐’–๐’• ๐Ÿ โˆ’ ๐ˆ ๐’ƒ๐’๐’–๐’•
๐’™๐Ÿ
๐‘Ž1 ๐‘ง1
๐’ƒ๐’๐’–๐’• ๐’›๐’๐’–๐’•
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐’›๐Ÿ = ๐ˆ ๐’‚๐Ÿ
๐’›๐’๐’–๐’• = ๐ˆ ๐’ƒ๐’๐’–๐’• = เท
๐’š
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ) ๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
๐’™๐Ÿ
๐‘Ž2 ๐‘ง2
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐’›๐Ÿ = ๐ˆ ๐’‚๐Ÿ
๐œน๐Ÿ = ๐ˆโ€ฒ
๐’‚๐Ÿ ๐œน๐’‹๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
= ๐ˆ ๐’‚๐Ÿ ๐Ÿ โˆ’ ๐ˆ ๐’‚๐Ÿ ๐œน๐’๐’–๐’•๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
= โˆ’๐ˆ ๐’‚๐Ÿ ๐Ÿ โˆ’ ๐ˆ ๐’‚๐Ÿ ๐’š โˆ’ ๐’›๐’๐’–๐’• ๐ˆ ๐’ƒ๐’๐’–๐’• ๐Ÿ โˆ’ ๐ˆ ๐’ƒ๐’๐’–๐’• ๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
0.4
0.1
๐œน๐Ÿ = โˆ’ ๐ŸŽ. ๐Ÿ“๐Ÿ’๐Ÿ—๐Ÿ– ๐Ÿ โˆ’ ๐ŸŽ. ๐Ÿ“๐Ÿ’๐Ÿ—๐Ÿ– ๐Ÿ โˆ’ ๐ŸŽ. ๐Ÿ“๐Ÿ๐Ÿ๐Ÿ• ๐ŸŽ. ๐Ÿ“๐Ÿ๐Ÿ๐Ÿ• ๐Ÿ โˆ’ ๐ŸŽ. ๐Ÿ“๐Ÿ๐Ÿ๐Ÿ• ๐ŸŽ. ๐Ÿ’ = โˆ’๐ŸŽ. ๐ŸŽ๐Ÿ๐Ÿ๐Ÿ•๐Ÿ–๐Ÿ—
๐œน๐Ÿ = โˆ’ ๐ŸŽ. ๐Ÿ•๐Ÿ๐ŸŽ๐Ÿ— ๐Ÿ โˆ’ ๐ŸŽ. ๐Ÿ•๐Ÿ๐ŸŽ๐Ÿ— ๐Ÿ โˆ’ ๐ŸŽ. ๐Ÿ“๐Ÿ๐Ÿ๐Ÿ• ๐ŸŽ. ๐Ÿ“๐Ÿ๐Ÿ๐Ÿ• ๐Ÿ โˆ’ ๐ŸŽ. ๐Ÿ“๐Ÿ๐Ÿ๐Ÿ• ๐ŸŽ. ๐Ÿ = โˆ’๐ŸŽ. ๐ŸŽ๐ŸŽ๐Ÿ๐Ÿ’๐Ÿ’๐Ÿ•
0
1
๐’š = ๐Ÿ
1
1
1
Back-propagation Algorithm (Update of Layer 1 weights)
๐‘พ๐’Š๐’‹
๐‘ต๐’†๐’˜
= ๐‘พ๐’Š๐’‹
๐’๐’๐’…
โˆ’ ๐œผ
๐๐‘ณ
๐๐‘พ๐’Š๐’‹
๐’๐’๐’…
๐œน๐Ÿ = โˆ’๐ŸŽ. ๐ŸŽ๐Ÿ๐Ÿ๐Ÿ•๐Ÿ–๐Ÿ—
๐œน๐Ÿ = โˆ’๐ŸŽ. ๐ŸŽ๐ŸŽ๐Ÿ๐Ÿ’๐Ÿ’๐Ÿ•
๐œผ = ๐ŸŽ. ๐Ÿ๐Ÿ“
0.5227
0.09101
0.9
๐ŸŽ. ๐Ÿ•๐Ÿ๐ŸŽ๐Ÿ—
๐ŸŽ. ๐Ÿ“๐Ÿ’๐Ÿ—๐Ÿ–
0.2
๐’™๐Ÿ
๐‘Ž1 ๐‘ง1
๐’ƒ๐’๐’–๐’• ๐’›๐’๐’–๐’•
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐’›๐Ÿ = ๐ˆ ๐’‚๐Ÿ
๐’›๐’๐’–๐’• = ๐ˆ ๐’ƒ๐’๐’–๐’• = เท
๐’š
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
1
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
1
1
๐’™๐Ÿ
๐‘Ž2 ๐‘ง2
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐’›๐Ÿ = ๐ˆ ๐’‚๐Ÿ
0.6
0.4
0.4
0.1
-0.1
-0.3
0
1
๐’š = ๐Ÿ
0.3
0.5
-0.2
๐๐‘ณ
๐๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
= ๐œน๐Ÿ๐’™๐Ÿ = โˆ’๐ŸŽ. ๐ŸŽ๐Ÿ๐Ÿ๐Ÿ•๐Ÿ–๐Ÿ— ร— ๐ŸŽ = ๐ŸŽ
๐๐‘ณ
๐๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
= ๐œน๐Ÿ๐’™๐Ÿ = โˆ’๐ŸŽ. ๐ŸŽ๐Ÿ๐Ÿ๐Ÿ•๐Ÿ–๐Ÿ— ร— ๐Ÿ = โˆ’๐ŸŽ. ๐ŸŽ๐Ÿ๐Ÿ๐Ÿ•๐Ÿ–๐Ÿ—
๐๐‘ณ
๐๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
= ๐œน๐Ÿ๐’™๐ŸŽ = โˆ’๐ŸŽ. ๐ŸŽ๐Ÿ๐Ÿ๐Ÿ•๐Ÿ–๐Ÿ— ร— ๐Ÿ = โˆ’๐ŸŽ. ๐ŸŽ๐Ÿ๐Ÿ๐Ÿ•๐Ÿ–๐Ÿ—
๐๐‘ณ
๐๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
= ๐œน๐Ÿ๐’™๐Ÿ = โˆ’๐ŸŽ. ๐ŸŽ๐ŸŽ๐Ÿ๐Ÿ’๐Ÿ’๐Ÿ• ร— ๐ŸŽ = ๐ŸŽ
๐๐‘ณ
๐๐‘พ๐Ÿ๐Ÿ
(๐Ÿ) = ๐œน๐Ÿ๐’™๐Ÿ = โˆ’๐ŸŽ. ๐ŸŽ๐ŸŽ๐Ÿ๐Ÿ’๐Ÿ’๐Ÿ• ร— ๐Ÿ = โˆ’๐ŸŽ. ๐ŸŽ02447
๐๐‘ณ
๐๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ) = ๐œน๐Ÿ๐’™๐ŸŽ = โˆ’๐ŸŽ. ๐ŸŽ๐ŸŽ๐Ÿ๐Ÿ’๐Ÿ’๐Ÿ• ร— ๐Ÿ = โˆ’๐ŸŽ. ๐ŸŽ๐ŸŽ2447
๐‘พ๐Ÿ๐Ÿ
๐Ÿ
๐’• + ๐Ÿ = ๐‘พ๐Ÿ๐Ÿ
๐Ÿ
๐’• โˆ’ ๐ŸŽ. ๐Ÿ๐Ÿ“
๐๐‘ณ
๐๐‘พ๐Ÿ๐Ÿ
๐Ÿ
๐’•
= ๐ŸŽ. ๐Ÿ” โˆ’ ๐ŸŽ. ๐Ÿ๐Ÿ“ ร— ๐ŸŽ = ๐ŸŽ. ๐Ÿ”
๐‘พ๐Ÿ๐Ÿ
๐Ÿ
๐’• + ๐Ÿ = ๐‘พ๐Ÿ๐Ÿ
๐Ÿ
๐’• โˆ’ ๐ŸŽ. ๐Ÿ๐Ÿ“
๐๐‘ณ
๐๐‘พ๐Ÿ๐Ÿ
๐Ÿ
๐’•
= โˆ’๐ŸŽ. ๐Ÿ + ๐ŸŽ. ๐Ÿ๐Ÿ“ ร— ๐ŸŽ. ๐ŸŽ๐Ÿ๐Ÿ๐Ÿ•๐Ÿ–๐Ÿ— = โˆ’๐ŸŽ. . ๐ŸŽ๐Ÿ—๐Ÿ•๐ŸŽ๐Ÿ“
๐‘พ๐Ÿ๐ŸŽ
๐Ÿ
๐’• + ๐Ÿ = ๐‘พ๐Ÿ๐ŸŽ
๐Ÿ
๐’• โˆ’ ๐ŸŽ. ๐Ÿ๐Ÿ“
๐๐‘ณ
๐๐‘พ๐Ÿ๐ŸŽ
๐Ÿ
๐’•
= ๐ŸŽ. ๐Ÿ‘ + ๐ŸŽ. ๐Ÿ๐Ÿ“ ร— ๐ŸŽ. ๐ŸŽ๐Ÿ๐Ÿ๐Ÿ•๐Ÿ–๐Ÿ— = ๐ŸŽ. ๐Ÿ‘๐ŸŽ๐Ÿ๐Ÿ—๐Ÿ“
๐‘พ๐Ÿ๐Ÿ
๐Ÿ
๐’• + ๐Ÿ = ๐‘พ๐Ÿ๐Ÿ
๐Ÿ
๐’• โˆ’ ๐ŸŽ. ๐Ÿ๐Ÿ“
๐๐‘ณ
๐๐‘พ๐Ÿ๐Ÿ
๐Ÿ
๐’•
= โˆ’๐ŸŽ. ๐Ÿ‘ โˆ’ ๐ŸŽ. ๐Ÿ๐Ÿ“ ร— ๐ŸŽ = โˆ’๐ŸŽ. ๐Ÿ‘
๐‘พ๐Ÿ๐Ÿ
๐Ÿ
๐’• + ๐Ÿ = ๐‘พ๐Ÿ๐Ÿ
๐Ÿ
๐’• โˆ’ ๐ŸŽ. ๐Ÿ๐Ÿ“
๐๐‘ณ
๐๐‘พ๐Ÿ๐Ÿ
๐Ÿ
๐’•
= ๐ŸŽ. ๐Ÿ’ + ๐ŸŽ. ๐Ÿ๐Ÿ“ ร— ๐ŸŽ. ๐ŸŽ๐ŸŽ๐Ÿ๐Ÿ’๐Ÿ’๐Ÿ• = ๐ŸŽ.4006125
๐‘พ๐Ÿ๐ŸŽ
๐Ÿ
๐’• + ๐Ÿ = ๐‘พ๐Ÿ๐ŸŽ
๐Ÿ
๐’• โˆ’ ๐ŸŽ. ๐Ÿ๐Ÿ“
๐๐‘ณ
๐๐‘พ๐Ÿ๐ŸŽ
๐Ÿ
๐’•
= ๐ŸŽ. ๐Ÿ“ + ๐ŸŽ. ๐Ÿ๐Ÿ“ ร— ๐ŸŽ. ๐ŸŽ๐ŸŽ๐Ÿ๐Ÿ’๐Ÿ’๐Ÿ• = ๐ŸŽ. ๐Ÿ“๐ŸŽ๐ŸŽ๐Ÿ”๐Ÿ๐Ÿ๐Ÿ“
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
Back-propagation Algorithm (Update of layer 2 Weights)
๐‘พ๐’Š๐’‹
๐‘ต๐’†๐’˜
= ๐‘พ๐’Š๐’‹
๐’๐’๐’…
โˆ’ ๐œผ
๐๐‘ณ
๐๐‘พ๐’Š๐’‹
๐’๐’๐’…
๐œน๐Ÿ = โˆ’๐ŸŽ. ๐ŸŽ๐Ÿ๐Ÿ๐Ÿ•๐Ÿ–๐Ÿ—
๐œน๐Ÿ = โˆ’๐ŸŽ. ๐ŸŽ๐ŸŽ๐Ÿ๐Ÿ’๐Ÿ’๐Ÿ•
0.5227
0.09101
๐ŸŽ. ๐Ÿ•๐Ÿ๐ŸŽ๐Ÿ—
0.9
0.2 ๐ŸŽ. ๐Ÿ“๐Ÿ’๐Ÿ—๐Ÿ–
๐’™๐Ÿ
๐‘Ž1 ๐‘ง1
๐’ƒ๐’๐’–๐’• ๐’›๐’๐’–๐’•
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐’›๐Ÿ = ๐ˆ ๐’‚๐Ÿ
๐’›๐’๐’–๐’• = ๐ˆ ๐’ƒ๐’๐’–๐’• = เท
๐’š
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
1
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
1
1
๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
๐’™๐Ÿ
๐‘Ž2 ๐‘ง2
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
๐’›๐Ÿ = ๐ˆ ๐’‚๐Ÿ
0.6
0.4
0.4
0.1
-0.1
-0.3
0
1
๐’š = ๐Ÿ
0.3
0.5
-0.2
๐’ƒ๐’๐’–๐’• = ๐’›๐Ÿ๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
+ ๐’›๐Ÿ๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
+ ๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
, ๐’›๐’๐’–๐’• = ๐ˆ ๐’ƒ๐’๐’–๐’• = เท
๐’š
๐๐‘ณ
๐๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
= โˆ’ ๐’š โˆ’ เท
๐’š
๐เท
๐’š
๐๐‘พ๐Ÿ๐Ÿ
๐Ÿ
= โˆ’ ๐’š โˆ’ เท
๐’š ๐ˆโ€ฒ
๐’ƒ๐’๐’–๐’• ๐’›๐Ÿ = ๐œน๐’๐’–๐’•๐’›๐Ÿ
๐๐‘ณ
๐๐‘พ๐Ÿ๐Ÿ
(๐Ÿ)
= โˆ’ ๐’š โˆ’ เท
๐’š
๐เท
๐’š
๐๐‘พ๐Ÿ๐Ÿ
๐Ÿ
= โˆ’ ๐’š โˆ’ เท
๐’š ๐ˆโ€ฒ
๐’ƒ๐’๐’–๐’• ๐’›๐Ÿ = ๐œน๐’๐’–๐’•๐’›๐Ÿ
๐๐‘ณ
๐๐‘พ๐Ÿ๐ŸŽ
(๐Ÿ)
= โˆ’ ๐’š โˆ’ เท
๐’š
๐เท
๐’š
๐๐‘พ๐Ÿ๐ŸŽ
๐Ÿ
= โˆ’ ๐’š โˆ’ เท
๐’š ๐ˆโ€ฒ
๐’ƒ๐’๐’–๐’• = ๐œน๐’๐’–๐’•
๐ˆโ€ฒ
๐’ƒ๐’๐’–๐’• = ๐ˆ ๐’ƒ๐’๐’–๐’• ๐Ÿ โˆ’ ๐ˆ ๐’ƒ๐’๐’–๐’•
๐œน๐‘ถ๐’–๐’• = โˆ’ ๐’š โˆ’ เท
๐’š ๐ˆ ๐’ƒ๐’๐’–๐’• ๐Ÿ โˆ’ ๐ˆ ๐’ƒ๐’๐’–๐’•
= โˆ’ ๐Ÿ โˆ’ ๐ŸŽ. ๐Ÿ“๐Ÿ๐Ÿ๐Ÿ• ๐ŸŽ. ๐Ÿ“๐Ÿ๐Ÿ๐Ÿ• ๐Ÿ โˆ’ ๐ŸŽ. ๐Ÿ“๐Ÿ๐Ÿ๐Ÿ•
= โˆ’๐ŸŽ. ๐Ÿ๐Ÿ๐Ÿ—๐ŸŽ๐Ÿ–
๐‘พ๐Ÿ๐Ÿ
๐Ÿ
๐’• + ๐Ÿ = ๐‘พ๐Ÿ๐Ÿ
๐Ÿ
๐’• โˆ’ ๐œผ
๐๐‘ณ
๐๐‘พ๐Ÿ๐Ÿ
๐Ÿ
= ๐ŸŽ. ๐Ÿ’ + ๐ŸŽ. ๐Ÿ๐Ÿ“ ร— ๐ŸŽ. ๐Ÿ๐Ÿ๐Ÿ—๐ŸŽ๐Ÿ– ร— ๐ŸŽ. ๐Ÿ“๐Ÿ’๐Ÿ—๐Ÿ– = ๐ŸŽ. ๐Ÿ’๐Ÿ๐Ÿ”๐Ÿ’
๐‘พ๐Ÿ๐Ÿ
๐Ÿ
๐’• + ๐Ÿ = ๐‘พ๐Ÿ๐Ÿ
๐Ÿ
๐’• โˆ’ ๐œผ
๐๐‘ณ
๐๐‘พ๐Ÿ๐Ÿ
๐Ÿ
= ๐ŸŽ. ๐Ÿ + ๐ŸŽ. ๐Ÿ๐Ÿ“ ร— ๐ŸŽ. ๐Ÿ๐Ÿ๐Ÿ—๐ŸŽ๐Ÿ– ร— ๐ŸŽ. ๐Ÿ•๐Ÿ๐ŸŽ๐Ÿ— = ๐ŸŽ. ๐Ÿ๐Ÿ๐Ÿ๐Ÿ
๐‘พ๐Ÿ๐ŸŽ
๐Ÿ
๐’• + ๐Ÿ = ๐‘พ๐Ÿ๐ŸŽ
๐Ÿ
๐’• โˆ’ ๐œผ
๐๐‘ณ
๐๐‘พ๐Ÿ๐ŸŽ
๐Ÿ
= โˆ’๐ŸŽ. ๐Ÿ + ๐ŸŽ. ๐Ÿ๐Ÿ“ ร— ๐ŸŽ. ๐Ÿ๐Ÿ๐Ÿ—๐ŸŽ๐Ÿ– = โˆ’๐ŸŽ. ๐Ÿ๐Ÿ•๐ŸŽ๐Ÿ๐Ÿ‘

Mais conteรบdo relacionado

Mais procurados

Adaptive Resonance Theory (ART)
Adaptive Resonance Theory (ART)Adaptive Resonance Theory (ART)
Adaptive Resonance Theory (ART)Amir Masoud Sefidian
ย 
HOPFIELD NETWORK
HOPFIELD NETWORKHOPFIELD NETWORK
HOPFIELD NETWORKankita pandey
ย 
Batch normalization presentation
Batch normalization presentationBatch normalization presentation
Batch normalization presentationOwin Will
ย 
Hebbian Learning
Hebbian LearningHebbian Learning
Hebbian LearningESCOM
ย 
Back propagation method
Back propagation methodBack propagation method
Back propagation methodProf. Neeta Awasthy
ย 
Artificial Neural Networks Lect3: Neural Network Learning rules
Artificial Neural Networks Lect3: Neural Network Learning rulesArtificial Neural Networks Lect3: Neural Network Learning rules
Artificial Neural Networks Lect3: Neural Network Learning rulesMohammed Bennamoun
ย 
Hopfield Networks
Hopfield NetworksHopfield Networks
Hopfield NetworksKanchana Rani G
ย 
Feed forward ,back propagation,gradient descent
Feed forward ,back propagation,gradient descentFeed forward ,back propagation,gradient descent
Feed forward ,back propagation,gradient descentMuhammad Rasel
ย 
Associative memory network
Associative memory networkAssociative memory network
Associative memory networkDr. C.V. Suresh Babu
ย 
Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural NetworkPrakash K
ย 
Deep Learning: Introduction & Chapter 5 Machine Learning Basics
Deep Learning: Introduction & Chapter 5 Machine Learning BasicsDeep Learning: Introduction & Chapter 5 Machine Learning Basics
Deep Learning: Introduction & Chapter 5 Machine Learning BasicsJason Tsai
ย 
Multi Layer Perceptron & Back Propagation
Multi Layer Perceptron & Back PropagationMulti Layer Perceptron & Back Propagation
Multi Layer Perceptron & Back PropagationSung-ju Kim
ย 
03 Single layer Perception Classifier
03 Single layer Perception Classifier03 Single layer Perception Classifier
03 Single layer Perception ClassifierTamer Ahmed Farrag, PhD
ย 
Perceptron & Neural Networks
Perceptron & Neural NetworksPerceptron & Neural Networks
Perceptron & Neural NetworksNAGUR SHAREEF SHAIK
ย 
Neural Networks
Neural NetworksNeural Networks
Neural NetworksAdri Jovin
ย 
Neural Networks: Multilayer Perceptron
Neural Networks: Multilayer PerceptronNeural Networks: Multilayer Perceptron
Neural Networks: Multilayer PerceptronMostafa G. M. Mostafa
ย 
04 Multi-layer Feedforward Networks
04 Multi-layer Feedforward Networks04 Multi-layer Feedforward Networks
04 Multi-layer Feedforward NetworksTamer Ahmed Farrag, PhD
ย 
A brief introduction to mutual information and its application
A brief introduction to mutual information and its applicationA brief introduction to mutual information and its application
A brief introduction to mutual information and its applicationHyun-hwan Jeong
ย 

Mais procurados (20)

Adaptive Resonance Theory (ART)
Adaptive Resonance Theory (ART)Adaptive Resonance Theory (ART)
Adaptive Resonance Theory (ART)
ย 
HOPFIELD NETWORK
HOPFIELD NETWORKHOPFIELD NETWORK
HOPFIELD NETWORK
ย 
Batch normalization presentation
Batch normalization presentationBatch normalization presentation
Batch normalization presentation
ย 
Hebbian Learning
Hebbian LearningHebbian Learning
Hebbian Learning
ย 
Back propagation method
Back propagation methodBack propagation method
Back propagation method
ย 
Artificial Neural Networks Lect3: Neural Network Learning rules
Artificial Neural Networks Lect3: Neural Network Learning rulesArtificial Neural Networks Lect3: Neural Network Learning rules
Artificial Neural Networks Lect3: Neural Network Learning rules
ย 
Hopfield Networks
Hopfield NetworksHopfield Networks
Hopfield Networks
ย 
Feed forward ,back propagation,gradient descent
Feed forward ,back propagation,gradient descentFeed forward ,back propagation,gradient descent
Feed forward ,back propagation,gradient descent
ย 
Associative memory network
Associative memory networkAssociative memory network
Associative memory network
ย 
Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural Network
ย 
Deep Learning: Introduction & Chapter 5 Machine Learning Basics
Deep Learning: Introduction & Chapter 5 Machine Learning BasicsDeep Learning: Introduction & Chapter 5 Machine Learning Basics
Deep Learning: Introduction & Chapter 5 Machine Learning Basics
ย 
Genetic algorithm
Genetic algorithmGenetic algorithm
Genetic algorithm
ย 
Multi Layer Perceptron & Back Propagation
Multi Layer Perceptron & Back PropagationMulti Layer Perceptron & Back Propagation
Multi Layer Perceptron & Back Propagation
ย 
03 Single layer Perception Classifier
03 Single layer Perception Classifier03 Single layer Perception Classifier
03 Single layer Perception Classifier
ย 
Perceptron & Neural Networks
Perceptron & Neural NetworksPerceptron & Neural Networks
Perceptron & Neural Networks
ย 
Neural Networks
Neural NetworksNeural Networks
Neural Networks
ย 
Neural Networks: Multilayer Perceptron
Neural Networks: Multilayer PerceptronNeural Networks: Multilayer Perceptron
Neural Networks: Multilayer Perceptron
ย 
Unit 1
Unit 1Unit 1
Unit 1
ย 
04 Multi-layer Feedforward Networks
04 Multi-layer Feedforward Networks04 Multi-layer Feedforward Networks
04 Multi-layer Feedforward Networks
ย 
A brief introduction to mutual information and its application
A brief introduction to mutual information and its applicationA brief introduction to mutual information and its application
A brief introduction to mutual information and its application
ย 

Semelhante a Lecture 5 backpropagation

Deep learning study 2
Deep learning study 2Deep learning study 2
Deep learning study 2San Kim
ย 
Deep Feed Forward Neural Networks and Regularization
Deep Feed Forward Neural Networks and RegularizationDeep Feed Forward Neural Networks and Regularization
Deep Feed Forward Neural Networks and RegularizationYan Xu
ย 
Solving Poisson Equation using Conjugate Gradient Method and its implementation
Solving Poisson Equation using Conjugate Gradient Methodand its implementationSolving Poisson Equation using Conjugate Gradient Methodand its implementation
Solving Poisson Equation using Conjugate Gradient Method and its implementationJongsu "Liam" Kim
ย 
Vibration Isolation of a LEGOยฎ plate
Vibration Isolation of a LEGOยฎ plateVibration Isolation of a LEGOยฎ plate
Vibration Isolation of a LEGOยฎ plateOpen Adaptronik
ย 
Passivity-based control of rigid-body manipulator
Passivity-based control of rigid-body manipulatorPassivity-based control of rigid-body manipulator
Passivity-based control of rigid-body manipulatorHancheol Choi
ย 
Lecture 6 radial basis-function_network
Lecture 6 radial basis-function_networkLecture 6 radial basis-function_network
Lecture 6 radial basis-function_networkParveenMalik18
ย 
Stochastic optimal control & rl
Stochastic optimal control & rlStochastic optimal control & rl
Stochastic optimal control & rlChoiJinwon3
ย 
ุชุทุจูŠู‚ุงุช ุงู„ู…ุนุงุฏู„ุงุช ุงู„ุชูุงุถู„ูŠุฉ
ุชุทุจูŠู‚ุงุช ุงู„ู…ุนุงุฏู„ุงุช ุงู„ุชูุงุถู„ูŠุฉุชุทุจูŠู‚ุงุช ุงู„ู…ุนุงุฏู„ุงุช ุงู„ุชูุงุถู„ูŠุฉ
ุชุทุจูŠู‚ุงุช ุงู„ู…ุนุงุฏู„ุงุช ุงู„ุชูุงุถู„ูŠุฉMohammedRazzaqSalman
ย 
MLU_DTE_Lecture_2.pptx
MLU_DTE_Lecture_2.pptxMLU_DTE_Lecture_2.pptx
MLU_DTE_Lecture_2.pptxRahulChaudhry15
ย 
Lecture Notes: EEEC4340318 Instrumentation and Control Systems - Fundamental...
Lecture Notes:  EEEC4340318 Instrumentation and Control Systems - Fundamental...Lecture Notes:  EEEC4340318 Instrumentation and Control Systems - Fundamental...
Lecture Notes: EEEC4340318 Instrumentation and Control Systems - Fundamental...AIMST University
ย 
Introduction to PyTorch
Introduction to PyTorchIntroduction to PyTorch
Introduction to PyTorchJun Young Park
ย 
Introduction to Neural Networks and Deep Learning from Scratch
Introduction to Neural Networks and Deep Learning from ScratchIntroduction to Neural Networks and Deep Learning from Scratch
Introduction to Neural Networks and Deep Learning from ScratchAhmed BESBES
ย 
Introduction to Neural Networks and Deep Learning
Introduction to Neural Networks and Deep LearningIntroduction to Neural Networks and Deep Learning
Introduction to Neural Networks and Deep LearningVahid Mirjalili
ย 
Direct solution of sparse network equations by optimally ordered triangular f...
Direct solution of sparse network equations by optimally ordered triangular f...Direct solution of sparse network equations by optimally ordered triangular f...
Direct solution of sparse network equations by optimally ordered triangular f...Dimas Ruliandi
ย 
Back propagation
Back propagationBack propagation
Back propagationSan Kim
ย 

Semelhante a Lecture 5 backpropagation (20)

Deep learning study 2
Deep learning study 2Deep learning study 2
Deep learning study 2
ย 
Deep Feed Forward Neural Networks and Regularization
Deep Feed Forward Neural Networks and RegularizationDeep Feed Forward Neural Networks and Regularization
Deep Feed Forward Neural Networks and Regularization
ย 
Solving Poisson Equation using Conjugate Gradient Method and its implementation
Solving Poisson Equation using Conjugate Gradient Methodand its implementationSolving Poisson Equation using Conjugate Gradient Methodand its implementation
Solving Poisson Equation using Conjugate Gradient Method and its implementation
ย 
Neural Networks
Neural NetworksNeural Networks
Neural Networks
ย 
Vibration Isolation of a LEGOยฎ plate
Vibration Isolation of a LEGOยฎ plateVibration Isolation of a LEGOยฎ plate
Vibration Isolation of a LEGOยฎ plate
ย 
Passivity-based control of rigid-body manipulator
Passivity-based control of rigid-body manipulatorPassivity-based control of rigid-body manipulator
Passivity-based control of rigid-body manipulator
ย 
Lecture 6 radial basis-function_network
Lecture 6 radial basis-function_networkLecture 6 radial basis-function_network
Lecture 6 radial basis-function_network
ย 
Stochastic optimal control & rl
Stochastic optimal control & rlStochastic optimal control & rl
Stochastic optimal control & rl
ย 
Lec10.pptx
Lec10.pptxLec10.pptx
Lec10.pptx
ย 
ุชุทุจูŠู‚ุงุช ุงู„ู…ุนุงุฏู„ุงุช ุงู„ุชูุงุถู„ูŠุฉ
ุชุทุจูŠู‚ุงุช ุงู„ู…ุนุงุฏู„ุงุช ุงู„ุชูุงุถู„ูŠุฉุชุทุจูŠู‚ุงุช ุงู„ู…ุนุงุฏู„ุงุช ุงู„ุชูุงุถู„ูŠุฉ
ุชุทุจูŠู‚ุงุช ุงู„ู…ุนุงุฏู„ุงุช ุงู„ุชูุงุถู„ูŠุฉ
ย 
MLU_DTE_Lecture_2.pptx
MLU_DTE_Lecture_2.pptxMLU_DTE_Lecture_2.pptx
MLU_DTE_Lecture_2.pptx
ย 
Lecture Notes: EEEC4340318 Instrumentation and Control Systems - Fundamental...
Lecture Notes:  EEEC4340318 Instrumentation and Control Systems - Fundamental...Lecture Notes:  EEEC4340318 Instrumentation and Control Systems - Fundamental...
Lecture Notes: EEEC4340318 Instrumentation and Control Systems - Fundamental...
ย 
Introduction to PyTorch
Introduction to PyTorchIntroduction to PyTorch
Introduction to PyTorch
ย 
Introduction to Neural Networks and Deep Learning from Scratch
Introduction to Neural Networks and Deep Learning from ScratchIntroduction to Neural Networks and Deep Learning from Scratch
Introduction to Neural Networks and Deep Learning from Scratch
ย 
Introduction to Neural Networks and Deep Learning
Introduction to Neural Networks and Deep LearningIntroduction to Neural Networks and Deep Learning
Introduction to Neural Networks and Deep Learning
ย 
Direct solution of sparse network equations by optimally ordered triangular f...
Direct solution of sparse network equations by optimally ordered triangular f...Direct solution of sparse network equations by optimally ordered triangular f...
Direct solution of sparse network equations by optimally ordered triangular f...
ย 
DNN and RBM
DNN and RBMDNN and RBM
DNN and RBM
ย 
Backpropagation - Elisa Sayrol - UPC Barcelona 2018
Backpropagation - Elisa Sayrol - UPC Barcelona 2018Backpropagation - Elisa Sayrol - UPC Barcelona 2018
Backpropagation - Elisa Sayrol - UPC Barcelona 2018
ย 
Back propagation
Back propagationBack propagation
Back propagation
ย 
Backpropagation
BackpropagationBackpropagation
Backpropagation
ย 

Mais de ParveenMalik18

Lecture 4 neural networks
Lecture 4 neural networksLecture 4 neural networks
Lecture 4 neural networksParveenMalik18
ย 
Lecture 3 fuzzy inference system
Lecture 3  fuzzy inference systemLecture 3  fuzzy inference system
Lecture 3 fuzzy inference systemParveenMalik18
ย 
Lecture 2 fuzzy inference system
Lecture 2  fuzzy inference systemLecture 2  fuzzy inference system
Lecture 2 fuzzy inference systemParveenMalik18
ย 
Lecture 1 computational intelligence
Lecture 1  computational intelligenceLecture 1  computational intelligence
Lecture 1 computational intelligenceParveenMalik18
ย 
Electrical and Electronic Measurement
Electrical and Electronic MeasurementElectrical and Electronic Measurement
Electrical and Electronic MeasurementParveenMalik18
ย 

Mais de ParveenMalik18 (10)

Lecture 4 neural networks
Lecture 4 neural networksLecture 4 neural networks
Lecture 4 neural networks
ย 
Lecture 3 fuzzy inference system
Lecture 3  fuzzy inference systemLecture 3  fuzzy inference system
Lecture 3 fuzzy inference system
ย 
Lecture 2 fuzzy inference system
Lecture 2  fuzzy inference systemLecture 2  fuzzy inference system
Lecture 2 fuzzy inference system
ย 
Lecture 1 computational intelligence
Lecture 1  computational intelligenceLecture 1  computational intelligence
Lecture 1 computational intelligence
ย 
Chapter8
Chapter8Chapter8
Chapter8
ย 
Chapter6
Chapter6Chapter6
Chapter6
ย 
Chapter5
Chapter5Chapter5
Chapter5
ย 
Chapter3
Chapter3Chapter3
Chapter3
ย 
Chapter2
Chapter2Chapter2
Chapter2
ย 
Electrical and Electronic Measurement
Electrical and Electronic MeasurementElectrical and Electronic Measurement
Electrical and Electronic Measurement
ย 

รšltimo

Call Girls Service Nagpur Tanvi Call 7001035870 Meet With Nagpur Escorts
Call Girls Service Nagpur Tanvi Call 7001035870 Meet With Nagpur EscortsCall Girls Service Nagpur Tanvi Call 7001035870 Meet With Nagpur Escorts
Call Girls Service Nagpur Tanvi Call 7001035870 Meet With Nagpur EscortsCall Girls in Nagpur High Profile
ย 
CCS335 _ Neural Networks and Deep Learning Laboratory_Lab Complete Record
CCS335 _ Neural Networks and Deep Learning Laboratory_Lab Complete RecordCCS335 _ Neural Networks and Deep Learning Laboratory_Lab Complete Record
CCS335 _ Neural Networks and Deep Learning Laboratory_Lab Complete RecordAsst.prof M.Gokilavani
ย 
UNIT-III FMM. DIMENSIONAL ANALYSIS
UNIT-III FMM.        DIMENSIONAL ANALYSISUNIT-III FMM.        DIMENSIONAL ANALYSIS
UNIT-III FMM. DIMENSIONAL ANALYSISrknatarajan
ย 
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escortsranjana rawat
ย 
UNIT-V FMM.HYDRAULIC TURBINE - Construction and working
UNIT-V FMM.HYDRAULIC TURBINE - Construction and workingUNIT-V FMM.HYDRAULIC TURBINE - Construction and working
UNIT-V FMM.HYDRAULIC TURBINE - Construction and workingrknatarajan
ย 
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...Christo Ananth
ย 
HARMONY IN THE NATURE AND EXISTENCE - Unit-IV
HARMONY IN THE NATURE AND EXISTENCE - Unit-IVHARMONY IN THE NATURE AND EXISTENCE - Unit-IV
HARMONY IN THE NATURE AND EXISTENCE - Unit-IVRajaP95
ย 
High Profile Call Girls Nagpur Isha Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Isha Call 7001035870 Meet With Nagpur EscortsHigh Profile Call Girls Nagpur Isha Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Isha Call 7001035870 Meet With Nagpur Escortsranjana rawat
ย 
(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Service
(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Service(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Service
(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Serviceranjana rawat
ย 
Booking open Available Pune Call Girls Koregaon Park 6297143586 Call Hot Ind...
Booking open Available Pune Call Girls Koregaon Park  6297143586 Call Hot Ind...Booking open Available Pune Call Girls Koregaon Park  6297143586 Call Hot Ind...
Booking open Available Pune Call Girls Koregaon Park 6297143586 Call Hot Ind...Call Girls in Nagpur High Profile
ย 
APPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICS
APPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICSAPPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICS
APPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICSKurinjimalarL3
ย 
UNIT-II FMM-Flow Through Circular Conduits
UNIT-II FMM-Flow Through Circular ConduitsUNIT-II FMM-Flow Through Circular Conduits
UNIT-II FMM-Flow Through Circular Conduitsrknatarajan
ย 
(SHREYA) Chakan Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Esc...
(SHREYA) Chakan Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Esc...(SHREYA) Chakan Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Esc...
(SHREYA) Chakan Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Esc...ranjana rawat
ย 
Structural Analysis and Design of Foundations: A Comprehensive Handbook for S...
Structural Analysis and Design of Foundations: A Comprehensive Handbook for S...Structural Analysis and Design of Foundations: A Comprehensive Handbook for S...
Structural Analysis and Design of Foundations: A Comprehensive Handbook for S...Dr.Costas Sachpazis
ย 
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINE
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINEMANUFACTURING PROCESS-II UNIT-2 LATHE MACHINE
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINESIVASHANKAR N
ย 
Microscopic Analysis of Ceramic Materials.pptx
Microscopic Analysis of Ceramic Materials.pptxMicroscopic Analysis of Ceramic Materials.pptx
Microscopic Analysis of Ceramic Materials.pptxpurnimasatapathy1234
ย 
Call Girls Service Nashik Vaishnavi 7001305949 Independent Escort Service Nashik
Call Girls Service Nashik Vaishnavi 7001305949 Independent Escort Service NashikCall Girls Service Nashik Vaishnavi 7001305949 Independent Escort Service Nashik
Call Girls Service Nashik Vaishnavi 7001305949 Independent Escort Service NashikCall Girls in Nagpur High Profile
ย 
Introduction to Multiple Access Protocol.pptx
Introduction to Multiple Access Protocol.pptxIntroduction to Multiple Access Protocol.pptx
Introduction to Multiple Access Protocol.pptxupamatechverse
ย 
Call for Papers - Educational Administration: Theory and Practice, E-ISSN: 21...
Call for Papers - Educational Administration: Theory and Practice, E-ISSN: 21...Call for Papers - Educational Administration: Theory and Practice, E-ISSN: 21...
Call for Papers - Educational Administration: Theory and Practice, E-ISSN: 21...Christo Ananth
ย 

รšltimo (20)

Call Girls Service Nagpur Tanvi Call 7001035870 Meet With Nagpur Escorts
Call Girls Service Nagpur Tanvi Call 7001035870 Meet With Nagpur EscortsCall Girls Service Nagpur Tanvi Call 7001035870 Meet With Nagpur Escorts
Call Girls Service Nagpur Tanvi Call 7001035870 Meet With Nagpur Escorts
ย 
CCS335 _ Neural Networks and Deep Learning Laboratory_Lab Complete Record
CCS335 _ Neural Networks and Deep Learning Laboratory_Lab Complete RecordCCS335 _ Neural Networks and Deep Learning Laboratory_Lab Complete Record
CCS335 _ Neural Networks and Deep Learning Laboratory_Lab Complete Record
ย 
UNIT-III FMM. DIMENSIONAL ANALYSIS
UNIT-III FMM.        DIMENSIONAL ANALYSISUNIT-III FMM.        DIMENSIONAL ANALYSIS
UNIT-III FMM. DIMENSIONAL ANALYSIS
ย 
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
(MEERA) Dapodi Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Escorts
ย 
UNIT-V FMM.HYDRAULIC TURBINE - Construction and working
UNIT-V FMM.HYDRAULIC TURBINE - Construction and workingUNIT-V FMM.HYDRAULIC TURBINE - Construction and working
UNIT-V FMM.HYDRAULIC TURBINE - Construction and working
ย 
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...
ย 
HARMONY IN THE NATURE AND EXISTENCE - Unit-IV
HARMONY IN THE NATURE AND EXISTENCE - Unit-IVHARMONY IN THE NATURE AND EXISTENCE - Unit-IV
HARMONY IN THE NATURE AND EXISTENCE - Unit-IV
ย 
High Profile Call Girls Nagpur Isha Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Isha Call 7001035870 Meet With Nagpur EscortsHigh Profile Call Girls Nagpur Isha Call 7001035870 Meet With Nagpur Escorts
High Profile Call Girls Nagpur Isha Call 7001035870 Meet With Nagpur Escorts
ย 
Roadmap to Membership of RICS - Pathways and Routes
Roadmap to Membership of RICS - Pathways and RoutesRoadmap to Membership of RICS - Pathways and Routes
Roadmap to Membership of RICS - Pathways and Routes
ย 
(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Service
(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Service(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Service
(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Service
ย 
Booking open Available Pune Call Girls Koregaon Park 6297143586 Call Hot Ind...
Booking open Available Pune Call Girls Koregaon Park  6297143586 Call Hot Ind...Booking open Available Pune Call Girls Koregaon Park  6297143586 Call Hot Ind...
Booking open Available Pune Call Girls Koregaon Park 6297143586 Call Hot Ind...
ย 
APPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICS
APPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICSAPPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICS
APPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICS
ย 
UNIT-II FMM-Flow Through Circular Conduits
UNIT-II FMM-Flow Through Circular ConduitsUNIT-II FMM-Flow Through Circular Conduits
UNIT-II FMM-Flow Through Circular Conduits
ย 
(SHREYA) Chakan Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Esc...
(SHREYA) Chakan Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Esc...(SHREYA) Chakan Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Esc...
(SHREYA) Chakan Call Girls Just Call 7001035870 [ Cash on Delivery ] Pune Esc...
ย 
Structural Analysis and Design of Foundations: A Comprehensive Handbook for S...
Structural Analysis and Design of Foundations: A Comprehensive Handbook for S...Structural Analysis and Design of Foundations: A Comprehensive Handbook for S...
Structural Analysis and Design of Foundations: A Comprehensive Handbook for S...
ย 
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINE
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINEMANUFACTURING PROCESS-II UNIT-2 LATHE MACHINE
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINE
ย 
Microscopic Analysis of Ceramic Materials.pptx
Microscopic Analysis of Ceramic Materials.pptxMicroscopic Analysis of Ceramic Materials.pptx
Microscopic Analysis of Ceramic Materials.pptx
ย 
Call Girls Service Nashik Vaishnavi 7001305949 Independent Escort Service Nashik
Call Girls Service Nashik Vaishnavi 7001305949 Independent Escort Service NashikCall Girls Service Nashik Vaishnavi 7001305949 Independent Escort Service Nashik
Call Girls Service Nashik Vaishnavi 7001305949 Independent Escort Service Nashik
ย 
Introduction to Multiple Access Protocol.pptx
Introduction to Multiple Access Protocol.pptxIntroduction to Multiple Access Protocol.pptx
Introduction to Multiple Access Protocol.pptx
ย 
Call for Papers - Educational Administration: Theory and Practice, E-ISSN: 21...
Call for Papers - Educational Administration: Theory and Practice, E-ISSN: 21...Call for Papers - Educational Administration: Theory and Practice, E-ISSN: 21...
Call for Papers - Educational Administration: Theory and Practice, E-ISSN: 21...
ย 

Lecture 5 backpropagation

  • 1. Parveen Malik Assistant Professor KIIT University Neural Networks Backpropagation
  • 2. Background โ€ข Perceptron Learning Algorithm , Hebbian Learning can classify input pattern if input patterns are linearly separable. โ€ข We need an algorithm which can train multilayer of perceptron or classify patterns which are not linearly separable. โ€ข Algorithm should also be able to use non-linear activation function. ๐’™๐Ÿ ๐’™๐Ÿ ๐‘ช๐’๐’‚๐’”๐’” ๐Ÿ ๐‘ช๐’๐’‚๐’”๐’” ๐Ÿ ๐‘ณ๐’Š๐’๐’†๐’‚๐’“๐’๐’š ๐‘บ๐’†๐’‘๐’†๐’“๐’‚๐’ƒ๐’๐’† ๐’™๐Ÿ ๐’™๐Ÿ ๐‘ช๐’๐’‚๐’”๐’” ๐Ÿ ๐‘ช๐’๐’‚๐’”๐’” ๐Ÿ ๐‘ณ๐’Š๐’๐’†๐’‚๐’“๐’๐’š ๐‘ต๐’๐’ โˆ’ ๐’”๐’†๐’‘๐’†๐’“๐’‚๐’ƒ๐’๐’† โ€ข Need non-linear boundaries โ€ข Perceptron algorithm can't be used โ€ข Variation of GD rule is used. โ€ข More layers are required โ€ข Non-linear activation function required Perceptron Algorithm โˆ’ ๐‘พ๐’Š+๐Ÿ ๐‘ต๐’†๐’˜ = ๐‘พ๐’Š ๐‘ต๐’†๐’˜ + (๐’• โˆ’ ๐’‚)๐’™๐’Š Gradient Descent Algorithm- ๐‘พ๐’Š+๐Ÿ ๐‘ต๐’†๐’˜ = ๐‘พ๐’Š ๐’๐’๐’… โˆ’ ๐œผ ๐๐‘ณ ๐๐’˜๐’Š ๐‘ณ๐’๐’”๐’” ๐’‡๐’–๐’๐’„๐’•๐’Š๐’๐’, ๐‘ณ = ๐Ÿ ๐Ÿ ๐’• โˆ’ ๐’‚ ๐Ÿ
  • 3. Background- Back Propagation โ€ข The perceptron learning rule of Frank Rosenblatt and the LMS algorithm of Bernard Widrow and Marcian Hoff were designed to train single-layer perceptron-like networks. โ€ข Single-layer networks suffer from the disadvantage that they are only able to solve linearly separable classification problems. Both Rosenblatt and Widrow were aware of these limitations and proposed multilayer networks that could overcome them, but they were not able to generalize their algorithms to train these more powerful networks. โ€ข First description of an algorithm to train multilayer networks was contained in the thesis of Paul Werbos in 1974 .This thesis presented the algorithm in the context of general networks, with neural networks as a special case, and was not disseminated in the neural network community. โ€ข It was not until the mid 1980s that the backpropagation algorithm was rediscovered and widely publicized. It was rediscovered independently by David Rumelhart, Geoffrey Hinton and Ronald Williams 1986, David Parker 1985 and Yann Le Cun 1985. โ€ข The algorithm was popularized by its inclusion in the book Parallel Distributed Processing [RuMc86], which described the work of the Parallel Distributed Processing Group led by psychologists David Rumelhart and James Mc-Clelland โ€ข The multilayer perceptron, trained by the backpropagation algorithm, is currently the most widely used neural network.
  • 4. Network Design Problem : Whether you watch a movie or not ? Step 1 : Design โ€“ Output can be Yes (1) or No (0). Therefore one neuron or perceptron is sufficient. Step -2 : Choose suitable activation function in the output along with a rule to update the weights. (Hard Limit function for perceptron learning algorithm, sigmoid for the Widrow-Hoff rule or delta rule.) ๐‘พ๐’Š+๐Ÿ ๐‘ต๐’†๐’˜ = ๐‘พ๐’Š ๐’๐’๐’… โˆ’ ๐œผ ๐๐‘ณ ๐๐’˜๐’Š ๐‘ณ = ๐Ÿ ๐Ÿ ๐’š โˆ’ เท ๐’š ๐Ÿ = ๐Ÿ ๐Ÿ ๐’š โˆ’ ๐’‡ ๐’˜๐’™ + ๐’ƒ ๐Ÿ ๐๐‘ณ ๐๐’˜ = ๐Ÿ โˆ— ๐Ÿ ๐Ÿ ๐’š โˆ’ ๐’‡ ๐’˜๐’™ + ๐’ƒ ๐๐’‡ ๐’˜๐’™ + ๐’ƒ ๐๐’˜ = โˆ’ ๐’š โˆ’ เท ๐’š ๐’‡โ€ฒ ๐’˜๐’™ + ๐’ƒ ๐’™ ๐‘ฅ เท ๐‘“ Director or Actor or Genre or IMDB w Yes (1) or No (0) ๐‘ค๐‘ฅ + ๐‘ เท ๐’š = ๐’‡ ๐’˜๐’™ + ๐’ƒ = ๐Ÿ ๐Ÿ + ๐’†โˆ’๐’˜๐’™+๐’ƒ ๐’‡ ๐’˜๐’™ + ๐’ƒ ๐‘ค0 = ๐‘ 1
  • 5. Network Design Problem : Sort the students in the 4 house based on their three qualities like lineage, choice and ethics ? Step 1 : Design โ€“ Here, the input vector is 3-D i.e for each students, ๐‘†๐‘ก๐‘ข๐‘‘๐‘’๐‘›๐‘ก 1 = ๐ฟ1 ๐ถ1 ๐ธ1 , ๐‘†๐‘ก๐‘ข๐‘‘๐‘’๐‘›๐‘ก 2 = ๐ฟ2 ๐ถ2 ๐ธ2 ๐’™๐Ÿ ๐’™๐Ÿ ๐’™๐Ÿ‘ N ๐’™๐Ÿ ๐’™๐Ÿ ๐’™๐Ÿ‘ ๐’™๐ŸŽ=1 ๐’˜๐Ÿ ๐’˜๐Ÿ ๐’˜๐Ÿ‘ ๐’˜๐ŸŽ = ๐’ƒ Yes (1) or No (0) ๐‘1 ๐‘2 เท ๐’š๐Ÿ = ๐’‡ ๐’˜๐Ÿ๐Ÿ๐’™๐Ÿ + ๐’˜๐Ÿ๐Ÿ๐’™๐Ÿ + ๐’˜๐Ÿ๐Ÿ‘๐’™๐Ÿ‘ + ๐’ƒ๐Ÿ ๐’™๐Ÿ ๐’™๐Ÿ ๐’™๐Ÿ‘ ๐’˜๐Ÿ๐Ÿ ๐’˜๐Ÿ๐Ÿ ๐’˜๐Ÿ๐Ÿ‘ ๐’˜๐Ÿ๐Ÿ ๐’˜๐Ÿ๐Ÿ ๐’˜๐Ÿ๐Ÿ‘ เท ๐’š๐Ÿ = ๐’‡ ๐’˜๐Ÿ๐Ÿ๐’™๐Ÿ + ๐’˜๐Ÿ๐Ÿ๐’™๐Ÿ + ๐’˜๐Ÿ๐Ÿ‘๐’™๐Ÿ‘ + ๐’ƒ๐Ÿ ๐’ƒ๐Ÿ ๐’ƒ๐Ÿ เทœ ๐‘ฆ1 เทœ ๐‘ฆ2 0 1 1 0 1 1 0 0 A B C D Houses ๐‘ฆ1 ๐‘ฆ2 Actual Output Target Output
  • 6. Network Design Step 2 : Choosing the activation function and rule to update weights Loss function, ๐ฟ = 1 2 ๐‘ฆ โˆ’ เทœ ๐‘ฆ 2 1 1 ๐‘1 ๐‘2 เท ๐’š๐Ÿ = ๐’‡ ๐’˜๐Ÿ๐Ÿ๐’™๐Ÿ + ๐’˜๐Ÿ๐Ÿ๐’™๐Ÿ + ๐’˜๐Ÿ๐Ÿ‘๐’™๐Ÿ‘ + ๐’ƒ๐Ÿ ๐’™๐Ÿ ๐’™๐Ÿ ๐’™๐Ÿ‘ ๐’˜๐Ÿ๐Ÿ ๐’˜๐Ÿ๐Ÿ ๐’˜๐Ÿ๐Ÿ‘ ๐’˜๐Ÿ๐Ÿ ๐’˜๐Ÿ๐Ÿ ๐’˜๐Ÿ๐Ÿ‘ เท ๐’š๐Ÿ = ๐’‡ ๐’˜๐Ÿ๐Ÿ๐’™๐Ÿ + ๐’˜๐Ÿ๐Ÿ๐’™๐Ÿ + ๐’˜๐Ÿ๐Ÿ‘๐’™๐Ÿ‘ + ๐’ƒ๐Ÿ ๐’ƒ๐Ÿ ๐’ƒ๐Ÿ เทœ ๐‘ฆ1 เทœ ๐‘ฆ2 0 1 1 0 0 0 A B C D Houses ๐‘ฆ1 ๐‘ฆ2 Actual Output Target Output ๐‘พ๐’Š๐’‹ ๐’• + ๐Ÿ = ๐‘พ๐’Š๐’‹ ๐’• โˆ’ ๐œผ ๐๐‘ณ ๐๐’˜๐’Š๐’‹ ๐๐‘ณ ๐๐’˜๐Ÿ๐Ÿ = ๐’š โˆ’ เท ๐’š ๐’‡โ€ฒ ๐’˜๐Ÿ๐Ÿ๐’™๐Ÿ + +๐’˜๐Ÿ๐Ÿ๐’™๐Ÿ + ๐’˜๐Ÿ๐Ÿ‘๐’™๐Ÿ‘ + ๐’ƒ๐Ÿ ๐’™๐Ÿ
  • 8. Network Architectures (More Complex) ๐‘ฅ1 ๐‘ฅ2 ๐‘ฅ2 ๐‘ฅ2 โ„Ž2 (1) โ„Ž1 (1) โ„Ž3 (1) โ„Ž1 (2) โ„Ž2 (2) โ„Ž3 (2) ๐‘ฆ1 ๐‘ฆ2 ๐‘พ(๐Ÿ) ๐‘พ(๐Ÿ) ๐‘พ(๐Ÿ‘)
  • 9. Input ๐๐‘ณ ๐๐‘พ๐’Š๐’ = ๐œน๐’Š๐’๐’ ๐œน๐’Š = ๐ˆโ€ฒ ๐’‚๐’Š เท ๐‘ฑ ๐œน๐’‹๐‘พ๐’‹๐’Š ๐‘Ž๐‘™ ๐‘ง๐‘™ ๐‘Ž๐‘– ๐‘ง๐‘– ๐‘Ž๐‘— ๐‘ง๐‘— โ‹ฎ โ‹ฎ โ‹ฎ โ‹ฎ โ‹ฎ โ‹ฎ ๐‘พ๐’Š๐’ ๐‘พ๐’‹๐’Š ๐œน๐’Š = ๐๐‘ณ ๐๐’‚๐’Š ๐œน๐’‹ = ๐๐‘ณ ๐๐’‚๐’‹ Cost Function ๐‘ณ = ๐Ÿ ๐Ÿ (๐’š โˆ’ เท ๐’š) ๐Ÿ Error to input layer ๐ˆ ๐’‚๐’Š ๐Ÿ โˆ’ ๐ˆ ๐’‚๐’Š Back-propagation Algorithm (Generalized Expression) ๐œน๐’‹ = ๐๐‘ณ ๐๐’‚๐’‹ = โˆ’ ๐’š โˆ’ เท ๐’š ๐เท ๐’š ๐๐’‚๐’‹ ๐’‚๐’Š = เท ๐’ ๐‘พ๐’Š๐’๐’๐’ ๐œน๐’Š = ๐๐‘ณ ๐๐’‚๐’Š = เท ๐‘ฑ ๐๐‘ณ ๐๐’‚๐’‹ ๐๐’‚๐’‹ ๐๐’‚๐’Š ๐๐’‚๐’‹ ๐๐’‚๐’Š = ๐๐’‚๐’‹ ๐๐’๐’Š ๐๐’๐’Š ๐๐’‚๐’Š = ๐‘พ๐’‹๐’Š๐ˆโ€ฒ ๐’‚๐’Š
  • 10. Back-propagation Algorithm ๐’™๐Ÿ ๐’™๐Ÿ เท ๐’š ๐‘Ž1 ๐œŽ ๐‘Ž1 ๐‘Ž2 ๐œŽ ๐‘Ž2 ๐‘1 ๐œŽ ๐‘1 1 1 1 0.6 0.4 -0.1 0.5 -0.3 0.3 0.4 0.1 -0.2
  • 11. Back-propagation Algorithm ๐’™๐Ÿ ๐’™๐Ÿ เท ๐’š ๐‘Ž1 ๐œŽ ๐‘Ž1 ๐‘Ž2 ๐œŽ ๐‘Ž2 ๐‘1 ๐œŽ ๐‘1 ๐‘ป๐’‚๐’“๐’ˆ๐’†๐’• ๐’š = ๐Ÿ 1 1 1 0 1 0.6 0.4 -0.1 0.5 -0.3 0.3 0.4 0.1 -0.2 ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ)
  • 12. Back-propagation Algorithm ๐’™๐Ÿ ๐’™๐Ÿ เท ๐’š ๐‘Ž1 ๐œŽ ๐‘Ž1 ๐‘Ž2 ๐œŽ ๐‘Ž2 ๐‘1 ๐œŽ ๐‘1 ๐‘ป๐’‚๐’“๐’ˆ๐’†๐’• ๐’š = ๐Ÿ 1 1 1 0 1 0.6 0.4 -0.1 0.5 -0.3 0.3 0.4 0.1 -0.2 ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘ณ = ๐Ÿ ๐Ÿ ๐’š โˆ’ เท ๐’š ๐Ÿ Loss function
  • 13. Back-propagation Algorithm ๐’™๐Ÿ ๐’™๐Ÿ เท ๐’š ๐‘Ž1 ๐œŽ ๐‘Ž1 ๐‘Ž2 ๐œŽ ๐‘Ž2 ๐‘1 ๐œŽ ๐‘1 ๐‘ป๐’‚๐’“๐’ˆ๐’†๐’• ๐’š = ๐Ÿ 1 1 1 0 1 0.6 0.4 -0.1 0.5 -0.3 0.3 0.4 0.1 -0.2 ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘ณ = ๐Ÿ ๐Ÿ ๐’š โˆ’ เท ๐’š ๐Ÿ Loss function ๐‘พ๐‘ต๐’†๐’˜ = ๐‘พ๐’๐’๐’… โˆ’ ๐œผ ๐๐‘ณ ๐๐‘พ๐’๐’๐’…
  • 14. Back-propagation Algorithm Step 1 : Forward pass ๐’™๐Ÿ ๐’™๐Ÿ ๐‘Ž1 ๐œŽ ๐‘Ž1 ๐’‚๐Ÿ=0.2 ๐‘Ž2 ๐œŽ ๐‘Ž2 ๐‘1 ๐œŽ ๐‘1 ๐‘ป๐’‚๐’“๐’ˆ๐’†๐’• ๐’š = ๐Ÿ 1 1 1 0 1 0.6 0.4 -0.1 0.5 -0.3 0.3 0.4 0.1 -0.2 ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) ๐‘Š 12 (1) ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘ณ = ๐Ÿ ๐Ÿ ๐’š โˆ’ เท ๐’š ๐Ÿ Loss function ๐ˆ ๐’‚๐Ÿ = ๐Ÿ ๐Ÿ + ๐’†โˆ’๐ŸŽ.๐Ÿ = ๐ŸŽ. ๐Ÿ“๐Ÿ’๐Ÿ—๐Ÿ– ๐‘พ๐‘ต๐’†๐’˜ = ๐‘พ๐’๐’๐’… โˆ’ ๐œผ ๐๐‘ณ ๐๐‘พ๐’๐’๐’…
  • 15. Back-propagation Algorithm Step 1 : Forward pass ๐’™๐Ÿ ๐’™๐Ÿ ๐‘Ž1 ๐œŽ ๐‘Ž1 ๐’‚๐Ÿ=0.2 ๐’‚๐Ÿ=0.9 ๐‘Ž2 ๐œŽ ๐‘Ž2 ๐‘1 ๐œŽ ๐‘1 ๐‘ป๐’‚๐’“๐’ˆ๐’†๐’• ๐’š = ๐Ÿ 1 1 1 0 1 0.6 0.4 -0.1 0.5 -0.3 0.3 0.4 0.1 -0.2 ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘ณ = ๐Ÿ ๐Ÿ ๐’š โˆ’ เท ๐’š ๐Ÿ Loss function ๐ˆ ๐’‚๐Ÿ = ๐Ÿ ๐Ÿ + ๐’†โˆ’๐ŸŽ.๐Ÿ = ๐ŸŽ. ๐Ÿ“๐Ÿ’๐Ÿ—๐Ÿ– ๐ˆ ๐’‚๐Ÿ = ๐Ÿ ๐Ÿ + ๐’†โˆ’๐ŸŽ.๐Ÿ— = ๐ŸŽ. ๐Ÿ•๐Ÿ๐ŸŽ๐Ÿ— เท ๐’š ๐‘พ๐‘ต๐’†๐’˜ = ๐‘พ๐’๐’๐’… โˆ’ ๐œผ ๐๐‘ณ ๐๐‘พ๐’๐’๐’…
  • 16. Back-propagation Algorithm Step 1 : Forward pass ๐’™๐Ÿ ๐’™๐Ÿ ๐‘Ž1 ๐œŽ ๐‘Ž1 ๐’‚๐Ÿ=0.2 ๐’‚๐Ÿ=0.9 ๐’ƒ๐Ÿ=0.09101 ๐‘Ž2 ๐œŽ ๐‘Ž2 ๐‘1 ๐œŽ ๐‘1 เท ๐’š = ๐ˆ ๐’ƒ๐Ÿ =0.5227 ๐‘ป๐’‚๐’“๐’ˆ๐’†๐’• ๐’š = ๐Ÿ 1 1 1 0 1 0.6 0.4 -0.1 0.5 -0.3 0.3 0.4 0.1 -0.2 ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘ณ = ๐Ÿ ๐Ÿ ๐’š โˆ’ เท ๐’š ๐Ÿ Loss function ๐ˆ ๐’‚๐Ÿ = ๐Ÿ ๐Ÿ + ๐’†โˆ’๐ŸŽ.๐Ÿ = ๐ŸŽ. ๐Ÿ“๐Ÿ’๐Ÿ—๐Ÿ– ๐ˆ ๐’‚๐Ÿ = ๐Ÿ ๐Ÿ + ๐’†โˆ’๐ŸŽ.๐Ÿ— = ๐ŸŽ. ๐Ÿ•๐Ÿ๐ŸŽ๐Ÿ—
  • 17. Back-propagation Algorithm Step 2 : Backpropagation of error ๐’™๐Ÿ ๐’™๐Ÿ ๐‘Ž1 ๐œŽ ๐‘Ž1 ๐’‚๐Ÿ=0.2 ๐’‚๐Ÿ=0.9 ๐’ƒ๐Ÿ=0.09101 ๐‘Ž2 ๐œŽ ๐‘Ž2 ๐‘1 ๐œŽ ๐‘1 เท ๐’š = ๐ˆ ๐’ƒ๐Ÿ =0.5227 ๐‘ป๐’‚๐’“๐’ˆ๐’†๐’• ๐’š = ๐Ÿ 1 1 1 0 1 0.6 0.4 -0.1 0.5 -0.3 0.3 0.4 0.1 -0.2 ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘ณ = ๐Ÿ ๐Ÿ ๐’š โˆ’ เท ๐’š ๐Ÿ Loss function ๐ˆ ๐’‚๐Ÿ = ๐Ÿ ๐Ÿ + ๐’†โˆ’๐ŸŽ.๐Ÿ = ๐ŸŽ. ๐Ÿ“๐Ÿ’๐Ÿ—๐Ÿ– ๐ˆ ๐’‚๐Ÿ = ๐Ÿ ๐Ÿ + ๐’†โˆ’๐ŸŽ.๐Ÿ— = ๐ŸŽ. ๐Ÿ•๐Ÿ๐ŸŽ๐Ÿ— ๐‘Ž๐‘™ ๐‘ง๐‘™ ๐‘Ž๐‘– ๐‘ง๐‘– ๐‘Ž๐‘— ๐‘ง๐‘— โ‹ฎ โ‹ฎ โ‹ฎ โ‹ฎ โ‹ฎ โ‹ฎ ๐๐‘ณ ๐๐‘พ๐’Š๐’ = ๐œน๐’Š๐’๐’ ๐œน๐’Š = ๐ˆโ€ฒ ๐’‚๐’Š เท ๐‘ฑ ๐œน๐’‹๐‘พ๐’‹๐’Š ๐‘พ๐’Š๐’ ๐‘พ๐’‹๐’Š Imagine
  • 18. Back-propagation Algorithm ๐๐‘ณ ๐๐‘พ๐’Š๐’ = ๐œน๐’Š๐’๐’ ๐œน๐’Š = ๐ˆโ€ฒ ๐’‚๐’Š เท ๐‘ฑ ๐œน๐’‹๐‘พ๐’‹๐’Š ๐‘Ž๐‘™ ๐‘ง๐‘™ ๐‘Ž๐‘– ๐‘ง๐‘– ๐‘Ž๐‘— ๐‘ง๐‘— โ‹ฎ โ‹ฎ โ‹ฎ โ‹ฎ โ‹ฎ โ‹ฎ ๐‘พ๐’Š๐’ ๐‘พ๐’‹๐’Š ๐œน๐’Š= ๐๐‘ณ ๐๐’‚๐’Š ๐œน๐’‹= ๐๐‘ณ ๐๐’‚๐’‹ ๐ˆโ€ฒ ๐’‚๐’Š = ๐ˆ ๐’‚๐’Š ๐Ÿ โˆ’ ๐ˆ ๐’‚๐’Š Cost/Error Function ๐‘ณ = ๐Ÿ ๐Ÿ (๐’š โˆ’ เท ๐’š) ๐Ÿ ๐’›๐’Š = ๐ˆ ๐’‚๐’Š ๐œน๐Ÿ = ๐ˆโ€ฒ ๐’‚๐Ÿ ๐œน๐’‹๐‘พ๐Ÿ๐Ÿ (๐Ÿ) = ๐ˆ ๐’‚๐Ÿ ๐Ÿ โˆ’ ๐ˆ ๐’‚๐Ÿ ๐œน๐’๐’–๐’•๐‘พ๐Ÿ๐Ÿ (๐Ÿ) = โˆ’๐ˆ ๐’‚๐Ÿ ๐Ÿ โˆ’ ๐ˆ ๐’‚๐Ÿ ๐’š โˆ’ ๐’›๐’๐’–๐’• ๐ˆ ๐’ƒ๐’๐’–๐’• ๐Ÿ โˆ’ ๐ˆ ๐’ƒ๐’๐’–๐’• ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐œน๐’‹= ๐๐‘ณ ๐๐’‚๐’‹ ๐œน๐’๐’–๐’• = ๐๐‘ณ ๐๐’ƒ๐’๐’–๐’• = โˆ’ ๐’š โˆ’ เท ๐’š ๐เท ๐’š ๐๐’ƒ๐’๐’–๐’• = โˆ’ ๐’š โˆ’ ๐’›๐’๐’–๐’• ๐๐’›๐’๐’–๐’• ๐๐’ƒ๐’๐’–๐’• = โˆ’ ๐’š โˆ’ ๐’›๐’๐’–๐’• ๐ˆโ€ฒ ๐’ƒ๐’๐’–๐’• = โˆ’ ๐’š โˆ’ ๐’›๐’๐’–๐’• ๐ˆ ๐’ƒ๐’๐’–๐’• ๐Ÿ โˆ’ ๐ˆ ๐’ƒ๐’๐’–๐’• ๐’™๐Ÿ ๐‘Ž1 ๐‘ง1 ๐’ƒ๐’๐’–๐’• ๐’›๐’๐’–๐’• ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐’›๐Ÿ = ๐ˆ ๐’‚๐Ÿ ๐’›๐’๐’–๐’• = ๐ˆ ๐’ƒ๐’๐’–๐’• = เท ๐’š ๐’™๐Ÿ ๐‘Ž2 ๐‘ง2 ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐’›๐Ÿ = ๐ˆ ๐’‚๐Ÿ ๐œน๐Ÿ = ๐ˆโ€ฒ ๐’‚๐Ÿ ๐œน๐’‹๐‘พ๐Ÿ๐Ÿ (๐Ÿ) = ๐ˆ ๐’‚๐Ÿ ๐Ÿ โˆ’ ๐ˆ ๐’‚๐Ÿ ๐œน๐’๐’–๐’•๐‘พ๐Ÿ๐Ÿ (๐Ÿ) = โˆ’๐ˆ ๐’‚๐Ÿ ๐Ÿ โˆ’ ๐ˆ ๐’‚๐Ÿ ๐’š โˆ’ ๐’›๐’๐’–๐’• ๐ˆ ๐’ƒ๐’๐’–๐’• ๐Ÿ โˆ’ ๐ˆ ๐’ƒ๐’๐’–๐’• ๐‘พ๐Ÿ๐Ÿ (๐Ÿ)
  • 19. Back-propagation Algorithm - error propagation (Update of Layer 1 weights) 0.5227 0.09101 ๐ŸŽ. ๐Ÿ•๐Ÿ๐ŸŽ๐Ÿ— 0.9 ๐ŸŽ. ๐Ÿ“๐Ÿ’๐Ÿ—๐Ÿ– 0.2 ๐œน๐Ÿ = ๐ˆโ€ฒ ๐’‚๐Ÿ ๐œน๐’‹๐‘พ๐Ÿ๐Ÿ (๐Ÿ) = ๐ˆ ๐’‚๐Ÿ ๐Ÿ โˆ’ ๐ˆ ๐’‚๐Ÿ ๐œน๐’๐’–๐’•๐‘พ๐Ÿ๐Ÿ (๐Ÿ) = โˆ’๐ˆ ๐’‚๐Ÿ ๐Ÿ โˆ’ ๐ˆ ๐’‚๐Ÿ ๐’š โˆ’ ๐’›๐’๐’–๐’• ๐ˆ ๐’ƒ๐’๐’–๐’• ๐Ÿ โˆ’ ๐ˆ ๐’ƒ๐’๐’–๐’• ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐œน๐’‹= ๐๐‘ณ ๐๐’‚๐’‹ ๐œน๐’๐’–๐’• = ๐๐‘ณ ๐๐’ƒ๐’๐’–๐’• = โˆ’ ๐’š โˆ’ เท ๐’š ๐เท ๐’š ๐๐’ƒ๐’๐’–๐’• = โˆ’ ๐’š โˆ’ ๐’›๐’๐’–๐’• ๐๐’›๐’๐’–๐’• ๐๐’ƒ๐’๐’–๐’• = โˆ’ ๐’š โˆ’ ๐’›๐’๐’–๐’• ๐ˆโ€ฒ ๐’ƒ๐’๐’–๐’• = โˆ’ ๐’š โˆ’ ๐’›๐’๐’–๐’• ๐ˆ ๐’ƒ๐’๐’–๐’• ๐Ÿ โˆ’ ๐ˆ ๐’ƒ๐’๐’–๐’• ๐’™๐Ÿ ๐‘Ž1 ๐‘ง1 ๐’ƒ๐’๐’–๐’• ๐’›๐’๐’–๐’• ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐’›๐Ÿ = ๐ˆ ๐’‚๐Ÿ ๐’›๐’๐’–๐’• = ๐ˆ ๐’ƒ๐’๐’–๐’• = เท ๐’š ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) ๐’™๐Ÿ ๐‘Ž2 ๐‘ง2 ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐’›๐Ÿ = ๐ˆ ๐’‚๐Ÿ ๐œน๐Ÿ = ๐ˆโ€ฒ ๐’‚๐Ÿ ๐œน๐’‹๐‘พ๐Ÿ๐Ÿ (๐Ÿ) = ๐ˆ ๐’‚๐Ÿ ๐Ÿ โˆ’ ๐ˆ ๐’‚๐Ÿ ๐œน๐’๐’–๐’•๐‘พ๐Ÿ๐Ÿ (๐Ÿ) = โˆ’๐ˆ ๐’‚๐Ÿ ๐Ÿ โˆ’ ๐ˆ ๐’‚๐Ÿ ๐’š โˆ’ ๐’›๐’๐’–๐’• ๐ˆ ๐’ƒ๐’๐’–๐’• ๐Ÿ โˆ’ ๐ˆ ๐’ƒ๐’๐’–๐’• ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) 0.4 0.1 ๐œน๐Ÿ = โˆ’ ๐ŸŽ. ๐Ÿ“๐Ÿ’๐Ÿ—๐Ÿ– ๐Ÿ โˆ’ ๐ŸŽ. ๐Ÿ“๐Ÿ’๐Ÿ—๐Ÿ– ๐Ÿ โˆ’ ๐ŸŽ. ๐Ÿ“๐Ÿ๐Ÿ๐Ÿ• ๐ŸŽ. ๐Ÿ“๐Ÿ๐Ÿ๐Ÿ• ๐Ÿ โˆ’ ๐ŸŽ. ๐Ÿ“๐Ÿ๐Ÿ๐Ÿ• ๐ŸŽ. ๐Ÿ’ = โˆ’๐ŸŽ. ๐ŸŽ๐Ÿ๐Ÿ๐Ÿ•๐Ÿ–๐Ÿ— ๐œน๐Ÿ = โˆ’ ๐ŸŽ. ๐Ÿ•๐Ÿ๐ŸŽ๐Ÿ— ๐Ÿ โˆ’ ๐ŸŽ. ๐Ÿ•๐Ÿ๐ŸŽ๐Ÿ— ๐Ÿ โˆ’ ๐ŸŽ. ๐Ÿ“๐Ÿ๐Ÿ๐Ÿ• ๐ŸŽ. ๐Ÿ“๐Ÿ๐Ÿ๐Ÿ• ๐Ÿ โˆ’ ๐ŸŽ. ๐Ÿ“๐Ÿ๐Ÿ๐Ÿ• ๐ŸŽ. ๐Ÿ = โˆ’๐ŸŽ. ๐ŸŽ๐ŸŽ๐Ÿ๐Ÿ’๐Ÿ’๐Ÿ• 0 1 ๐’š = ๐Ÿ 1 1 1
  • 20. Back-propagation Algorithm (Update of Layer 1 weights) ๐‘พ๐’Š๐’‹ ๐‘ต๐’†๐’˜ = ๐‘พ๐’Š๐’‹ ๐’๐’๐’… โˆ’ ๐œผ ๐๐‘ณ ๐๐‘พ๐’Š๐’‹ ๐’๐’๐’… ๐œน๐Ÿ = โˆ’๐ŸŽ. ๐ŸŽ๐Ÿ๐Ÿ๐Ÿ•๐Ÿ–๐Ÿ— ๐œน๐Ÿ = โˆ’๐ŸŽ. ๐ŸŽ๐ŸŽ๐Ÿ๐Ÿ’๐Ÿ’๐Ÿ• ๐œผ = ๐ŸŽ. ๐Ÿ๐Ÿ“ 0.5227 0.09101 0.9 ๐ŸŽ. ๐Ÿ•๐Ÿ๐ŸŽ๐Ÿ— ๐ŸŽ. ๐Ÿ“๐Ÿ’๐Ÿ—๐Ÿ– 0.2 ๐’™๐Ÿ ๐‘Ž1 ๐‘ง1 ๐’ƒ๐’๐’–๐’• ๐’›๐’๐’–๐’• ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐’›๐Ÿ = ๐ˆ ๐’‚๐Ÿ ๐’›๐’๐’–๐’• = ๐ˆ ๐’ƒ๐’๐’–๐’• = เท ๐’š ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) 1 ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) 1 1 ๐’™๐Ÿ ๐‘Ž2 ๐‘ง2 ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐’›๐Ÿ = ๐ˆ ๐’‚๐Ÿ 0.6 0.4 0.4 0.1 -0.1 -0.3 0 1 ๐’š = ๐Ÿ 0.3 0.5 -0.2 ๐๐‘ณ ๐๐‘พ๐Ÿ๐Ÿ (๐Ÿ) = ๐œน๐Ÿ๐’™๐Ÿ = โˆ’๐ŸŽ. ๐ŸŽ๐Ÿ๐Ÿ๐Ÿ•๐Ÿ–๐Ÿ— ร— ๐ŸŽ = ๐ŸŽ ๐๐‘ณ ๐๐‘พ๐Ÿ๐Ÿ (๐Ÿ) = ๐œน๐Ÿ๐’™๐Ÿ = โˆ’๐ŸŽ. ๐ŸŽ๐Ÿ๐Ÿ๐Ÿ•๐Ÿ–๐Ÿ— ร— ๐Ÿ = โˆ’๐ŸŽ. ๐ŸŽ๐Ÿ๐Ÿ๐Ÿ•๐Ÿ–๐Ÿ— ๐๐‘ณ ๐๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) = ๐œน๐Ÿ๐’™๐ŸŽ = โˆ’๐ŸŽ. ๐ŸŽ๐Ÿ๐Ÿ๐Ÿ•๐Ÿ–๐Ÿ— ร— ๐Ÿ = โˆ’๐ŸŽ. ๐ŸŽ๐Ÿ๐Ÿ๐Ÿ•๐Ÿ–๐Ÿ— ๐๐‘ณ ๐๐‘พ๐Ÿ๐Ÿ (๐Ÿ) = ๐œน๐Ÿ๐’™๐Ÿ = โˆ’๐ŸŽ. ๐ŸŽ๐ŸŽ๐Ÿ๐Ÿ’๐Ÿ’๐Ÿ• ร— ๐ŸŽ = ๐ŸŽ ๐๐‘ณ ๐๐‘พ๐Ÿ๐Ÿ (๐Ÿ) = ๐œน๐Ÿ๐’™๐Ÿ = โˆ’๐ŸŽ. ๐ŸŽ๐ŸŽ๐Ÿ๐Ÿ’๐Ÿ’๐Ÿ• ร— ๐Ÿ = โˆ’๐ŸŽ. ๐ŸŽ02447 ๐๐‘ณ ๐๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) = ๐œน๐Ÿ๐’™๐ŸŽ = โˆ’๐ŸŽ. ๐ŸŽ๐ŸŽ๐Ÿ๐Ÿ’๐Ÿ’๐Ÿ• ร— ๐Ÿ = โˆ’๐ŸŽ. ๐ŸŽ๐ŸŽ2447 ๐‘พ๐Ÿ๐Ÿ ๐Ÿ ๐’• + ๐Ÿ = ๐‘พ๐Ÿ๐Ÿ ๐Ÿ ๐’• โˆ’ ๐ŸŽ. ๐Ÿ๐Ÿ“ ๐๐‘ณ ๐๐‘พ๐Ÿ๐Ÿ ๐Ÿ ๐’• = ๐ŸŽ. ๐Ÿ” โˆ’ ๐ŸŽ. ๐Ÿ๐Ÿ“ ร— ๐ŸŽ = ๐ŸŽ. ๐Ÿ” ๐‘พ๐Ÿ๐Ÿ ๐Ÿ ๐’• + ๐Ÿ = ๐‘พ๐Ÿ๐Ÿ ๐Ÿ ๐’• โˆ’ ๐ŸŽ. ๐Ÿ๐Ÿ“ ๐๐‘ณ ๐๐‘พ๐Ÿ๐Ÿ ๐Ÿ ๐’• = โˆ’๐ŸŽ. ๐Ÿ + ๐ŸŽ. ๐Ÿ๐Ÿ“ ร— ๐ŸŽ. ๐ŸŽ๐Ÿ๐Ÿ๐Ÿ•๐Ÿ–๐Ÿ— = โˆ’๐ŸŽ. . ๐ŸŽ๐Ÿ—๐Ÿ•๐ŸŽ๐Ÿ“ ๐‘พ๐Ÿ๐ŸŽ ๐Ÿ ๐’• + ๐Ÿ = ๐‘พ๐Ÿ๐ŸŽ ๐Ÿ ๐’• โˆ’ ๐ŸŽ. ๐Ÿ๐Ÿ“ ๐๐‘ณ ๐๐‘พ๐Ÿ๐ŸŽ ๐Ÿ ๐’• = ๐ŸŽ. ๐Ÿ‘ + ๐ŸŽ. ๐Ÿ๐Ÿ“ ร— ๐ŸŽ. ๐ŸŽ๐Ÿ๐Ÿ๐Ÿ•๐Ÿ–๐Ÿ— = ๐ŸŽ. ๐Ÿ‘๐ŸŽ๐Ÿ๐Ÿ—๐Ÿ“ ๐‘พ๐Ÿ๐Ÿ ๐Ÿ ๐’• + ๐Ÿ = ๐‘พ๐Ÿ๐Ÿ ๐Ÿ ๐’• โˆ’ ๐ŸŽ. ๐Ÿ๐Ÿ“ ๐๐‘ณ ๐๐‘พ๐Ÿ๐Ÿ ๐Ÿ ๐’• = โˆ’๐ŸŽ. ๐Ÿ‘ โˆ’ ๐ŸŽ. ๐Ÿ๐Ÿ“ ร— ๐ŸŽ = โˆ’๐ŸŽ. ๐Ÿ‘ ๐‘พ๐Ÿ๐Ÿ ๐Ÿ ๐’• + ๐Ÿ = ๐‘พ๐Ÿ๐Ÿ ๐Ÿ ๐’• โˆ’ ๐ŸŽ. ๐Ÿ๐Ÿ“ ๐๐‘ณ ๐๐‘พ๐Ÿ๐Ÿ ๐Ÿ ๐’• = ๐ŸŽ. ๐Ÿ’ + ๐ŸŽ. ๐Ÿ๐Ÿ“ ร— ๐ŸŽ. ๐ŸŽ๐ŸŽ๐Ÿ๐Ÿ’๐Ÿ’๐Ÿ• = ๐ŸŽ.4006125 ๐‘พ๐Ÿ๐ŸŽ ๐Ÿ ๐’• + ๐Ÿ = ๐‘พ๐Ÿ๐ŸŽ ๐Ÿ ๐’• โˆ’ ๐ŸŽ. ๐Ÿ๐Ÿ“ ๐๐‘ณ ๐๐‘พ๐Ÿ๐ŸŽ ๐Ÿ ๐’• = ๐ŸŽ. ๐Ÿ“ + ๐ŸŽ. ๐Ÿ๐Ÿ“ ร— ๐ŸŽ. ๐ŸŽ๐ŸŽ๐Ÿ๐Ÿ’๐Ÿ’๐Ÿ• = ๐ŸŽ. ๐Ÿ“๐ŸŽ๐ŸŽ๐Ÿ”๐Ÿ๐Ÿ๐Ÿ“ ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ)
  • 21. Back-propagation Algorithm (Update of layer 2 Weights) ๐‘พ๐’Š๐’‹ ๐‘ต๐’†๐’˜ = ๐‘พ๐’Š๐’‹ ๐’๐’๐’… โˆ’ ๐œผ ๐๐‘ณ ๐๐‘พ๐’Š๐’‹ ๐’๐’๐’… ๐œน๐Ÿ = โˆ’๐ŸŽ. ๐ŸŽ๐Ÿ๐Ÿ๐Ÿ•๐Ÿ–๐Ÿ— ๐œน๐Ÿ = โˆ’๐ŸŽ. ๐ŸŽ๐ŸŽ๐Ÿ๐Ÿ’๐Ÿ’๐Ÿ• 0.5227 0.09101 ๐ŸŽ. ๐Ÿ•๐Ÿ๐ŸŽ๐Ÿ— 0.9 0.2 ๐ŸŽ. ๐Ÿ“๐Ÿ’๐Ÿ—๐Ÿ– ๐’™๐Ÿ ๐‘Ž1 ๐‘ง1 ๐’ƒ๐’๐’–๐’• ๐’›๐’๐’–๐’• ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐’›๐Ÿ = ๐ˆ ๐’‚๐Ÿ ๐’›๐’๐’–๐’• = ๐ˆ ๐’ƒ๐’๐’–๐’• = เท ๐’š ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) 1 ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) 1 1 ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) ๐’™๐Ÿ ๐‘Ž2 ๐‘ง2 ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) ๐’›๐Ÿ = ๐ˆ ๐’‚๐Ÿ 0.6 0.4 0.4 0.1 -0.1 -0.3 0 1 ๐’š = ๐Ÿ 0.3 0.5 -0.2 ๐’ƒ๐’๐’–๐’• = ๐’›๐Ÿ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) + ๐’›๐Ÿ๐‘พ๐Ÿ๐Ÿ (๐Ÿ) + ๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) , ๐’›๐’๐’–๐’• = ๐ˆ ๐’ƒ๐’๐’–๐’• = เท ๐’š ๐๐‘ณ ๐๐‘พ๐Ÿ๐Ÿ (๐Ÿ) = โˆ’ ๐’š โˆ’ เท ๐’š ๐เท ๐’š ๐๐‘พ๐Ÿ๐Ÿ ๐Ÿ = โˆ’ ๐’š โˆ’ เท ๐’š ๐ˆโ€ฒ ๐’ƒ๐’๐’–๐’• ๐’›๐Ÿ = ๐œน๐’๐’–๐’•๐’›๐Ÿ ๐๐‘ณ ๐๐‘พ๐Ÿ๐Ÿ (๐Ÿ) = โˆ’ ๐’š โˆ’ เท ๐’š ๐เท ๐’š ๐๐‘พ๐Ÿ๐Ÿ ๐Ÿ = โˆ’ ๐’š โˆ’ เท ๐’š ๐ˆโ€ฒ ๐’ƒ๐’๐’–๐’• ๐’›๐Ÿ = ๐œน๐’๐’–๐’•๐’›๐Ÿ ๐๐‘ณ ๐๐‘พ๐Ÿ๐ŸŽ (๐Ÿ) = โˆ’ ๐’š โˆ’ เท ๐’š ๐เท ๐’š ๐๐‘พ๐Ÿ๐ŸŽ ๐Ÿ = โˆ’ ๐’š โˆ’ เท ๐’š ๐ˆโ€ฒ ๐’ƒ๐’๐’–๐’• = ๐œน๐’๐’–๐’• ๐ˆโ€ฒ ๐’ƒ๐’๐’–๐’• = ๐ˆ ๐’ƒ๐’๐’–๐’• ๐Ÿ โˆ’ ๐ˆ ๐’ƒ๐’๐’–๐’• ๐œน๐‘ถ๐’–๐’• = โˆ’ ๐’š โˆ’ เท ๐’š ๐ˆ ๐’ƒ๐’๐’–๐’• ๐Ÿ โˆ’ ๐ˆ ๐’ƒ๐’๐’–๐’• = โˆ’ ๐Ÿ โˆ’ ๐ŸŽ. ๐Ÿ“๐Ÿ๐Ÿ๐Ÿ• ๐ŸŽ. ๐Ÿ“๐Ÿ๐Ÿ๐Ÿ• ๐Ÿ โˆ’ ๐ŸŽ. ๐Ÿ“๐Ÿ๐Ÿ๐Ÿ• = โˆ’๐ŸŽ. ๐Ÿ๐Ÿ๐Ÿ—๐ŸŽ๐Ÿ– ๐‘พ๐Ÿ๐Ÿ ๐Ÿ ๐’• + ๐Ÿ = ๐‘พ๐Ÿ๐Ÿ ๐Ÿ ๐’• โˆ’ ๐œผ ๐๐‘ณ ๐๐‘พ๐Ÿ๐Ÿ ๐Ÿ = ๐ŸŽ. ๐Ÿ’ + ๐ŸŽ. ๐Ÿ๐Ÿ“ ร— ๐ŸŽ. ๐Ÿ๐Ÿ๐Ÿ—๐ŸŽ๐Ÿ– ร— ๐ŸŽ. ๐Ÿ“๐Ÿ’๐Ÿ—๐Ÿ– = ๐ŸŽ. ๐Ÿ’๐Ÿ๐Ÿ”๐Ÿ’ ๐‘พ๐Ÿ๐Ÿ ๐Ÿ ๐’• + ๐Ÿ = ๐‘พ๐Ÿ๐Ÿ ๐Ÿ ๐’• โˆ’ ๐œผ ๐๐‘ณ ๐๐‘พ๐Ÿ๐Ÿ ๐Ÿ = ๐ŸŽ. ๐Ÿ + ๐ŸŽ. ๐Ÿ๐Ÿ“ ร— ๐ŸŽ. ๐Ÿ๐Ÿ๐Ÿ—๐ŸŽ๐Ÿ– ร— ๐ŸŽ. ๐Ÿ•๐Ÿ๐ŸŽ๐Ÿ— = ๐ŸŽ. ๐Ÿ๐Ÿ๐Ÿ๐Ÿ ๐‘พ๐Ÿ๐ŸŽ ๐Ÿ ๐’• + ๐Ÿ = ๐‘พ๐Ÿ๐ŸŽ ๐Ÿ ๐’• โˆ’ ๐œผ ๐๐‘ณ ๐๐‘พ๐Ÿ๐ŸŽ ๐Ÿ = โˆ’๐ŸŽ. ๐Ÿ + ๐ŸŽ. ๐Ÿ๐Ÿ“ ร— ๐ŸŽ. ๐Ÿ๐Ÿ๐Ÿ—๐ŸŽ๐Ÿ– = โˆ’๐ŸŽ. ๐Ÿ๐Ÿ•๐ŸŽ๐Ÿ๐Ÿ‘