Perceptron and its learning rule, Limitations of perceptron

Machine Learning Using Python Artificial Neural Network
12 minutes
Share the link to this page
Copied
  Completed
You need to have access to the item to view this lesson.
One-time Fee
$69.99
List Price:  $99.99
You save:  $30
€59.73
List Price:  €85.34
You save:  €25.60
£51.81
List Price:  £74.02
You save:  £22.20
CA$96.84
List Price:  CA$138.36
You save:  CA$41.51
A$106.75
List Price:  A$152.51
You save:  A$45.75
S$89.95
List Price:  S$128.50
You save:  S$38.55
HK$545.64
List Price:  HK$779.53
You save:  HK$233.88
CHF 55.86
List Price:  CHF 79.80
You save:  CHF 23.94
NOK kr703.17
List Price:  NOK kr1,004.57
You save:  NOK kr301.40
DKK kr445.91
List Price:  DKK kr637.04
You save:  DKK kr191.13
NZ$118.76
List Price:  NZ$169.67
You save:  NZ$50.90
د.إ257.03
List Price:  د.إ367.21
You save:  د.إ110.17
৳8,495.52
List Price:  ৳12,136.98
You save:  ৳3,641.45
₹6,172.17
List Price:  ₹8,817.76
You save:  ₹2,645.59
RM295.67
List Price:  RM422.40
You save:  RM126.73
₦106,514.70
List Price:  ₦152,170.38
You save:  ₦45,655.68
₨19,808.40
List Price:  ₨28,298.93
You save:  ₨8,490.52
฿2,243.26
List Price:  ฿3,204.79
You save:  ฿961.53
₺2,883.76
List Price:  ₺4,119.83
You save:  ₺1,236.07
B$378.92
List Price:  B$541.35
You save:  B$162.42
R1,231.03
List Price:  R1,758.69
You save:  R527.66
Лв116.90
List Price:  Лв167.02
You save:  Лв50.11
₩97,041.13
List Price:  ₩138,636.13
You save:  ₩41,595
₪233.01
List Price:  ₪332.89
You save:  ₪99.87
₱3,968.43
List Price:  ₱5,669.43
You save:  ₱1,701
¥10,316.66
List Price:  ¥14,738.72
You save:  ¥4,422.05
MX$1,309.87
List Price:  MX$1,871.33
You save:  MX$561.45
QR255.14
List Price:  QR364.50
You save:  QR109.36
P938.51
List Price:  P1,340.79
You save:  P402.27
KSh9,032.87
List Price:  KSh12,904.65
You save:  KSh3,871.78
E£3,397.07
List Price:  E£4,853.17
You save:  E£1,456.09
ብር9,985.48
List Price:  ብር14,265.58
You save:  ብር4,280.10
Kz64,180.83
List Price:  Kz91,690.83
You save:  Kz27,510
CLP$67,863.28
List Price:  CLP$96,951.70
You save:  CLP$29,088.41
CN¥499.22
List Price:  CN¥713.21
You save:  CN¥213.98
RD$4,414.63
List Price:  RD$6,306.89
You save:  RD$1,892.25
DA9,089.11
List Price:  DA12,985
You save:  DA3,895.89
FJ$157.67
List Price:  FJ$225.25
You save:  FJ$67.58
Q535.38
List Price:  Q764.86
You save:  Q229.48
GY$14,604.61
List Price:  GY$20,864.62
You save:  GY$6,260.01
ISK kr8,549.85
List Price:  ISK kr12,214.60
You save:  ISK kr3,664.74
DH634.40
List Price:  DH906.33
You save:  DH271.92
L1,171.66
List Price:  L1,673.87
You save:  L502.21
ден3,674.28
List Price:  ден5,249.20
You save:  ден1,574.92
MOP$561.02
List Price:  MOP$801.49
You save:  MOP$240.47
N$1,234.02
List Price:  N$1,762.97
You save:  N$528.94
C$2,569.13
List Price:  C$3,670.35
You save:  C$1,101.21
रु9,858.12
List Price:  रु14,083.64
You save:  रु4,225.51
S/245.88
List Price:  S/351.28
You save:  S/105.39
K291.36
List Price:  K416.25
You save:  K124.88
SAR262.47
List Price:  SAR374.98
You save:  SAR112.50
ZK1,666.62
List Price:  ZK2,380.99
You save:  ZK714.36
L303.21
List Price:  L433.17
You save:  L129.96
Kč1,456.45
List Price:  Kč2,080.74
You save:  Kč624.28
Ft23,442.21
List Price:  Ft33,490.30
You save:  Ft10,048.09
SEK kr657.57
List Price:  SEK kr939.43
You save:  SEK kr281.85
ARS$95,451.09
List Price:  ARS$136,364.55
You save:  ARS$40,913.45
Bs482.36
List Price:  Bs689.11
You save:  Bs206.75
COP$278,383.76
List Price:  COP$397,708.14
You save:  COP$119,324.37
₡35,369.65
List Price:  ₡50,530.25
You save:  ₡15,160.59
L1,828.90
List Price:  L2,612.83
You save:  L783.92
₲503,139.27
List Price:  ₲718,801.20
You save:  ₲215,661.92
$U2,810.15
List Price:  $U4,014.67
You save:  $U1,204.52
zł253.83
List Price:  zł362.63
You save:  zł108.80
Already have an account? Log In

Transcript

Hello everyone, welcome to the course of machine learning with Python. In this video, we shall start a new module called artificial neural network. And in this particular video, we shall discuss about perceptron model. So this is the brief history of neural learning system. In 1940s McCulloch and Pitt's come up with the concept of electronic brain in 1950 Rosenblatt came up with the concept of perceptron. In 1960, we have come up with the concept of atoms.

However, in 1970, Minsky and pat pepper showed that perceptron model is unable to solve certain classes of problem known as XOR problem or nonlinear the separable data set problem. So, this creates a huge setback in the AI community. And this is what he calls the dark age or the AI winter from 1970 T two t 1986. In the 1986 rule heart, intern and Williams come up with the concept of backpropagation learning algorithm in multi layer perceptron, which will eventually solve this XOR problem and give these field a boost. However, because of the lack of computational instruments at that point of time, this was not a huge success in 1995, they actually come up with the concept of support vector machine. And in 2006 onwards, deep neural network was a trend because now people have that much of computational power to train deep neural network and people started getting amazing results.

Today, every research in the field of artificial intelligence is extremely concerned with deep neural network. So let's see from where we have got the inspiration to build an app. neural network. The biological neuron is basically the building block of our nervous system. This is called the cell body of the biological neuron and from the cell body there are so many extensions. So these small extensions are known as dendrites, and there isn't long extension that is known as axon.

So, what does this did right or dangerous do it the impulses from the other cell bodies of the other neurons is actually transferred to another neuron by the means of these dendrites, okay. And the dendrites of other neurons or the axon of other noodles is connected with the dendrite of this neuron and the connection is known as the signups. So through axon the impulse is basically carried away from the cell body, okay, this is the basic concept of violence you can move on. So, what we have understood so there is As his body, the cell body basically get impulses from the other neurons by means of dendrites, and it passes through the acceleration or the impulse through axon to some other neuron. Okay? Now, artificial neuron.

So this is what he called the cell body, and these are basically the dendrites. So, through dendrites, we're actually taking the sensation from the outside world or the actual from another neuron. So these are nothing but numbers or the values of the features. Okay, and they are synetic weights, denoted by W zero, W one or w two. In cell body, we compute the weighted summation of all the inputs from the other neurons. This weighted solution then passed through a function called the activation function.

And finally, we get the output So, this is how we have modeled our original biological neuron to some mathematical model known as artificial neuron. So, B here is known as the bias, why is that called a signup tick weights are simply weights and this is a very crude approximation of original biological neuron. Actual biological neurons are even more complex, this is known as the activation function. For example, the activation function could be a sigmoid activation function or a logistic activation function. So, perceptron So, this model of neuron is also called macula fits model. However, Rosenblatt extended this idea and give a learning rule to update the weight parameters from a given data set.

This is called perceptron. The activation function if in this case is a hard limiter or signum function, let V is nothing but a weighted summation of all the inputs Then output Y hat is calculated as it is plus one, if v is greater than equals to zero, it is minus one if E is less than zero, okay, so that is how the activation function if map's v to Y hat, so if he's value is non negative, then y hat or the output from the cell body, we will be one or plus one. If V is negative, then the output from the cell body will be minus one. So this is the scale of the activation function as we can see V is greater than zero, then the output is plus one. If he's less than zero, the output is minus one. Now comes perceptron learning rule.

So there are a few assumptions. First, we considered these a binary classification problem. The data set is x, and there are two classes denoted by C one and C two respectively. class level of a data point x vector is basically denoted by y. So, x vector is a data point inside the data set capital X, it is denoted as following. Now, y is equals to plus one if these expected belongs to class one and y is equals to minus one, if expected, we have two plus two.

So, that is how we have label the classes the waves are denoted by the vector w, the bias is denoted by B. Now, this weight and bias could be the lead can be considered as a weight vector w, if we consider another input x zero of Hello always plus one, then this whole wi x i plus b can be written as the dot product of this w vector and this is nothing one W transpose X no algorithm. So first, we should initialize the word vector with some small random number and the iteration count. P is equals to one. Now, we shall repeat for x vector in the data set capital X we shall compute y hat which is nothing but signal function of W transpose X or the dot product between w vector and x. This white hat is also called the computed output or the label.

Now, in t plus one is iteration the weight vector is nothing but the winter active iteration plus eater times y minus y hat or multiplied with the X Factor. What is why why is called the actual output or the level the actual output is obtained from the label training data. And it is platform if expected belongs to plus one and it is minus one if x vector belongs to plus two. So we deplete these tapes until, in two successive iterations, the award victory is not changed. So, how do we measure that a weight vector has not changed. So, let's say we compute the norm of the difference between the weight vector at t plus one iteration and with vector A t iteration, and if this norm is less than some predefined tolerance or very small number epsilon, then we shall see that our algorithm has fun first.

Now, we shall discuss about this update rule a little bit if the actual output that is y is same as the predicted output Y hat, then there is no way to update However, if the actual output and the predicted output differ from each other, then there is a word update, okay. However, if the algorithm is not converged, then we increase the iteration count by one and repeat the whole process. Okay, here he typed in the wind upgrade rule is called a learning rate and it is usually used defined and it is easily a small positive number perceptron decision boundary is a hyperplane defined by W transpose X equals to zero consider binary classification problem data set is shown in the figure, no random initialization of weight vector at the starting of the learning algorithm randomly feeds a hyperplane as shown beside as we keep updating the weights using the learning tune, the decision boundary keeps on changing till all the training data points fall correctly in the either side of the decision boundary Okay, does perceptron acts as a linear classifier and can classify linearly separable data set with high accuracy the perceptron with sigmoid activation function, not we have used hard limiter or the signal function as activation function.

However, we can choose sigmoid function as our activation function as well. And in this case, the single noodle call response exactly to the input output mapping defined by the logistic regression okay. So, this is how the sigmoid function looks like and with the sigmoid activation function perceptron is nothing but a logistic regression. So, limitations of perceptron the limitation of perceptron is not able to classify the nonlinear data set. So, here we shall demonstrate the problem using something called the XOR problem. So, this is XOR data set Okay, the input one and input two that means this x and y if both are zero then the output is zero.

If the input is zero and input two is one, then the output is one if input one is one and input two is one, then the output is zero. And if input one is one and input two is zero, then output is again one. Okay, so that is how they saw data set looks like. And this is our perceptron. Note that the perceptron works fairly well for linearly separable data. said it failed to solve the exact problem and other nonlinear data sets because there is no straight line that can feed the data points or that can form a decision boundary between the two classes.

So, let's say this is a possible decision boundary, but however, we can see that there is one data point of Class Zero in this side of the decision boundary and another data point of Class Zero in this side of the decision boundary, okay. So, this single region boundary is not able to classify this data set properly. Similarly, if we place the decision boundary anywhere in this plane will be unable to solve this problem, okay. So, this is known as exhort problem or non linearly separable data set problem, and this has traveled the AI community for many years until in 1986, Geoffrey Hinton, and his to come up with the concept of backpropagation learning algorithm, which eventually solve the problem using multi layer perceptron. In the next video we shall discuss about multi layer perceptron. So see you in the next lecture.

Thank you

Sign Up

Share

Share with friends, get 20% off
Invite your friends to LearnDesk learning marketplace. For each purchase they make, you get 20% off (upto $10) on your next purchase.