Dimensionality and its problem. Linear algebra review: Eigen Value Decomposition

Machine Learning Using Python Dimensionality Reduction
9 minutes
Share the link to this page
Copied
  Completed
You need to have access to the item to view this lesson.
One-time Fee
$69.99
List Price:  $99.99
You save:  $30
€59.73
List Price:  €85.34
You save:  €25.60
£51.80
List Price:  £74.01
You save:  £22.20
CA$96.84
List Price:  CA$138.36
You save:  CA$41.51
A$106.75
List Price:  A$152.51
You save:  A$45.75
S$89.95
List Price:  S$128.50
You save:  S$38.55
HK$545.64
List Price:  HK$779.53
You save:  HK$233.88
CHF 55.85
List Price:  CHF 79.79
You save:  CHF 23.94
NOK kr703.17
List Price:  NOK kr1,004.57
You save:  NOK kr301.40
DKK kr445.91
List Price:  DKK kr637.04
You save:  DKK kr191.13
NZ$118.73
List Price:  NZ$169.63
You save:  NZ$50.89
د.إ257.03
List Price:  د.إ367.21
You save:  د.إ110.17
৳8,495.52
List Price:  ৳12,136.98
You save:  ৳3,641.45
₹6,172.17
List Price:  ₹8,817.76
You save:  ₹2,645.59
RM295.67
List Price:  RM422.40
You save:  RM126.73
₦107,084.70
List Price:  ₦152,984.70
You save:  ₦45,900
₨19,808.40
List Price:  ₨28,298.93
You save:  ₨8,490.52
฿2,243.26
List Price:  ฿3,204.79
You save:  ฿961.53
₺2,883.76
List Price:  ₺4,119.83
You save:  ₺1,236.07
B$378.92
List Price:  B$541.35
You save:  B$162.42
R1,231.03
List Price:  R1,758.69
You save:  R527.66
Лв116.90
List Price:  Лв167.02
You save:  Лв50.11
₩97,041.13
List Price:  ₩138,636.13
You save:  ₩41,595
₪234.19
List Price:  ₪334.57
You save:  ₪100.38
₱3,968.43
List Price:  ₱5,669.43
You save:  ₱1,701
¥10,316.87
List Price:  ¥14,739.02
You save:  ¥4,422.15
MX$1,309.87
List Price:  MX$1,871.33
You save:  MX$561.45
QR255.14
List Price:  QR364.50
You save:  QR109.36
P938.51
List Price:  P1,340.79
You save:  P402.27
KSh9,032.87
List Price:  KSh12,904.65
You save:  KSh3,871.78
E£3,400.11
List Price:  E£4,857.51
You save:  E£1,457.40
ብር9,985.48
List Price:  ብር14,265.58
You save:  ብር4,280.10
Kz64,180.83
List Price:  Kz91,690.83
You save:  Kz27,510
CLP$67,522.15
List Price:  CLP$96,464.35
You save:  CLP$28,942.20
CN¥499.22
List Price:  CN¥713.21
You save:  CN¥213.98
RD$4,414.63
List Price:  RD$6,306.89
You save:  RD$1,892.25
DA9,089.11
List Price:  DA12,985
You save:  DA3,895.89
FJ$157.67
List Price:  FJ$225.25
You save:  FJ$67.58
Q535.38
List Price:  Q764.86
You save:  Q229.48
GY$14,604.61
List Price:  GY$20,864.62
You save:  GY$6,260.01
ISK kr8,549.27
List Price:  ISK kr12,213.77
You save:  ISK kr3,664.50
DH634.40
List Price:  DH906.33
You save:  DH271.92
L1,171.66
List Price:  L1,673.87
You save:  L502.21
ден3,674.28
List Price:  ден5,249.20
You save:  ден1,574.92
MOP$561.02
List Price:  MOP$801.49
You save:  MOP$240.47
N$1,234.02
List Price:  N$1,762.97
You save:  N$528.94
C$2,569.13
List Price:  C$3,670.35
You save:  C$1,101.21
रु9,858.12
List Price:  रु14,083.64
You save:  रु4,225.51
S/245.88
List Price:  S/351.28
You save:  S/105.39
K291.36
List Price:  K416.25
You save:  K124.88
SAR262.47
List Price:  SAR374.98
You save:  SAR112.50
ZK1,666.62
List Price:  ZK2,380.99
You save:  ZK714.36
L303.21
List Price:  L433.17
You save:  L129.96
Kč1,456.45
List Price:  Kč2,080.74
You save:  Kč624.28
Ft23,442.21
List Price:  Ft33,490.30
You save:  Ft10,048.09
SEK kr657.57
List Price:  SEK kr939.43
You save:  SEK kr281.85
ARS$95,110.21
List Price:  ARS$135,877.55
You save:  ARS$40,767.34
Bs482.36
List Price:  Bs689.11
You save:  Bs206.75
COP$278,383.76
List Price:  COP$397,708.14
You save:  COP$119,324.37
₡35,369.65
List Price:  ₡50,530.25
You save:  ₡15,160.59
L1,828.90
List Price:  L2,612.83
You save:  L783.92
₲504,728.30
List Price:  ₲721,071.34
You save:  ₲216,343.03
$U2,810.15
List Price:  $U4,014.67
You save:  $U1,204.52
zł253.83
List Price:  zł362.63
You save:  zł108.80
Already have an account? Log In

Transcript

Hello everyone, welcome to the course of machine learning with Python. In this video, we shall learn about dimensionality reduction. So, problem with very high dimensional data. This is also known as curse of dimensionality data at very high dimensional space becomes pass for highly sparse data machine learning algorithm does not give satisfactory results always this is simply because the available training data set might not have every possible combinations of the features. Hence, machine learning algorithm struggles to find any meaningful relationship among the features and the target variables. For example, consider a data set which has the mini categorical features with key many possible values, then, the number of possible combinations of the data point is k to the power d, which grows exponentially as the dimension increases.

Our training data set may not contain all sorts of possible combinations. So, what is the workaround of this problem? dimensionality reduction to find a suitable representation of the data in one dimensional space, which will eventually help us for machine learning, such as classification, clustering, regression and data visualization. There are several types of dimensionality reduction techniques. However, all those techniques can be broadly classified into two categories feature selection technique and feature extraction techniques. So, feature selection techniques, we select few features based on certain criteria, and thus get rid of redundant or less important features to make our data set less bulky.

Note that feature selection depends on the data set as well as on the task we want to perform with the data set feature extraction techniques we project or original data set which belongs to a high dimensional space to a low dimensional space such that the inherent information inside the data is not lost. This projection could be either linear or nonlinear. Based on which we have linear or nonlinear feature extraction techniques a very simple example of feature selection let's say a school student in a particular school can have following features x one name of the student x two contact details of the student x three class export roll number x five marks in mathematics, x six marks in language group x seven marks in social science group x eight marks in natural science group x nine height extent bodyweight excellent blood group etc. Each student can be represented by a vector such as x equals two x one x two x three whole transpose y transpose because we are considering column vectors.

Now, suppose our job is to find the suitable set of students for Science Olympiad which features play the important role in selecting the students for Science Olympiad. So, obviously not all of these features are important in this case, the most important features are marks in mathematics and marks in natural science cool. Now, let us want to select the students for sports to numb it. In this case, we have to select the students based on the height weight fitness and skill in that particular sports. So, based on the problems in hand, we can discuss some of the features and consider others. So, what is the objective of feature selection to opt in a set of features key out of in many features in the data set key usual is done in such that the selected features are based suitable for a required task D. Now, the actual value of k depends on many things such as the task itself.

That is if it is for visualization, then maybe we'll go for two or three features, the amount of the data that machine can handle also determines the value of k number of available training data points so that the data set becomes less parts in the reduced feature space. Excellent. QRA determines how much should be the value of key. That means in the modified he just pays how many features should present. Now how to select the features, we can use domain knowledge and experience to see a set of basic features suitable for a particular task exhaustive sets select the best possible set of K mini features out of all possible combinations of the features, these methods might work well if the number of features in the data set is not very high. Otherwise this method does not scale well.

If the number of features in the data set is too high heuristic search in this case, we keep on adding features which will provide based value of some criterion function. We stopped the algorithm till the required number of features are selected. Other such strategies such as randomized search are also used for feature selection. Now feature extraction let the original data point x belong to a higher dimension. Space capital D, we want to reduce the dimensionality of the data set into small d small d is less than capital D. in feature extraction, we shall project the data set from the higher dimension capital D to the lower dimension small d using some function which maps the higher dimensional space to learn what dimensional space such that the information contained inside the data set is preserved significantly. Here we are not selecting or discarding any particular feature.

The whole idea is to project the data set into some smaller dimensional subspace such that the interrelationship of the data points are preserved late capital X is our original data set in dimension capital D and small x vector is a data point late in the transform space, the dimension is small d and x prime is the corresponding data point that means, f of x vector is nothing but Vector prime if x one vector and x two vector A to neighboring points in the origin data set x then in the transform Space X one prime vector and x two prime vector should also be neighbors. There are various feature extraction techniques, few techniques employ linear transformation or projection of the data set for example, principal component analysis and linear discriminant analysis some techniques involve nonlinear transformation of the data set, for example, locally linear embedding or kernel PCA.

Now, let's look at some mathematical prerequisite Egon values and eigenvectors of a square matrix matrix vector multiplication consider the vector x one x one is denoted by two comma three in these two dimensional space. Now, the matrix E which is written over here if it is post quantified with x one then we are Another vector x one prime. So x one prime is equal to E x one, we can see that x one prime after computing this multiplication, it is nothing but five, seven. Okay. So in general, we can say that multiplying a vector with a matrix changes the direction and the magnitude of the vector. However, for a given square matrix, there are some factors whose direction remain unchanged.

After multiplying with the matrix, consider the fixed term x two which is nothing but one one. So this is our x two vector. If we multiply a with x two, then the transform factor x two prime is nothing but three three, which is nothing but three times one one, okay. So we can see that the magnitude of the victor has been changed after multiplying the vector with the matrix, but the direction remains unchanged. That is the transformed vector x two prime is nothing but a scalar multiplication of the original vector x to know for a given square matrix E, if we find a vector v, such that the mighty plane a with vector v does not change the direction of the vector v, but scales it, then the vector v is known as the eigenvector of a and the scalar value by which it will be scaled after being multiplied with a is known as the corresponding Eigen value.

Mathematically, if e times V is equals to lambda times p, then lambda is called the eigenvalue and v is the Eigen vector of matrix A note that lambda here is a scalar. Now, e times v is equal to lambda times V, we can write this as e minus lambda times I hold times P is equals to zero. Here AI is nothing but the identity matrix of the same Order of matrix A for non trivial solution of the eigenvectors determining value of A minus lambda I should be equals to zero, the above equation is a polynomial of lambda of degree same as the order of the matrix A. This equation is also known as the character characteristic equation of matrix A, we can calculate the eigenvalues by solving the characteristic equation Putting the value of the Eigen value one can calculate the corresponding Eigen vectors. In the next video, we shall learn about principal component analysis.

So, see you in the next lecture. Thank you

Sign Up

Share

Share with friends, get 20% off
Invite your friends to LearnDesk learning marketplace. For each purchase they make, you get 20% off (upto $10) on your next purchase.