Implementation of Decision tree classifier using python

Machine Learning Using Python Other classification Algorithms
7 minutes
Share the link to this page
Copied
  Completed
You need to have access to the item to view this lesson.
One-time Fee
$69.99
List Price:  $99.99
You save:  $30
€65.39
List Price:  €93.41
You save:  €28.02
£55.92
List Price:  £79.90
You save:  £23.97
CA$96.01
List Price:  CA$137.16
You save:  CA$41.15
A$107.15
List Price:  A$153.08
You save:  A$45.93
S$95.13
List Price:  S$135.90
You save:  S$40.77
HK$547.14
List Price:  HK$781.66
You save:  HK$234.52
CHF 63.86
List Price:  CHF 91.23
You save:  CHF 27.37
NOK kr775.40
List Price:  NOK kr1,107.76
You save:  NOK kr332.36
DKK kr487.78
List Price:  DKK kr696.86
You save:  DKK kr209.07
NZ$118.01
List Price:  NZ$168.60
You save:  NZ$50.58
د.إ257.06
List Price:  د.إ367.24
You save:  د.إ110.18
৳7,680.49
List Price:  ৳10,972.60
You save:  ৳3,292.11
₹5,842.03
List Price:  ₹8,346.11
You save:  ₹2,504.08
RM332.86
List Price:  RM475.54
You save:  RM142.67
₦86,437.65
List Price:  ₦123,487.65
You save:  ₦37,050
₨19,491.96
List Price:  ₨27,846.85
You save:  ₨8,354.89
฿2,586.09
List Price:  ฿3,694.58
You save:  ฿1,108.48
₺2,265.39
List Price:  ₺3,236.41
You save:  ₺971.02
B$363.53
List Price:  B$519.35
You save:  B$155.82
R1,302.64
List Price:  R1,861
You save:  R558.35
Лв127.90
List Price:  Лв182.73
You save:  Лв54.82
₩96,270.48
List Price:  ₩137,535.16
You save:  ₩41,264.67
₪262.29
List Price:  ₪374.71
You save:  ₪112.42
₱4,033.94
List Price:  ₱5,763.02
You save:  ₱1,729.07
¥10,867.12
List Price:  ¥15,525.12
You save:  ¥4,658
MX$1,187.12
List Price:  MX$1,695.96
You save:  MX$508.84
QR254.93
List Price:  QR364.20
You save:  QR109.27
P994.08
List Price:  P1,420.18
You save:  P426.09
KSh9,360.69
List Price:  KSh13,372.99
You save:  KSh4,012.30
E£3,358.63
List Price:  E£4,798.26
You save:  E£1,439.62
ብር4,003.77
List Price:  ብር5,719.92
You save:  ብር1,716.15
Kz58,546.63
List Price:  Kz83,641.63
You save:  Kz25,095
CLP$67,216.99
List Price:  CLP$96,028.39
You save:  CLP$28,811.40
CN¥506.70
List Price:  CN¥723.89
You save:  CN¥217.19
RD$4,073.53
List Price:  RD$5,819.58
You save:  RD$1,746.04
DA9,418.34
List Price:  DA13,455.35
You save:  DA4,037.01
FJ$158.31
List Price:  FJ$226.17
You save:  FJ$67.86
Q543.96
List Price:  Q777.12
You save:  Q233.16
GY$14,650.29
List Price:  GY$20,929.88
You save:  GY$6,279.59
ISK kr9,815.39
List Price:  ISK kr14,022.59
You save:  ISK kr4,207.20
DH707.71
List Price:  DH1,011.06
You save:  DH303.35
L1,237.78
List Price:  L1,768.33
You save:  L530.55
ден4,025.24
List Price:  ден5,750.59
You save:  ден1,725.35
MOP$563.96
List Price:  MOP$805.69
You save:  MOP$241.73
N$1,304.33
List Price:  N$1,863.42
You save:  N$559.08
C$2,570.38
List Price:  C$3,672.13
You save:  C$1,101.75
रु9,397.27
List Price:  रु13,425.24
You save:  रु4,027.97
S/263.43
List Price:  S/376.35
You save:  S/112.91
K270.11
List Price:  K385.89
You save:  K115.77
SAR262.49
List Price:  SAR375.01
You save:  SAR112.51
ZK1,873.89
List Price:  ZK2,677.10
You save:  ZK803.21
L325.37
List Price:  L464.84
You save:  L139.46
Kč1,643.47
List Price:  Kč2,347.91
You save:  Kč704.44
Ft25,458.03
List Price:  Ft36,370.18
You save:  Ft10,912.14
SEK kr764.90
List Price:  SEK kr1,092.76
You save:  SEK kr327.86
ARS$61,327.27
List Price:  ARS$87,614.14
You save:  ARS$26,286.87
Bs483.57
List Price:  Bs690.85
You save:  Bs207.27
COP$273,218.78
List Price:  COP$390,329.27
You save:  COP$117,110.49
₡35,710.66
List Price:  ₡51,017.42
You save:  ₡15,306.75
L1,733.65
List Price:  L2,476.75
You save:  L743.09
₲524,442.73
List Price:  ₲749,236.02
You save:  ₲224,793.28
$U2,683.09
List Price:  $U3,833.15
You save:  $U1,150.06
zł283.24
List Price:  zł404.64
You save:  zł121.40
Already have an account? Log In

Transcript

Hello, everyone. Welcome to the course of machine learning with Python. In this video, we shall learn how to implement decision tree classifier in Python. So we shall use it as data set. So first we'll be importing the necessary libraries. So from SK learn dot data sets, we'll be importing load underscore IDs, we'll be importing NumPy as NP, we'll be importing random.

And we'll be importing matplotlib.pi plot as PLT. So let's go ahead and run this particular cell. Now. We shall be importing the data sets into the matrix X and the vector y. So x is basically the feature matrix and y is the label vector. So x comma y will be equals to load underscore IDs, return underscore capital X underscore small y is equals to true.

So here is the feature set and why is that so let's go ahead and run this particular So, now we shall be dividing the data into train and test. So, here we shall take 75% data as training sample and 25% data as test sample. So, from SK learn dot model selection, we should be important train test split. So, extreme comma x test comma y 10 comma y test will be equals to create tests pleat, x comma y and here we have specified test size is equal to 0.25. That means 25% of the data set will be randomly picked up and will be considered as a test sample. So, let's go ahead and run this particular cell.

Now, the shape of the training data set is 112 comma four that means it contains hundred and 12 training samples and four features or four attributes. Similarly, x test dot shape will give us 38 test samples Each test sample contains four features or attributes. Now, we shall be importing the model and feed the data set into the model. So, from a schema dot three, we shall be importing decision tree classifier, please note the camel font d capital T capital and C capital of the decision tree classifier. So, we will be importing this class CLF is the opposite of the decision tree classifier and we'll be fitting x train and y train. Okay.

So, there are so many attributes of this decision tree classifier. So let's look at the important one. So criteria by default is Gini. However, it can mention criterion equals to entropy or criterion equals to some other index in order to specify anything other than ci but by default, the criteria of speed is decided by Gini index we can also specify maximum depth maximum features maximum leaf nodes etc okay. So, there are so many attributes that we can specify over here however, we are choosing criterion Gini which is by default the attribute of the decision tree classifier object okay. So, now we shall be testing the model.

So why underscore predict is equals to this CLF dot predict exist. Now, we shall be evaluating the models so from SK learn dot matrix, we'll be importing confusion matrix. So our confusion matrix here is equals to confusion matrix. The test data set are the actual or the ground truth values which is why underscore test comma y underscore credit which is our predicted values. So, this is our confusion matrix. So, correct is equals to NP dot Chris of the confusion matrix that means it is the sum of the diagonal elements of the confusion matrix and the total will be nothing but the total number of elements in the validation matrix.

So, correctly identified is nothing but the credits and the total is nothing but this total value of this entire confusion matrix. So, here the correctly identified 38 and total is also 38. So, all the test samples has been correctly identified. So, what will be the accuracy, so, surely the accuracy will be hundred percent. Now from SK learn dot matrix we'll be importing classification from the school report. And we shall be printing this classification underscore report which is a function which accepts the argument whitest and white predict.

So, let's go ahead and run this particular cell. So we can see that the appreciation of class zero is one the call is one and If one score is one, similarly the precision of class one and two is all one. And accuracy is also very high, which is one. Now we can visualize the decision tree. And for these we need to install graph trees. The complete documentation on how to install graph fees can be found here.

If we click on this particular link, you will be redirected to this installation site of graph fees. So the comment is PIP installed lovelies. If you open Anaconda prompt and type the install graph with the graphics will be instructed. So after the installation of the Gulf ease, you can import graphviz and from SK learn dot three you can import export underscore breakfast. So let's go ahead and do this imports. Now from ipython dot display input display and this display function accepts these graphics cards.

Source within bracket, export route fees within bracket our model CLS. So let's go ahead and run this particular cell and as you can see, this is our decision tree created okay as you can see here all the splitting criteria are changing okay. So g equals to zero means it is a pure or homogeneous node okay. So this is how the entire decision tree has been built. So it has a very nice interpretability So, x within bracket three means it is nothing but the fourth feature, since the value of the fourth feature is less than point eight. Okay we'll be checking on this criteria it could be either true or false.

If it is true then We'll be going with Class Zero. If it is false then again we check on the feature value for whether it is less than or equals to 1.75 or not. So, that is how the entire decision tree has been built. Okay. So, I will recommend to change the criteria to entropy and see how the decision tree behaves. So, in the next video we shall explore another classifier known as random forest classifier.

So, see you in the next lecture. Thank you

Sign Up

Share

Share with friends, get 20% off
Invite your friends to LearnDesk learning marketplace. For each purchase they make, you get 20% off (upto $10) on your next purchase.