ECE 5984 Introduction to Machine Learning Topics DecisionClassification

  • Slides: 66
Download presentation
ECE 5984: Introduction to Machine Learning Topics: – Decision/Classification Trees Readings: Murphy 16. 1

ECE 5984: Introduction to Machine Learning Topics: – Decision/Classification Trees Readings: Murphy 16. 1 -16. 2; Hastie 9. 2 Dhruv Batra Virginia Tech

Project Proposals Graded • Mean 3. 6/5 = 72% (C) Dhruv Batra 2

Project Proposals Graded • Mean 3. 6/5 = 72% (C) Dhruv Batra 2

Administrativia • Project Mid-Sem Spotlight Presentations – Friday: 5 -7 pm, 3 -5 pm

Administrativia • Project Mid-Sem Spotlight Presentations – Friday: 5 -7 pm, 3 -5 pm Whittemore 654 457 A – 5 slides (recommended) – 4 minute time (STRICT) + 1 -2 min Q&A – – (C) Dhruv Batra Tell the class what you’re working on Any results yet? Problems faced? Upload slides on Scholar 3

Recap of Last Time (C) Dhruv Batra 4

Recap of Last Time (C) Dhruv Batra 4

Convolution Explained • http: //setosa. io/ev/image-kernels/ • https: //github. com/bruckner/deep. Viz (C) Dhruv Batra

Convolution Explained • http: //setosa. io/ev/image-kernels/ • https: //github. com/bruckner/deep. Viz (C) Dhruv Batra 5

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 6

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 6

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 7

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 7

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 8

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 8

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 9

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 9

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 10

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 10

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 11

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 11

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 12

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 12

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 13

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 13

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 14

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 14

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 15

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 15

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 16

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 16

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 17

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 17

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 18

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 18

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 19

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 19

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 20

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 20

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 21

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 21

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 22

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 22

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 23

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 23

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 24

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 24

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 25

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 25

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 26

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 26

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 27

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 27

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 28

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 28

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 29

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 29

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 30

(C) Dhruv Batra Slide Credit: Marc'Aurelio Ranzato 30

Convolutional Nets • Example: – http: //yann. lecun. com/exdb/lenet/index. html (C) Dhruv Batra Image

Convolutional Nets • Example: – http: //yann. lecun. com/exdb/lenet/index. html (C) Dhruv Batra Image Credit: Yann Le. Cun, Kevin Murphy 31

Visualizing Learned Filters (C) Dhruv Batra Figure Credit: [Zeiler & Fergus ECCV 14] 32

Visualizing Learned Filters (C) Dhruv Batra Figure Credit: [Zeiler & Fergus ECCV 14] 32

Visualizing Learned Filters (C) Dhruv Batra Figure Credit: [Zeiler & Fergus ECCV 14] 33

Visualizing Learned Filters (C) Dhruv Batra Figure Credit: [Zeiler & Fergus ECCV 14] 33

Visualizing Learned Filters (C) Dhruv Batra Figure Credit: [Zeiler & Fergus ECCV 14] 34

Visualizing Learned Filters (C) Dhruv Batra Figure Credit: [Zeiler & Fergus ECCV 14] 34

Addressing non-linearly separable data – Option 1, non-linear features • Choose non-linear features, e.

Addressing non-linearly separable data – Option 1, non-linear features • Choose non-linear features, e. g. , – Typical linear features: w 0 + i wi xi – Example of non-linear features: • Degree 2 polynomials, w 0 + i wi xi + ij wij xi xj • Classifier hw(x) still linear in parameters w – As easy to learn – Data is linearly separable in higher dimensional spaces – Express via kernels (C) Dhruv Batra Slide Credit: Carlos Guestrin 35

Addressing non-linearly separable data – Option 2, non-linear classifier • Choose a classifier hw(x)

Addressing non-linearly separable data – Option 2, non-linear classifier • Choose a classifier hw(x) that is non-linear in parameters w, e. g. , – Decision trees, neural networks, … • More general than linear classifiers • But, can often be harder to learn (nonconvex/concave optimization required) • Often very useful (outperforms linear classifiers) • In a way, both ideas are related (C) Dhruv Batra Slide Credit: Carlos Guestrin 36

New Topic: Decision Trees (C) Dhruv Batra 37

New Topic: Decision Trees (C) Dhruv Batra 37

Synonyms • Decision Trees • Classification and Regression Trees (CART) • Algorithms for learning

Synonyms • Decision Trees • Classification and Regression Trees (CART) • Algorithms for learning decision trees: – ID 3 – C 4. 5 • Random Forests – Multiple decision trees (C) Dhruv Batra 38

Decision Trees • Demo – http: //www. cs. technion. ac. il/~rani/Loc. Boost/ (C) Dhruv

Decision Trees • Demo – http: //www. cs. technion. ac. il/~rani/Loc. Boost/ (C) Dhruv Batra 39

Pose Estimation • Random Forests! – Multiple decision trees – http: //youtu. be/HNkb. G

Pose Estimation • Random Forests! – Multiple decision trees – http: //youtu. be/HNkb. G 3 Ks. Y 84 (C) Dhruv Batra 40

Slide Credit: Pedro Domingos, Tom Mitchel, Tom Dietterich

Slide Credit: Pedro Domingos, Tom Mitchel, Tom Dietterich

Slide Credit: Pedro Domingos, Tom Mitchel, Tom Dietterich

Slide Credit: Pedro Domingos, Tom Mitchel, Tom Dietterich

Slide Credit: Pedro Domingos, Tom Mitchel, Tom Dietterich

Slide Credit: Pedro Domingos, Tom Mitchel, Tom Dietterich

Slide Credit: Pedro Domingos, Tom Mitchel, Tom Dietterich

Slide Credit: Pedro Domingos, Tom Mitchel, Tom Dietterich

A small dataset: Miles Per Gallon Suppose we want to predict MPG 40 Records

A small dataset: Miles Per Gallon Suppose we want to predict MPG 40 Records From the UCI repository (thanks to Ross Quinlan) (C) Dhruv Batra Slide Credit: Carlos Guestrin 45

A Decision Stump (C) Dhruv Batra Slide Credit: Carlos Guestrin 46

A Decision Stump (C) Dhruv Batra Slide Credit: Carlos Guestrin 46

The final tree (C) Dhruv Batra Slide Credit: Carlos Guestrin 47

The final tree (C) Dhruv Batra Slide Credit: Carlos Guestrin 47

Comments • Not all features/attributes need to appear in the tree. • A features/attribute

Comments • Not all features/attributes need to appear in the tree. • A features/attribute Xi may appear in multiple branches. • On a path, no feature may appear more than once. – Not true for continuous features. We’ll see later. • Many trees can represent the same concept • But, not all trees will have the same size! – e. g. , Y = (A^B) ( A^C) (C) Dhruv Batra (A and B) or (not A and C) 48

Learning decision trees is hard!!! • Learning the simplest (smallest) decision tree is an

Learning decision trees is hard!!! • Learning the simplest (smallest) decision tree is an NP-complete problem [Hyafil & Rivest ’ 76] • Resort to a greedy heuristic: – Start from empty decision tree – Split on next best attribute (feature) – Recurse • “Iterative Dichotomizer” (ID 3) • C 4. 5 (ID 3+improvements) (C) Dhruv Batra Slide Credit: Carlos Guestrin 49

Recursion Step Records in which cylinders =4 Take the Original Dataset. . And partition

Recursion Step Records in which cylinders =4 Take the Original Dataset. . And partition it according to the value of the attribute we split on Records in which cylinders =5 Records in which cylinders =6 Records in which cylinders =8 (C) Dhruv Batra Slide Credit: Carlos Guestrin 50

Recursion Step Build tree from These records. . Records in which cylinders =4 (C)

Recursion Step Build tree from These records. . Records in which cylinders =4 (C) Dhruv Batra Records in which cylinders =5 Records in which cylinders =6 Slide Credit: Carlos Guestrin Records in which cylinders =8 51

Second level of tree Recursively build a tree from the seven records in which

Second level of tree Recursively build a tree from the seven records in which there are four cylinders and the maker was based in Asia (C) Dhruv Batra Slide Credit: Carlos Guestrin (Similar recursion in the other cases) 52

The final tree (C) Dhruv Batra Slide Credit: Carlos Guestrin 53

The final tree (C) Dhruv Batra Slide Credit: Carlos Guestrin 53

Choosing a good attribute (C) Dhruv Batra Slide Credit: Carlos Guestrin X 1 T

Choosing a good attribute (C) Dhruv Batra Slide Credit: Carlos Guestrin X 1 T T T X 2 T F T Y T T F F F T F T T F F F 54

Measuring uncertainty • Good split if we are more certain about classification after split

Measuring uncertainty • Good split if we are more certain about classification after split – Deterministic good (all true or all false) – Uniform distribution bad 1 P(Y=F | X 1= T) = 0 P(Y=T | X 1= T) = 1 0. 5 0 F T 1 P(Y=F | X 2=F) = 1/2 P(Y=T | X 2=F) = 1/2 0. 5 0 (C) Dhruv Batra 55

Entropy H(X) of a random variable Y More uncertainty, more entropy! Information Theory interpretation:

Entropy H(X) of a random variable Y More uncertainty, more entropy! Information Theory interpretation: H(Y) is the expected number of bits needed to encode a randomly drawn value of Y (under most efficient code) (C) Dhruv Batra Slide Credit: Carlos Guestrin 56

Information gain • Advantage of attribute – decrease in uncertainty – Entropy of Y

Information gain • Advantage of attribute – decrease in uncertainty – Entropy of Y before you split – Entropy after split • Weight by probability of following each branch, i. e. , normalized number of records • Information gain is difference – (Technically it’s mutual information; but in this context also referred to as information gain) (C) Dhruv Batra Slide Credit: Carlos Guestrin 57

Learning decision trees • Start from empty decision tree • Split on next best

Learning decision trees • Start from empty decision tree • Split on next best attribute (feature) – Use, for example, information gain to select attribute – Split on • Recurse (C) Dhruv Batra Slide Credit: Carlos Guestrin 58

Suppose we want to predict MPG Look at all the information gains… (C) Dhruv

Suppose we want to predict MPG Look at all the information gains… (C) Dhruv Batra Slide Credit: Carlos Guestrin 59

When do we stop? (C) Dhruv Batra 60

When do we stop? (C) Dhruv Batra 60

Base Case One Don’t split a node if all matching records have the same

Base Case One Don’t split a node if all matching records have the same output value (C) Dhruv Batra Slide Credit: Carlos Guestrin 61

Base Case Two: No attributes can distinguish (C) Dhruv Batra Slide Credit: Carlos Guestrin

Base Case Two: No attributes can distinguish (C) Dhruv Batra Slide Credit: Carlos Guestrin Don’t split a node if none of the attributes can create multiple nonempty children 62

Base Cases • Base Case One: If all records in current data subset have

Base Cases • Base Case One: If all records in current data subset have the same output then don’t recurse • Base Case Two: If all records have exactly the same set of input attributes then don’t recurse (C) Dhruv Batra Slide Credit: Carlos Guestrin 63

Base Cases: An idea • Base Case One: If all records in current data

Base Cases: An idea • Base Case One: If all records in current data subset have the same output then don’t recurse • Base Case Two: If all records have exactly the same set of input attributes then don’t recurse Proposed Base Case 3: If all attributes have zero information gain then don’t recurse • Is this a good idea? (C) Dhruv Batra Slide Credit: Carlos Guestrin 64

The problem with Base Case 3 y = a XOR b The information gains:

The problem with Base Case 3 y = a XOR b The information gains: (C) Dhruv Batra The resulting decision tree: Slide Credit: Carlos Guestrin 65

If we omit Base Case 3: y = a XOR b The resulting decision

If we omit Base Case 3: y = a XOR b The resulting decision tree: (C) Dhruv Batra Slide Credit: Carlos Guestrin 66