Lecture Slides for INTRODUCTION TO MACHINE LEARNING 3
- Slides: 26
Lecture Slides for INTRODUCTION TO MACHINE LEARNING 3 RD EDITION ETHEM ALPAYDIN © The MIT Press, 2014 alpaydin@boun. edu. tr http: //www. cmpe. boun. edu. tr/~ethem/i 2 ml 3 e
CHAPTER 12: LOCAL MODELS
Introduction 3 Divide the input space into local regions and learn simple (constant/linear) models in each patch Unsupervised: Competitive, online clustering Supervised: Radial-basis functions, mixture of experts
Competitive Learning 4
Winner-take-all network 5
Adaptive Resonance Theory Incremental; add a new cluster if not covered; defined by vigilance, ρ (Carpenter and Grossberg, 1988) 6
Self-Organizing Maps 7 Units have a neighborhood defined; mi is “between” mi-1 and mi+1, and are all updated together (Kohonen, 1990) One-dim map:
Radial-Basis Functions Locally-tuned units: 8
9 Local vs Distributed Representation
Training RBF 10 Hybrid learning: � First layer centers and spreads: Unsupervised k-means � Second layer weights: Supervised gradient-descent Fully supervised (Broomhead and Lowe, 1988; Moody and Darken, 1989)
Regression 11
Classification 12
Rules and Exceptions 13 Exceptions Default rule
Rule-Based Knowledge 14 Incorporation of prior knowledge (before training) Rule extraction (after training) (Tresp et al. , 1997) Fuzzy membership functions and fuzzy rules
Normalized Basis Functions 15
Competitive Basis Functions 16 Mixture model:
Regression 17
Classification 18
EM for RBF (Supervised EM) 19 E-step: M-step:
Learning Vector Quantization 20 H units per class prelabeled (Kohonen, 1990) Given x, mi is the closest: x mi mj
Mixture of Experts In RBF, each local fit is a constant, wih, second layer weight In Mo. E, each local fit is a linear function of x, a local expert: (Jacobs et al. , 1991) 21
Mo. E as Models Combined Radial gating: Softmax gating: 22
Cooperative Mo. E 23 Regression
Competitive Mo. E: Regression 24
Competitive Mo. E: Classification 25
Hierarchical Mixture of Experts 26 Tree of Mo. E where each Mo. E is an expert in a higher-level Mo. E Soft decision tree: Takes a weighted (gating) average of all leaves (experts), as opposed to using a single path and a single leaf Can be trained using EM (Jordan and Jacobs, 1994)
- Machine learning introduction slides
- Introduction to machine learning slides
- Machine learning lecture notes
- Ethem alpaydin
- A small child slides down the four frictionless slides
- Change in energy quick check
- Principles of economics powerpoint lecture slides
- Business communication lecture slides
- 01:640:244 lecture notes - lecture 15: plat, idah, farad
- Concept learning task in machine learning
- Analytical learning in machine learning
- Pac learning model in machine learning
- Pac learning model in machine learning
- Inductive and analytical learning in machine learning
- Inductive and analytical learning
- Instance based learning in machine learning
- Inductive learning machine learning
- First order rule learning in machine learning
- Lazy and eager learning
- Deep learning vs machine learning
- Introduction to machine learning ethem alpaydin
- Machine learning andrew ng
- Introduction to machine learning andrew ng
- Mike mozer
- Introduction to machine learning ethem alpaydin
- Introduction to machine learning and data mining
- Lstm lecture