KTH ROYAL INSTITUTE OF TECHNOLOGY Machine learning V
- Slides: 24
KTH ROYAL INSTITUTE OF TECHNOLOGY Machine learning V Artificial Neural Networks
On the Shoulder of Giants Much of the material in this slide set is based upon: - ”Automated Learning techniques in Power Systems” by L. Wehenkel, Université Liege - ”Machine Learning” course by Andrew Ng, Associate Professor, Stanford University 2
Contents Feedforward Artificial Neural Networks Back propagation in Artificial Neural Networks 3
Regression VS classification Regression (linear and polynomial): for prediction Classification: - Coping with over/under fitting Source: ”Machine learning ” course, Andrew Ng 4
Linear regression: bias unit and normalization Source: ”Machine learning ” course, Andrew Ng 5
Classification (logistic regression) - Non-linear perceptron normally uses a threhold function for the output, to limit the extreme values. Source: ”Machine learning ” course, Andrew Ng 6
Example (multi-layer perceptron): XNOR function Source: ”Machine learning ” course, Andrew Ng 7
Multi Layer Perceptrons (MLP) - A network of interconnected Perceptrons in several layers - First layer recives input, forwards to second layer etc. - Normally one hidden layer is sufficient to create good mappings Source: Automatic Learning techniques in Power Systems, L. Wehenkel 8
Contents Feedforward Artificial Neural Networks Back propagation in Artificial Neural Networks 9
How to tune the weights? (Backpropagation algorithm)
How to tune the weights? (Backpropagation algorithm)
How to tune the weights? (Learning)
Cost function and Regularization 1) Cost function for Linear regression
Cost function and Regularization 2) Cost function for classification (Logistic regression)
Gradient decent • Gradient decent for classification (Logistic regression), for the output layer
Gradient computation
Gradient computation: Backpropagation algorithm
Gradient computation: Backpropagation algorithm
How to tune the weights? (Learning)
Try different initial random weights If cost function is non-convex, trying different initial random weights can help to find the global minimum (e. g. , by applying optimization techniques like PSO to find the best weights).
Example: One Machine Infinite Bus (OMIB) system - We randomly sample values for Pu and Qu creating a database with 5000 samples (objects) and for each object we have a set of attributes (Pu, Qu, V 1, P 1, Vinf, Xinf, CCT). Source: Automatic Learning techniques in Power Systems, L. Wehenkel 21
ANN structure and tuned weights - After tuning the weights Source: Automatic Learning techniques in Power Systems, L. Wehenkel 22
Assignment 23
Questions? Lars Nordström larsno@kth. se Kaveh Paridari paridari@kth. se 24
- Kth royal institute of technology notable alumni
- Kth machine learning
- Deep learning kth
- Royal netherlands meteorological institute
- Concept learning task in machine learning
- Analytical learning in machine learning
- Pac learning model in machine learning
- Machine learning t mitchell
- Inductive and analytical learning in machine learning
- Analytical learning in machine learning
- Instance based learning in machine learning
- Inductive learning machine learning
- First order rule learning in machine learning
- Lazy learners vs eager learner
- Deep learning vs machine learning
- Cuadro comparativo e-learning m-learning b-learning
- Calc institute of technology
- Vidhyadeep institute of engineering and technology
- Waterford institute of technology vacancies
- Spike unist
- Assistive technology implementation plan sample
- Masdar institute of science and technology
- Madhav institute of technology and science
- Grenoble
- Gastechnology