Intelligent Leaning A Brief Introduction to Artificial Neural

Intelligent Leaning -- A Brief Introduction to Artificial Neural Networks Chiung-Yao Fang

Learning n What is leaning? ¨ Machine learning is programming computers to optimize a performance criterion using example data or past experience. ¨ There is no need to “learn” to calculate payroll ¨ Incremental learning, active learning, … n The type of learning ¨ Supervised learning ¨ Unsupervised learning ¨ Reinforcement learning 1/10/2022 2

Understanding the Brain n Levels of analysis (Marr, 1982) 1. Computational theory 2. Representation and algorithm 3. Hardware implementation Example: sorting ¨ The same computational theory may have multiple representations and algorithms. ¨ A given representation and algorithm may have multiple hardware implementations. Reverse engineering: From hardware to theory 1/10/2022 3

Understanding the Brain n n Parallel processing: SIMD vs MIMD ¨ SIMD: single instruction multiple data machines n All processors execute the same instruction but on different pieces of data ¨ MIMD: multiple instruction multiple data machines n Different processors may execute different instructions on different data Neural net: ¨ NIMD: neural instruction multiple data machines n Each processor corresponds to a neuron, local parameters correspond to its synaptic weights, and the whole structure is a neural network. 1/10/2022 n Learning: Update by training/experience n Learning from examples 4

Biological-Type Neural Networks 突觸(兩個神經原的相接處) ( 樹狀突 軸突 1/10/2022 5

Application-Driven Neural Networks n Three main characteristics: ¨ Adaptiveness and self-organization ¨ Nonlinear network processing ¨ Parallel processing 1/10/2022 6

Perceptron (Rosenblatt, 1962) Bias(偏壓) 1/10/2022 7

1/10/2022 8

What a Perceptron Does n Regression: y=wx+w 0 y Connection weight y w x x x 0=+1 n x 0: bias unit 1/10/2022 9

What a Perceptron Does n n Classification: y = s(wx+w 0>0) Define s(. ) as the threshold function y s w 0 w x n Choose C 1 if s(wx+w 0)>0 else choose C 2 1/10/2022 10

K Outputs 1/10/2022 11

Learning Boolean AND 1/10/2022 12

XOR (Minsky and Papert, 1969) n No w 0, w 1, w 2 satisfy: r w 0 x 0=1 1/10/2022 w 1 x 1 w 2 x 2 13

Multilayer Perceptrons (Rumelhart et al. , 1986) 1/10/2022 14

y x 1 XOR x 2 = (x 1 AND ~x 2) OR (~x 1 AND x 2) 1/10/2022 15

Structures of Neural Networks 1/10/2022 16

Connection Structures n Four types of weighted connections: ¨ Feedforward connections ¨ Feedback connections ¨ Lateral connections ¨ Time-delay connections 1/10/2022 17

Connection Structures n Single-layer example 1/10/2022 18

Taxonomy of Neural Networks HAM SOM 1/10/2022 19

Supervised and Unsupervised Networks 1/10/2022 20

A Top-down Perspective 1/10/2022 21

1/10/2022 22

Applications: Association n Association: ¨ Auto-Association ¨ Hetero-Association 1/10/2022 23

Applications: Classification n Classification: ¨ Unsupervised classification (clustering) ¨ Supervised classification 1/10/2022 24

Applications: Pattern Completions n Two kinds of pattern completion problems: ¨ Static n pattern completion Multilayer nets, Boltzmann machines, and Hopfield nets ¨ Temporal n 1/10/2022 pattern completion Markov models and time-delay dynamic networks 25

Applications: Regression and Generalization 1/10/2022 26

Applications: Optimization 1/10/2022 27

Examples: A Toy OCR n Optical character recognition (OCR) ¨ Supervised n n 1/10/2022 learning The retrieving phase The training phase 28

Examples: A Toy OCR 1/10/2022 29

1/10/2022 30

Supervised Learning Neural Networks Backpropagation HAM

Backpropagation 1/10/2022 32

Regression Forward Backward x 1/10/2022 33

Hidden Layer n Do we have more hidden layers? Yes! But complicate. ¨ “Long and narrow” network vs “Short and fat” network ¨ n Two hidden layer example: For every input case of region, that region can be delimited by hyperplanes on all sides using hidden units on the first hidden layer. ¨ A hidden unit in the second layer than ANDs them together to bound the region. ¨ n It has been proven that an MLP (multi-layer perceptrons) with one hidden layer can learn any nonlinear function of the input. 1/10/2022 34

HAM (Hetero-Associative Memory) Neural Network (1) v 1 v 2 vi i Excitatory connection vn Output layer (Competitive layer) wij xj j Input layer 1/10/2022 35

HAM Neural Network (2) n The input to neuron ni due to input stimuli x: nc: the winner after the competition 1/10/2022 36

Training Patterns for HAM 1/10/2022 37

1/10/2022 38

Unsupervised Learning Neural Networks SOM ART 1 ART 2

Self-organization Feature Maps 1/10/2022 40

1/10/2022 41

1/10/2022 42

An Assembly of SSO Neural Networks for Character Recognition 1/10/2022 43

An Assembly of SSO Neural Networks for Character Recognition 1/10/2022 44

ART 1 Neural Networks 1/10/2022 45

ART 2 Neural Networks (1) Orienting subsystem Signal generator S Reset + signal Attentional subsystem Category representation field F 2 y Input representation field F 1 + q + r + - G G + - + p + + + G - + G + v + u + + x - + w + G + Input vector i 1/10/2022 46

ART 2 Neural Network (2) n The activities on each of the six sublayers on F 1: where I is an input pattern where the J th node on F 2 is the winner 1/10/2022 47

ART 2 Neural Network (3) n n 1/10/2022 Initial weights: ¨ Top-down weights: ¨ Bottom-up weights: Parameters: 48

Road Sign Recognition System 1/10/2022 49

Classification Results of ART 2 Training Set Test Set 1/10/2022 50

Conclusions 1/10/2022 51

STA Neural Networks

STA (Spatial-temporal attention ) Neural Network a a i Output layer (Attention layer) nk ni Inhibitory connection Excitatory connection k wij xj nj Input layer 1/10/2022 53

STA Neural Network n The input to attention neuron ni due to input stimuli Gaussian function G x: Attention layer nk rk wkj ni corresponding neurons nj Input neuron 1/10/2022 The linking strengths between the input and the attention layers 54

STA Neural Network n The input to attention neuron ni due to lateral interaction: Interaction + Lateral distance “Mexican-hat” function of lateral interaction 1/10/2022 55

STA Neural Network n The net input to attention neuron ni : : a threshold to limit the effects of noise where 1< d <0 1/10/2022 56

STA Neural Network (5) stimulus activation t 1 pd The activation of an attention neuron in response to a stimulus. 1/10/2022 57

Results of STA Neural Networks 1/10/2022 58

Experimental Results 1/10/2022 59

Results of STA Neural Networks 1/10/2022 60
- Slides: 60