Neural Networks Capabilities and Examples L Manevitz Computer

Neural Networks: Capabilities and Examples L. Manevitz Computer Science Department HIACS Research Center University of Haifa L. Manevitz U. Haifa 1

What Are Neural Networks? What Are They Good for? How Do We Use Them? • Definitions and some history • Basics – Basic Algorithms – Examples • Recent Examples • Future Directions L. Manevitz U. Haifa 2

Natural versus Artificial Neuron • Natural Neuron L. Manevitz Mc. Cullough Pitts Neuron U. Haifa 3

Definitions and History • Mc. Cullough –Pitts Neuron • Perceptron • Adaline • Linear Separability • Multi-Level Neurons • Neurons with Loops L. Manevitz U. Haifa 4

Sample Feed forward Network (No loops) • Weights • Output • Weights • Input • Wji • Vik F(S wji xj L. Manevitz U. Haifa 5

Replacement of Threshold Neurons with Sigmoid or Differentiable Neurons • Sigmoid • Threshold L. Manevitz U. Haifa 6

Reason for Explosion of Interest • Two co-incident affects (around 1985 – 87) – (Re-)discovery of mathematical tools and algorithms for handling large networks – Availability (hurray for Intel and company!) of sufficient computing power to make experiments practical. L. Manevitz U. Haifa 7

Some Properties of NNs • Universal: Can represent and accomplish any task. • Uniform: “Programming” is changing weights • Automatic: Algorithms for Automatic Programming; Learning L. Manevitz U. Haifa 8

Networks are Universal • All logical functions represented by three level (non-loop) network (Mc. Cullough-Pitts) • All continuous (and more) functions represente by three level feed-forward networks (Cybenko al. ) • Networks can self organize (without teacher). • Networks serve as associative memories L. Manevitz U. Haifa 9

Universality • Mc. Cullough-Pitts: Adaptive Logic Gates; can represent any logic function • Cybenko: Any continuous function representable by three-level NN. L. Manevitz U. Haifa 10

Networks can “LEARN” and Generalize (Algorithms) • One Neuron (Perceptron and Adaline) Very popular in 1960 s – early 70 s – Limited by representability (only linearly separable • Feed forward networks (Back Propagation) – Currently most popular network (1987 –now) • Kohonen self-Organizing Network (1980 s – now)(loops) • Attractor Networks (loops) L. Manevitz U. Haifa 11

Learnability (Automatic Programming) • One neuron: Perceptron and Adaline algorithms (Rosenblatt and Widrow-Hoff) (1960 s –now) Feed forward Networks: Backpropagation (1987 – now) Associative Memories and Looped Networks (“Attractors”) (1990 s – now) L. Manevitz U. Haifa 12

Generalizability • Typically train a network on a sample set of examples • Use it on general class • Training can be slow; but execution is fast. L. Manevitz U. Haifa 13

Perceptron • weights • Pattern Identification • (Note: Neuron is trained) L. Manevitz U. Haifa 14

Feed Forward Network • weights L. Manevitz • weights U. Haifa 15

Classical Applications (1986 – 1997) • “Net Talk” : text to speech • ZIPcodes: handwriting analysis • Glovetalk: Sign Language to speech • Data and Picture Compression: “Bottleneck” • Steering of Automobile (up to 55 m. p. h) • Market Predictions • Associative Memories • Cognitive Modeling: (especially reading, …) L. Manevitz U. (Finnish) Haifa • Phonetic Typewriter 16

Neural Network • Once the architecture is fixed; the only free parameters are the weights • Thus Uniform Programming • Potentially Automatic Programming • Search for Learning Algorithms L. Manevitz U. Haifa 17

Programming: Just find the weights! • AUTOMATIC PROGRAMMING • One Neuron: Perceptron or Adaline • Multi-Level: Gradient Descent on Continuous Neuron (Sigmoid instead of step function). L. Manevitz U. Haifa 18

Prediction • delay • Input/Output • NN L. Manevitz U. Haifa • Compare 19

Training NN to Predict L. Manevitz U. Haifa 20

Finite Element Method • Numerical Method for solving p. d. e. s • Many user chosen parameters • Replace user expertise with NNs. L. Manevitz U. Haifa 21

FEM Flow chart L. Manevitz U. Haifa 22

Problems and Methods L. Manevitz U. Haifa 23

Finite Element Method and Neural Networks • Place mesh on body • Predict where to adapt mesh L. Manevitz U. Haifa 24

Placing Mesh on Body (Manevitz, Givoli and Yousef) • Need to place geometry on topology • Method: Use Kohonen algorithm • Idea: Identify neurons with FEM nodes – Identify weights of nodes with geometric location – Identify topology with adjaceny – RESULT: Equi-probably placement L. Manevitz U. Haifa 25

Kohonen Placement for FEM • Include slide from Malik’s work. L. Manevitz U. Haifa 26

Self-Organizing Network • Weights from input to neurons • Topology between neurons L. Manevitz U. Haifa 27

Self-Organizing Network • Weights from input give “location” to neuron • Kohonen algorithm results in “winner” neuron • After training, close input patterns have topologically close winners • Results in Equiprobable Continuous Mapping (without teacher) L. Manevitz U. Haifa 28

Placement of Mesh via Self Organizing NNs L. Manevitz U. Haifa 29

Placement of Mesh via Self Organizing NNs 2 Iteration 0 Iteration 6000; Quality =223 Iteration 500; Quality =288 Iteration 12000; Quality = 208 L. Manevitz U. Haifa Iteration 2000; Quality = 238 Iteration 30000; Quality =20230

Comparison of NN and PLTMG (249 nodes) NN (225 nodes); Quality = 279 Node Value Error Pltmg 2. 4 E-02 4. 51 E-02 NN 7. 5 E-03 9. 09 E-03 L. Manevitz U. Haifa 31

FEM Temporal Adaptive Meshes L. Manevitz U. Haifa 32

Prediction of Refinement of Elements • Method simulates time • Current adaptive method uses gradient • Can just MISS all the action. • We use NNs to PREDICT the gradient. • Under development with Manevitz, Givoli and Bitar. L. Manevitz U. Haifa 33

Training NN to Predict 2 L. Manevitz U. Haifa 34

Refinement Predictors • Need to choose features • Need to identify kinds of elements L. Manevitz U. Haifa 35

Other Predictions? • Stock Market (really!) • Credit Card Fraud (Master Card, USA) L. Manevitz U. Haifa 36

Surfer’s Apprentice Program • Manevitz and Yousef • Make a “model” of user for retrieving information from internet. • Many issues: here focus on retrieval of new pages similar to other pages of interest to user. Note ONLY POSITIVE DATA. L. Manevitz U. Haifa 37

L. Manevitz U. Haifa 38

Bottleneck Network • Train to Identity on Sample Data • Should be identity only on similar data • NOVELTY FILTER L. Manevitz U. Haifa 39

How well does it work? • Tested on Standard Reuter’s Data Base. • Used 25% for training • Withholding information on representation • The best method for retrieval using only positive training. (Better than SVM, etc. ) L. Manevitz U. Haifa 40

How to help Intel? (Make Billions? Reset NASDAQ) • Branch prediction? • (Note similarity to FEM refinement. ) • Perhaps can use to give predictor that is even user or application dependent. • (Note: Neural activity is, I am told, natural for VLSI design and there have been several such chips produced. ) L. Manevitz U. Haifa 41

Other Different Directions • Modify basic model to handle temporal adaptivity. (Occurs in real neurons according to latest biological information. ) • Apply to model human diseases, etc. L. Manevitz U. Haifa 42
- Slides: 42