Recurrent Neural Network RNN Example Application Slot Filling
- Slides: 89
Recurrent Neural Network (RNN)
Example Application • Slot Filling I would like to arrive Taipei on November 2 nd. ticket booking system Slot Destination: Taipei time of arrival: November 2 nd
Example Application Solving slot filling by Feedforward network? Input: a word (Each word is represented as a vector) Taipei
1 -of-N encoding How to represent each word as a vector? 1 -of-N Encoding lexicon = {apple, bag, cat, dog, elephant} The vector is lexicon size. Each dimension corresponds to a word in the lexicon apple = [ 1 0 0] bag = [ 0 1 0 0 0] cat = [ 0 0 1 0 0] dog = [ 0 0 0 1 0] The dimension for the word elephant = [ 0 0 is 1, and others are 0 1]
Beyond 1 -of-N encoding Dimension for “Other” 1 1 26 X 26 … 1 … w = “Gandalf” w = “Sauron” … 1 p-p-l 0 0 … … “other” p-l-e … 0 a-p-p … elephant cat a-a-a a-a-b … dog 0 0 … apple bag Word hashing w = “apple” 5
Example Application Solving slot filling by Feedforward network? Input: a word (Each word is represented as a vector) Output: Probability distribution that the input word belonging to the slots Taipei dest time of departure
Example Application arrive Taipei on November 2 nd other dest other time Problem? leave Taipei on November 2 nd place of departure Neural network needs memory! Taipei dest time of departure
Recurrent Neural Network (RNN) The output of hidden layer are stored in the memory. store Memory can be considered as another input.
Input sequence: Example output sequence: 4 4 2 2 store given Initial 0 values 0 All the weights are “ 1”, no bias All activation functions are linear 1 1
Input sequence: Example output sequence: 12 12 6 6 store 2 2 All the weights are “ 1”, no bias All activation functions are linear 1 1
Input sequence: Example output sequence: Changing the sequence order will change the output. 32 32 16 16 store 6 6 All the weights are “ 1”, no bias All activation functions are linear 2 2
RNN The same network is used again and again. Probability of “arrive” in each slot Probability of “on” in each slot y 2 y 1 a 1 x 1 Probability of “Taipei” in each slot store y 3 a 2 a 1 x 2 store a 3 a 2 x 3 arrive Taipei on November 2 nd
RNN Prob of “leave” in each slot y 1 a 1 Different Prob of “Taipei” in each slot y 2 …… store a 2 a 1 x 2 leave Taipei Prob of “arrive” Prob of “Taipei” in each slot y 2 …… y 1 a 1 …… …… store a 2 a 1 x 2 arrive Taipei The values stored in the memory is different. …… ……
Of course it can be deep … yt yt+1 yt+2 …… …… …… xt xt+1 xt+2
Elman Network & Jordan Network Elman Network yt yt yt+1 Wo Wo Wi Wi Wh Wo …… …… Wi xt Wh xt+1 xt Wi xt+1
Bidirectional RNN xt xt+2 xt+1 …… …… yt+1 yt yt+2 …… …… xt xt+1 xt+2
Long Short-term Memory (LSTM) Other part of the network Signal control the output gate (Other part of the network) Signal control the input gate (Other part of the network) Output Gate Special Neuron: 4 inputs, 1 output Memory Cell Forget Gate Input Gate LSTM Other part of the network Signal control the forget gate (Other part of the network)
multiply Activation function f is usually a sigmoid function Between 0 and 1 Mimic open and close gate c multiply
LSTM - Example 0 1 0 0 0 3 1 0 3 2 0 0 3 4 1 0 7 2 0 0 7 1 0 1 7 3 -1 0 0 6 1 0 1 0 0 0 7 0 0 6 When x 2 = 1, add the numbers of x 1 into the memory When x 2 = -1, reset the memory When x 3 = 1, output the number in the memory.
x 1 0 x 2 0 y 0 + x 3 100 0 7 0 4 1 0 2 0 0 1 3 -1 0 0 x 1 1 -10 0 x 1 0 x 2 100 0 + 100 x 2 0 x 3 + 10 1 x 3 0 1 -10 + 1 x 1 0 0 0 x 2 x 3 1 0
3 0 1 0 y ≈0 0 + 7 0 4 1 0 2 0 0 1 3 -1 0 0 3 1 -10 30 3 0 1 100 0 3 ≈0 0 100 0 + ≈1 0 0 1 -10 + ≈1 3 0 0 3 10 1 + 1 3 0 1 100 1 0 0 0 1 3 1 0
4 0 1 0 y ≈0 0 + 7 0 4 1 0 2 0 0 1 3 -1 0 0 4 1 -10 37 4 0 1 100 0 7 ≈0 0 100 0 + ≈1 0 0 1 -10 + ≈1 4 0 0 4 10 1 + 1 4 0 1 100 1 0 0 0 1 3 1 0
2 0 0 0 y ≈0 0 + 7 0 4 1 0 2 0 0 1 3 -1 0 0 2 1 -10 7 2 0 0 100 0 7 ≈0 0 100 0 + ≈0 0 0 1 -10 + ≈1 0 0 0 2 10 1 + 1 2 0 0 100 0 0 1 3 1 0
1 0 0 0 y ≈7 0 + 7 0 4 1 0 2 0 0 1 3 -1 0 0 1 1 -10 7 1 0 0 100 0 7 ≈1 1 100 0 + ≈0 1 -10 + ≈1 0 0 1 1 10 1 + 1 1 0 0 100 0 1 1 3 1 0
3 0 -1 0 y ≈0 0 + 7 0 4 1 0 2 0 0 1 3 -1 0 0 3 1 -10 07 3 0 -1 100 0 0 ≈0 0 100 0 + ≈0 0 0 1 -10 ≈0 100 -1 0 0 0 3 10 1 + 1 3 + 0 0 0 -1 0 1 3 1 0
Original Network: ØSimply replace the neurons with LSTM …… …… x 1 x 2 Input
+ + + + 4 times of parameters x 1 x 2 Input
LSTM ct-1 …… vector zf zi z xt zo 4 vectors
LSTM yt zo ct-1 zf zf zi z xt zo zi z
Extension: “peephole” LSTM yt yt+1 ct ct-1 zf ct-1 zi ht-1 z xt ct+1 zf zo ct zi ht z xt+1 zo
Multiple-layer LSTM Don’t worry if you cannot understand this. Keras can handle it. Keras supports “LSTM”, “GRU”, “Simple. RNN” layers This is quite standard now. https: //img. komicolle. org/2015 -09 -20/src/14426967627131. gif
Learning Target dest other 0 … 1 … 0 y 1 y 2 y 3 a 1 copy a 2 a 1 copy a 3 a 2 Wi x 1 Training Sentences: x 3 x 2 arrive Taipei on November 2 nd other dest other time
Learning Backpropagation through time (BPTT) copy
Unfortunately …… 感謝 曾柏翔 同學 提供實驗結果 • RNN-based network is not always easy to learn Real experiments on Language modeling Total Loss sometimes Lucky Epoch
The error surface is rough. The error surface is either very flat or very steep. Total Loss Cost Clipping w 2 w 1 [Razvan Pascanu, ICML’ 13]
Why? Toy Example y 1 y 2 y 3 1 1 1 w 1 1 0 1 …… w w Small Learning rate? Large Learning rate? 999 =w y 1000 0 1 0
Helpful Techniques • Long Short-term Memory (LSTM) • Can deal with gradient vanishing (not gradient explode) Ø Memory and input are added Ø The influence never disappears unless forget gate is closed No Gradient vanishing (If forget gate is opened. ) Gated Recurrent Unit (GRU): [Cho, EMNLP’ 14] add
Helpful Techniques Clockwise RNN [Jan Koutnik, JMLR’ 14] Structurally Constrained Recurrent Network (SCRN) [Tomas Mikolov, ICLR’ 15] Vanilla RNN Initialized with Identity matrix + Re. LU activation function [Quoc V. Le, ar. Xiv’ 15] Ø Outperform or be comparable with LSTM in 4 different tasks
More Applications …… Probability of “arrive” in each slot y 1 Probability of “Taipei” in each slot Probability of “on” in each slot y 2 y 3 Input and output are both sequences store 1 2 a a 3 a with the same length 1 a a 2 RNN can do more than that! x 1 x 2 x 3 arrive Taipei on November 2 nd
Many to one [Shen & Lee, Interspeech 16] • Input is a vector sequence, but output is only one vector OT Key Term Extraction … V 2 V 3 V 4 … V 1 Key Terms: DNN, LSTN VT Output Layer x 1 x. T Hidden Layer VT Σαi. Vi Embedding Layer V 1 V 2 V 3 V 4 … document Embedding Layer … x 2 x 3 x 4 α 1 α 2 α 3 α 4 … αT
Many to Many (Output is shorter) • Both input and output are both sequences, but the output is shorter. • E. g. Speech Recognition Output: “好棒” (character sequence) Trimming Problem? 好好好棒棒棒棒棒 Why can’t it be “ 好棒棒” Input: (vector sequence )
Many to Many (Output is shorter) • Both input and output are both sequences, but the output is shorter. • Connectionist Temporal Classification (CTC) [Alex Graves, ICML’ 06][Alex Graves, ICML’ 14][Haşim Sak, Interspeech’ 15][Jie Li, Interspeech’ 15][Andrew Senior, ASRU’ 15] “好棒” Add an extra symbol “φ” representing “null” 好 φ φ 棒 φ φ “好棒棒” 好 φ φ 棒 φ φ
Many to Many (Output is shorter) • CTC: Training 好 φ 棒 φ φ φ Acoustic Features: Label: 好 棒 All possible alignments are considered as correct. 好 φ φ 棒 φ φ 好 φ φ φ 棒 φ ……
Many to Many (Output is shorter) • CTC: example HIS FRIEND’S φφ φ φ φ φφ φφ Graves, Alex, and Navdeep Jaitly. "Towards end-to-end speech recognition with recurrent neural networks. " Proceedings of the 31 st International Conference on Machine Learning (ICML-14). 2014.
Many to Many (No Limitation) • Both input and output are both sequences with different lengths. → Sequence to sequence learning • E. g. Machine Translation (machine learning→機器學習) learning machine Containing all information about input sequence
Many to Many (No Limitation) • Both input and output are both sequences with different lengths. → Sequence to sequence learning • E. g. Machine Translation (machine learning→機器學習) 機 器 學 習 慣 性 …… …… learning machine Don’t know when to stop
Many to Many (No Limitation) • Both input and output are both sequences with different lengths. → Sequence to sequence learning • E. g. Machine Translation (machine learning→機器學習) 器 學 習 === 機 learning machine Add a symbol “===“ (斷) [Ilya Sutskever, NIPS’ 14][Dzmitry Bahdanau, ar. Xiv’ 15]
https: //arxiv. org/pdf/1612. 01744 v 1. pdf Many to Many (No Limitation) • Both input and output are both sequences with different lengths. → Sequence to sequence learning • E. g. Machine Translation (machine learning→機器學習)
Beyond Sequence • Syntactic parsing john has a dog Oriol Vinyals, Lukasz Kaiser, Terry Koo, Slav Petrov, Ilya Sutskever, Geoffrey Hinton, Grammar as a Foreign Language, NIPS 2015
Sequence-to-sequence Auto-encoder - Text • To understand the meaning of a word sequence, the order of the words can not be ignored. white blood cells destroying an infection exactly the same bag-of-word an infection destroying white blood cells positive different meaning negative
Sequence-to-sequence Auto-encoder - Text Li, Jiwei, Minh-Thang Luong, and Dan Jurafsky. "A hierarchical neural autoencoder for paragraphs and documents. " ar. Xiv preprint ar. Xiv: 1506. 01057(2015).
Sequence-to-sequence Auto-encoder - Text Li, Jiwei, Minh-Thang Luong, and Dan Jurafsky. "A hierarchical neural autoencoder for paragraphs and documents. " ar. Xiv preprint ar. Xiv: 1506. 01057(2015).
Sequence-to-sequence Auto-encoder - Speech • Dimension reduction for a sequence with variable length audio segments (word-level) Fixed-length vector dog never Yu-An Chung, Chao-Chung Wu, Chia-Hao Shen, dogs Hung-Yi Lee, Lin-Shan Lee, Audio Word 2 Vec: Unsupervised Learning of Audio Segment Representations using Sequence-to-sequence Autoencoder, Interspeech 2016 ever never
Sequence-to-sequence Auto-encoder - Speech Audio archive divided into variablelength audio segments Off-line Audio Segment to Vector Spoken Query Audio Segment to Vector On-line Similarity Search Result
Sequence-to-sequence Auto-encoder - Speech vector audio segment RNN Encoder The values in the memory represent the whole audio segment The vector we want How to train RNN Encoder? x 1 x 2 x 3 x 4 acoustic features audio segment
Sequence-to-sequence Input acoustic features Auto-encoder The RNN encoder and decoder are jointly trained. x 1 x 2 x 3 x 4 y 1 y 2 y 3 y 4 RNN Encoder x 1 x 2 x 3 x 4 acoustic features audio segment RNN Decoder
Sequence-to-sequence Auto-encoder - Speech • Visualizing embedding vectors of the words fear fame near
Demo: Chat-bot 電視影集 (~40, 000 sentences)、美國總統大選辯論
Demo: Chat-bot • Develop Team Interface design: Prof. Lin-Lin Chen & Arron Lu Web programming: Shi-Yun Huang Data collection: Chao-Chuang Shih System implementation: Kevin Wu, Derek Chuang, & Zhi -Wei Lee (李致緯), Roy Lu (盧柏儒) • System design: Richard Tsai & Hung-Yi Lee • • 61
Demo: Video Caption Generation A girl is running. Video A group of people is knocked by a tree. A group of people is walking in the forest.
Demo: Video Caption Generation • Can machine describe what it see from video? • Demo: 台大語音處理實驗室 曾柏翔、吳柏瑜、 盧宏宗 • Video: 莊舜博、楊棋宇、黃邦齊、萬家宏
Demo: Image Caption Generation • Input an image, but output a sequence of words a woman is === A vector for whole image [Kelvin Xu, ar. Xiv’ 15][Li Yao, ICCV’ 15] …… CNN Input image Caption Generation
Demo: Image Caption Generation • Can machine describe what it see from image? • Demo: 台大電機系 大四 蘇子睿、林奕辰、徐翊 祥、陳奕安 MTK 產學大聯盟
http: //news. ltn. com. tw/photo/politics/breakingnews/975542_1
Attention-based Model What you learned in these lectures Breakfast today What is deep learning? Answer Organize http: //henrylo 1605. blogspot. tw/2015/05/blog-post_56. html summer vacation 10 years ago
Attention-based Model Input output DNN/RNN Reading Head Controller Reading Head …… …… Machine’s Memory Ref: http: //speech. ee. ntu. edu. tw/~tlkagk/courses/MLDS_2015_2/Lecture/Attain%20(v 3). ecm. mp 4/index. html
Attention-based Model v 2 Input output DNN/RNN Reading Head Controller Writing Head Controller Reading Head …… …… Machine’s Memory Neural Turing Machine
Reading Comprehension Query DNN/RNN answer Reading Head Controller Semantic Analysis …… …… Each sentence becomes a vector.
Reading Comprehension • End-To-End Memory Networks. S. Sukhbaatar, A. Szlam, J. Weston, R. Fergus. NIPS, 2015. The position of reading head: Keras has example: https: //github. com/fchollet/keras/blob/master/examples/babi_me mnn. py
Visual Question Answering source: http: //visualqa. org/
Visual Question Answering Query DNN/RNN answer Reading Head Controller CNN A vector for each region
Speech Question Answering • TOEFL Listening Comprehension Test by Machine • Example: Audio Story: (The original story is 5 min long. ) Question: “ What is a possible origin of Venus’ clouds? ” Choices: (A) gases released as a result of volcanic activity (B) chemical reactions caused by high surface temperatures (C) bursts of radio energy from the plane's surface (D) strong winds that blow dust into the atmosphere
Model Architecture Answer Select the choice most Attention similar to the answer Attention Question Semantics Semantic Analysis Question: “what is a possible origin of Venus‘ Everything is learned from training examples …… It be quite possible that this be due to volcanic eruption because volcanic eruption often emit gas. If that be the case volcanism could very well be the root cause of Venus 's thick cloud cover. And also we have observe burst of radio energy from the planet 's surface. These burst be similar to what we see when volcano erupt on earth …… Speech Recognition Audio Story: Semantic Analysis
Simple Baselines (4) the choice with semantic most similar to others Accuracy (%) (2) select the shortest choice as answer Experimental setup: 717 for training, 124 for validation, 122 for testing random (1) (2) (3) (4) Naive Approaches (5) (6) (7)
Memory Network Accuracy (%) Memory Network: 39. 2% (proposed by FB AI group) (1) (2) (3) (4) Naive Approaches (5) (6) (7)
Proposed Approach [Tseng & Lee, Interspeech 16] [Fang & Hsu & Lee, SLT 16] Proposed Approach: 48. 8% Accuracy (%) Memory Network: 39. 2% (proposed by FB AI group) (1) (2) (3) (4) Naive Approaches (5) (6) (7)
To Learn More …… • The Unreasonable Effectiveness of Recurrent Neural Networks • http: //karpathy. github. io/2015/05/21/rnneffectiveness/ • Understanding LSTM Networks • http: //colah. github. io/posts/2015 -08 Understanding-LSTMs/
Deep & Structured
RNN v. s. Structured Learning • RNN, LSTM • HMM, CRF, Structured • Unidirectional RNN does Perceptron/SVM NOT consider the whole sequence • Cost and error not always related • Deep • Using Viterbi, so consider the whole sequence ? • How about Bidirectional RNN? • Can explicitly consider the label dependency • Cost is the upper bound of error
Integrated Together HMM, CRF, Structured Perceptron/SVM Deep • Explicitly model the dependency • Cost is the upper bound of error RNN, LSTM http: //photo 30. bababian. com/upload 1/20100415/42 E 9331 A 6580 A 46 A 5 F 89 E 98638 B 8 FD 76. jpg
Integrated together • Speech Recognition: CNN/LSTM/DNN + HMM P(a|xl) P(b|xl) P(c|xl) …… …… RNN xl …… …… RNN Count
Integrated together • Semantic Tagging: Bi-directional LSTM + CRF/Structured SVM …… …… …… RNN …… …… …… Testing:
Is structured learning practical? • Considering GAN Real x Noise From Gaussian Generator Inference: Problem 2 x Generated Problem 3: you know how to learn F(x) Discriminator Evaluation Function F(x) Problem 1 1/0
Is structured learning practical? • Conditional GAN x Generator Real (x, y) pair y x Discriminator 1/0
Deep and Structured will be the future. Sounds crazy? People do think in this way … • Connect Energy-based model with GAN: • A Connection Between Generative Adversarial Networks, Inverse Reinforcement Learning, and Energy. Based Models • Deep Directed Generative Models with Energy-Based Probability Estimation • ENERGY-BASED GENERATIVE ADVERSARIAL NETWORKS • Deep learning model for inference • Deep Unfolding: Model-Based Inspiration of Novel Deep Architectures • Conditional Random Fields as Recurrent Neural Networks
Machine learning and having it deep and structured (MLDS) • 和 ML 的不同 • 在這學期 ML 中有提過的內容 (DNN, CNN …),在 MLDS 中不再重複,只做必要的復習 • 教科書: “Deep Learning” (http: //www. deeplearningbook. org/) • Part II 是講 deep learning 、Part III 就是講 structured learning
Machine learning and having it deep and structured (MLDS) • 所有作業都 2 ~ 4 人一組,可以先組好隊後一起來修 • MLDS 的作業和之前不同 • RNN (把之前 MLDS 的三個作業合為一個)、Attentionbased model 、Deep Reinforcement Learning 、Deep Generative Model 、Sequence-to-sequence learning • MLDS 初選不開放加簽,以組為單位加簽,作業 0的內容 是做一個 DNN (可用現成套件)
- Extensions of recurrent neural network language model
- Colah lstm
- Recurrent neural network based language model
- Lstm andrew ng
- Rnn slot
- Pixel rnn
- Simple recurrent network
- Backpropagation example
- Region filling can be performed in following approaches
- Rnn backward pass
- Translate.google.com
- Vanilla rnn
- Rnn gradient
- Rnn types
- Vanilla rnn
- Mxnet rnn
- Rnn andrew ng
- Single headed attention rnn: stop thinking with your head
- Cnn rnn
- Localization of function ib psychology
- Least mean square algorithm in neural network
- Student teacher deep learning
- Cost function neural network
- Tlu neural network
- Meshnet: mesh neural network for 3d shape representation
- Pengertian artificial neural network
- Neural network in r
- Neural network matlab toolbox
- Spss neural network
- Xkcd neural network
- Convolution neural network ppt
- Limitations of perceptron
- Artificial neural network in data mining
- Least mean square algorithm in neural network
- Weka mlp
- Difference between adaline and madaline
- Decision boundary of neural network
- Ann unsupervised learning
- Parallel tacotron 2
- Neural network terminology
- Principal component analysis jmp
- Playground.tensorflow.or
- Instar and outstar in neural network
- Grossberg neural network
- Neural network in data mining
- Auto associative memory
- Alternatives to convolutional neural networks
- Liran szlak
- Tic tac toe neural network
- Neural network data preprocessing
- Network application design
- Multilayer neural network
- Adaptive learning neural network
- Nnlm neural network
- Convolutional neural network
- Maxnet neural network
- Neural network blackjack
- Matt damon
- Tensor playground
- Machine learning playground
- Graph neural network lecture
- Edlut
- Artificial neural network conclusion
- Multiple slot substitution drill examples
- Cadaveric position of vocal cord
- Visualizing and understanding recurrent networks
- Deep venous palmar arch
- Cricothyroid membrane
- Recurrent stroke causes
- Recurrent stroke causes
- Recurrent artery of huebner
- Cs 231 n
- Bert question generation
- 6 elements of hair design
- Coarcitation of aorta
- A recurrent image in a literary work is the
- Part 135 recurrent training
- Partial vs complete recurrent laryngeal nerve injury
- Jolls triangle thyroid
- Khepara
- Cellulitis nose
- Foot arteries
- Rolo tracking
- 24 slot 2 pole winding diagram
- Face of weld
- Welding symbols worksheet
- Bucky slot cover purpose
- Umpire hand signals baseball
- Slot tekst voorbeeld
- Integral vs modular design