AUTOENCODER Hungyi Lee Selfsupervised Learning Framework Autoencoder Selfsupervised






































- Slides: 38

AUTO-ENCODER Hung-yi Lee 李宏毅

Self-supervised Learning Framework Auto-encoder Self-supervised Learning Pre-train Selfsupervised Learned by tasks not requiring label data Model for Task 1 Model for Task 2 Downstream Tasks Model for Task 3

Outline Basic Idea of Auto-encoder Feature Disentanglement Discrete Latent Representation More Applications

Unlabeled Images Auto-encoder Sounds familiar? We have seen the same idea in Cycle GAN. As close as possible (reconstruction) high dim old feature vector NN Encoder NN Decoder low dim (bottleneck) Dimension reduction Embedding, Representation, Code New feature for downstream tasks

More Dimension Reduction (not based on deep learning) https: //youtu. be/iwh 5 o_M 4 BNU https: //youtu. be/GBUEjkpox. Xc PCA t-SNE

Why Auto-encoder? 《神鵰俠侶》

Why Auto-encoder? As close as possible (reconstruction) NN Encoder 3 x 3 Find pattern, and then compress NN Decoder 2 dim 3 x 3

Why Auto-encoder? NN Encoder 《神鵰俠侶》

Auto-encoder is not a new idea Hinton, Geoffrey E. , and Ruslan R. Salakhutdinov. "Reducing the dimensionality of data with neural networks. " Science 313. 5786 (2006): 504 -507

De-noising Auto-encoder As close as possible Add noises vector NN Encoder NN Decoder The idea sounds familiar? Vincent, Pascal, et al. "Extracting and composing robust features with denoising autoencoders. " ICML, 2008.

Reconstruction Review: BERT A de-noising auto-encoder softmax Decoder minimize cross entropy 灣 Linear Embedding Encoder Add noise BERT 台 灣 大 學

Outline Basic Idea of Auto-encoder Feature Disentanglement Discrete Latent Representation More Applications

Representation includes information of different aspects Encoder input image Decoder object, texture …. Encoder reconstructed Decoder input audio reconstructed content, speaker … Encoder input sentence Decoder syntax, semantic … reconstructed

Feature Disentangle Encoder Decoder input audio reconstructed content, speaker … content information Encoder input audio https: //arxiv. org/abs/1904. 05742 https: //arxiv. org/abs/1804. 02812 https: //arxiv. org/abs/1905. 05879 Decoder reconstructed speaker information

Application: Voice Conversion

In the past Speaker A Speaker B How are you? Good morning Today Speaker A 天氣真好 再見囉 Speaker B How are you? Good morning Speakers A and B are talking about completely different things.

Application: Voice Conversion Encoder How are you? Decoder How are you? Hello Encoder Hello Decoder Hello

Application: Voice Conversion Encoder How are you? Decoder Hello Encoder Hello How are you?

Application: Voice Conversion 感謝解正平同學提供 實驗結果 Encoder Do you … Ph. D. ? Decoder Japanese Encoder Do you … Ph. D. ?

Application: Voice Conversion 感謝解正平同學提供 實驗結果 Encoder Do you … Ph. D. ? Japanese Decoder Japanese Encoder Japanese

Outline Basic Idea of Auto-encoder Feature Disentanglement Discrete Latent Representation More Applications

Discrete Representation Real numbers 0. 9 0. 1 0. 3 0. 7 NN Encoder NN Decoder Binary 1 0 0 1 NN Encoder glasses NN Decoder female One-hot NN Encoder 0 0 1 0 NN Decoder class

Discrete Representation https: //arxiv. org/abs/1711. 00937 • Vector Quantized Variational Auto-encoder (VQVAE) vector 3 NN Decoder vector 5 vector 4 vector 3 vector 2 vector 1 Codebook (a set of vectors) Learn from data vector NN Encoder (c. f. attention) Compute similarity The most similar one is the input of decoder. For speech, the codebook represents phonetic information https: //arxiv. org/pdf/1901. 08810. pdf

https: //arxiv. org/abs/1810. 02851 Text as Representation seq 2 seq auto-encoder Only need a lot of documents to train the model Unsupervised Summarization word sequence document En Seq 2 seq not readable … document De Summary? Seq 2 seq

This is cycle GAN Text as Representation Human written summaries Let discriminator considers my output as real Real or not Di Discriminator word sequence Readable document En Seq 2 seq document De Summary? Seq 2 seq



Tree as Embedding https: //arxiv. org/abs/1806. 07832 https: //arxiv. org/abs/1904. 03746

Outline Basic Idea of Auto-encoder Feature Disentanglement Discrete Latent Representation More Applications

Generator As close as possible vector NN Encoder NN Decoder Is it a generator? vector Randomly generate a vector from a distribution NN Decoder Image ? With some modification, we have variational auto-encoder (VAE).

Compression As close as possible vector NN Encoder NN Decoder Low-dim Compression https: //arxiv. org/abs/1708. 00838 https: //arxiv. org/abs/1703. 00395 Decompression Lossy

Anomaly Detection • Anomaly Detector normal outlier, novelty, exceptions anomaly

Anomaly Detection Training Data: anomaly

Anomaly Detection • Binary Classification? We only have one class. Training auto-encoder

Approach: Auto-encoder Training Using real human faces to learn an autoencoder As close as possible Encoder Decoder Training data Testing Can be reconstructed Encoder normal Decoder

Approach: Auto-encoder Training Using real human faces to learn an autoencoder As close as possible Encoder Decoder Training data Testing cannot be reconstructed Encoder anomaly Decoder

More about Anomaly Detection • Part 1: https: //youtu. be/g. Dp 2 LXGn. VLQ • Part 2: https: //youtu. be/c. Yr. Nj. Lxko. Xs • Part 3: https: //youtu. be/ue. Dlm 2 Fk. Cnw • Part 4: https: //youtu. be/Xwk. HOUPbc 0 Q • Part 5: https: //youtu. be/Fh 1 x. FBkt. RLQ • Part 6: https: //youtu. be/Lm. FWzmn 2 r. FY • Part 7: https: //youtu. be/6 W 8 Fq. UGYy. Do

Concluding Remarks Basic Idea of Auto-encoder Feature Disentanglement Discrete Latent Representation More Applications