CNN Alex Net Sungjoon Choi Alex Net Not
- Slides: 23
CNN / Alex. Net Sungjoon Choi
Alex. Net + Not so minor (actually SUPER IMPORTANT) heuristics. § § Re. Lu Nonlinearity Local Response Normalization Data Augmentation Dropout
What is Image. Net?
ILSVRC 2010 Image. Net Large Scale Visual Recognition Challenge (ILSVRC) It uses a subset of Image. Net with roughly 1 M images with 1 K categories. Are these all just cats? (Of course, some are super cute!) These are all different categories! (Egyptian, Persian, Tiger, Siamese, and Tabby cat)
Convolution Neural Network This is pretty much everything about the convolutional neural network. Convolution + Subsampling + Full Connection
What is CNN? CNNs are basically layers of convolutions followed by subsampling and fully connected layers. Intuitively speaking, convolutions and subsampling layers works as feature extraction layers while a fully connected layer classifies which category current input belongs to using extracted features. http: //www. wildml. com/2015/11/understanding-convolutional-neural-networks-for-nlp/
Why is CNN so powerful? Local Invariance Loosely speaking, as the convolution filters are ‘sliding’ over the input image, the exact location of the object we want to find does not matter much. Compositionality There is a hierarchy in CNNs. It is GOOD! Huge representation capacity! https: //starwarsanon. wordpress. com/tag/darth-sidious-vs-yoda/
What is Convolution? http: //deeplearning. stanford. edu/wiki/index. php/Feature_extraction_using_convolution
Details of Convolution Zero. Stride padding Channel
Conv: Zero-padding? What is the size of the input? What is the size of the output? What is the size of the filter? What is the size of the zero-padding?
Stride?
Conv: Stride? (Left) Stride size: 1 (Right) Stride size: 2 If stride size equals the filter size, there will be no overlapping.
Conv: Channel [batch, in_height, in_width, in_chnnel] [filter_height, filter_width, in_channels, out_channels] [batch, in_height=4, in_width=4, in_chnnel=3] [filter_height=3, filter_width=3, in_channels=3, out_channels=7]
Conv: Channel [batch, in_height=4, in_width=4, in_chnnel=3] [filter_height=3, filter_width=3, in_channels=3, out_channels=7] What is the number of parameters in this convolution layer?
Alex. Net What is the number of parameters? Why are layers divided into two parts?
Alex. Net
Re. LU Nonlinearity Re. LU tanh Faster Convergence!
Local Response Normalization It implements a form of lateral inhibition inspired by real neurons.
Reducing Overfitting It is often called regularization in machine learning literatures. More details will be handled in next week. In the Alex. Net, two regularization methods are used. § Data augmentation § Dropout
Reg: Data Augmentation http: //www. slideshare. net/Ken. Chatfield/chatfield 14 -devil
Reg: Data Augmentation 1
Reg: Data Augmentation 2 Color variation Probabilistically, not a single patch will be same at the training phase! (a factor of infinity!)
Reg: Dropout Original dropout [1] sets the output of each hidden neuron with certain probability. In this paper, they simply multiply the outputs by 0. 5. [1] G. E. Hinton, N. Srivastava, A. Krizhevsky, I. Sutskever, and R. R. Salakhutdinov. Improving neural networks by preventing coadaptation of feature detectors. ar. Xiv, 2012. http: //www. eddyazar. com/the-regrets-of-a-dropout-and-why-you-should-drop-out-too/
- Choi sungjoon
- Alex net
- Alex burns cnn
- Uncontrolled, lacking in restraint
- Simple present verbs
- Sorry, i ______ hear you at the door.
- Normalized cut loss for weakly-supervised cnn segmentation
- Normalized cut loss for weakly-supervised cnn segmentation
- Cnn misleading graphs
- Backpropagation cnn
- What is stride in cnn
- Font detector
- Cnn 10 september 7
- Cascade r-cnn: delving into high quality object detection
- Text classification with cnn
- First neighbor
- Jay contoy cnn
- Cnn
- A new backbone that can enhance learning capability of cnn
- Cnn 10 worksheet
- Vsepr exercise
- Backpropagation cnn
- Cnn
- Yaniv altshuler