SlideShare a Scribd company logo
ImageNet Classification with Deep
Convolutional Neural Networks
신우철
Introduction
1. Trained one of the largest CNN on ImageNet data. The advantages of CNN
are 1) CNN’s prior knowledge, which are stationarity of statistics and
locality of pixel dependencies, 2) its easiness to be controlled, varying its
depth and breath, contributing to fewer parameters and easier training.
2. Implemented highly-optimized GPU implementation to facilitate the
training of large CNNs on high resolution images.
3. Introduced new features to improve performance, reduce training time,
and prevent overfitting.
Dataset
• Down-sampled ImageNet images to 256 x 256. Trained on centered raw
RGB values of pixels.
1) Rescaled the image such that the shorter side was of length 256
2) Cropped out the central 256 x 256 patch from the resulting image of 1).
3) Subtracted the mean activity over the training set from each pixel.
cf)
NORB
MNIST
LabelMe
Architecture
• 8 layers = 5 Convolutional + 3 Fully-connected
• Newly introduced features
1) ReLU Nonlinearity
• Much faster to train since it is non-saturating
• Nonlinear |tanh(x)| function focuses on preventing overfitting, while ReLU
focuses on fast learning of large models on large datasets
Architecture
2) Training on two GPUs
• Cross-GPU parallelization, while the GPUs only communicate in certain
layers. This is to tune the amount of computation by communication.
Architecture
3) Local Response Normalization
• Since the architecture uses ReLU, one high activation value can affect
adjacent activation values in convolution or pooling. Therefore, LRN is
conducted.
filter
: the activity of a neuron computed by applying kernel i at position (x,y)
: the response-normalized activity
N : #all kernels
n : adjacent #kernels at position (x,y)
Architecture
3) Local Response Normalization
filter 0 filter 1 filter 2 filter 3
1 2 3 1 2 1 2 1 2 4 2 1
4 5 6 2 3 2 3 2 3 5 2 1
7 8 9 3 4 3 4 3 4 2 2 4
0.50 0.25 0.30 0.17 0.22 0.07 0.10 0.11 0.33 0.20 0.40 0.20
0.20 0.15 0.15 0.07 0.08 0.04 0.08 0.12 0.21 0.15 0.25 0.10
0.12 0.10 0.10 0.04 0.04 0.03 0.14 0.10 0.10 0.10 0.15 0.13
k alpha beta n N
0 1 1 2 4
=
2
{0 + 1 x 12 + 22 + 42 }1
Architecture
4) Overlapping Pooling
• Overlapping reduces overfitting compared to non-overlapping pooling.
Architecture
5) Overall architecture
Convolutional layer
Kernel size = 11
Stride = 4
Filter = 96
Zero-padding = 0
(227 – 11) / 4 + 1 = 55
Maxpooling
Kernel size(z) = 3
Stride(s) = 2
Convolutional layer
Kernel size = 5
Stride = 1
Filter = 256
Zero-padding = 2
(55 – 3) / 2 + 1 = 27
(27 +2 * 2 – 5) / 1 + 1 = 27
Local response normalization
Convolutional layer
Kernel size = 3
Stride = 1
Filter = 384
Zero-padding = 1
(27 + 1 * 2– 3) / 1 + 1 = 27
Convolutional layer
Kernel size = 3
Stride = 1
Filter = 384
Zero-padding = 1
(13 +1 * 2– 3) / 1 + 1 = 13
Maxpooling
Kernel size(z) = 3
Stride = 2
(27 – 3) / 2 + 1 = 13
Local response normalization
Convolutional layer
Kernel size = 3
Stride = 1
Filter = 256
Zero-padding = 1
(13 +1 * 2– 3) / 1 + 1 = 13
Maxpooling
Kernel size(z) = 3
Stride = 2
(13 – 3) / 2 + 1 = 6
Flatten 6 * 6 * 256 = 9216
Fully connected 4096
Fully connected 4096
Fully connected 1000(softmax)
Reducing Overfitting
1) Data Augmentation
(1) Image translations and horizontal reflections
Train set
• Image translations (x (256-224) * (256-224))
• Horizontal reflections (x 2)
Total : (256-224) * (256-224) * 2 = 2048
Test set
• Image translations (x 5)
• Horizontal reflections (x 2)
• Total: 5 * 2 = 10
Reducing Overfitting
1) Data Augmentation
(2) Altering intensity of RGB channels (performing PCA)
2) Dropout
• Applied dropout on first two FC layers with p = 0.5
Pi : eigen vector
 : eigen value
 : random
=
Details of Learning
• SGD with batch size of 128 examples
• Momentum = 0.9
• Weight decay = 0.0005
• Weight initialization : N(0, 0.012
)
• Neuron biases initialization:
Conv layers = 0
FC layers = 1
• Learning rate
Initialized at 0.01 and reduced three times prior to termination. Reduction was
done by dividing learning rate by 10 when the validation error rate stopped
improving with the current learning rate.
Results
Results
• Result of restricted connectivity between two GPUs result in specialization.
Kernels on GPU 1 are largely color-agnostic, while kernels on GPU 2 are
largely color-specific.
• Perform kNN at the last 4096-dimensional hidden layer shows that images
are semantically similar.
Ad

More Related Content

What's hot (20)

Kintinuous review
Kintinuous reviewKintinuous review
Kintinuous review
Dong-Won Shin
 
Reading group - Week 2 - Trajectory Pooled Deep-Convolutional Descriptors (TDD)
Reading group - Week 2 - Trajectory Pooled Deep-Convolutional Descriptors (TDD)Reading group - Week 2 - Trajectory Pooled Deep-Convolutional Descriptors (TDD)
Reading group - Week 2 - Trajectory Pooled Deep-Convolutional Descriptors (TDD)
Saimunur Rahman
 
DTAM: Dense Tracking and Mapping in Real-Time, Robot vision Group
DTAM: Dense Tracking and Mapping in Real-Time, Robot vision GroupDTAM: Dense Tracking and Mapping in Real-Time, Robot vision Group
DTAM: Dense Tracking and Mapping in Real-Time, Robot vision Group
Lihang Li
 
Depth estimation using deep learning
Depth estimation using deep learningDepth estimation using deep learning
Depth estimation using deep learning
University of Oklahoma
 
Seed net automatic seed generation with deep reinforcement learning for robus...
Seed net automatic seed generation with deep reinforcement learning for robus...Seed net automatic seed generation with deep reinforcement learning for robus...
Seed net automatic seed generation with deep reinforcement learning for robus...
NAVER Engineering
 
Motion Estimation in h.264 encoder
Motion Estimation in h.264 encoderMotion Estimation in h.264 encoder
Motion Estimation in h.264 encoder
Talal Khaliq
 
Model compression
Model compressionModel compression
Model compression
Nanhee Kim
 
FastCampus 2018 SLAM Workshop
FastCampus 2018 SLAM WorkshopFastCampus 2018 SLAM Workshop
FastCampus 2018 SLAM Workshop
Dong-Won Shin
 
30th コンピュータビジョン勉強会@関東 DynamicFusion
30th コンピュータビジョン勉強会@関東 DynamicFusion30th コンピュータビジョン勉強会@関東 DynamicFusion
30th コンピュータビジョン勉強会@関東 DynamicFusion
Hiroki Mizuno
 
Thesis Presentation
Thesis PresentationThesis Presentation
Thesis Presentation
Reuben Feinman
 
Yol ov2
Yol ov2Yol ov2
Yol ov2
Bang Tsui Liou
 
Single Image Depth Estimation using frequency domain analysis and Deep learning
Single Image Depth Estimation using frequency domain analysis and Deep learningSingle Image Depth Estimation using frequency domain analysis and Deep learning
Single Image Depth Estimation using frequency domain analysis and Deep learning
Ahan M R
 
A Novel Background Subtraction Algorithm for Dynamic Texture Scenes
A Novel Background Subtraction Algorithm for Dynamic Texture ScenesA Novel Background Subtraction Algorithm for Dynamic Texture Scenes
A Novel Background Subtraction Algorithm for Dynamic Texture Scenes
IJMER
 
Semantic segmentation with Convolutional Neural Network Approaches
Semantic segmentation with Convolutional Neural Network ApproachesSemantic segmentation with Convolutional Neural Network Approaches
Semantic segmentation with Convolutional Neural Network Approaches
UMBC
 
VJAI Paper Reading#3-KDD2019-ClusterGCN
VJAI Paper Reading#3-KDD2019-ClusterGCNVJAI Paper Reading#3-KDD2019-ClusterGCN
VJAI Paper Reading#3-KDD2019-ClusterGCN
Dat Nguyen
 
3D Reconstruction from Multiple uncalibrated 2D Images of an Object
3D Reconstruction from Multiple uncalibrated 2D Images of an Object3D Reconstruction from Multiple uncalibrated 2D Images of an Object
3D Reconstruction from Multiple uncalibrated 2D Images of an Object
Ankur Tyagi
 
Geometry Batching Using Texture-Arrays
Geometry Batching Using Texture-ArraysGeometry Batching Using Texture-Arrays
Geometry Batching Using Texture-Arrays
Matthias Trapp
 
Final Review
Final ReviewFinal Review
Final Review
Afzal Meeran
 
3D reconstruction
3D reconstruction3D reconstruction
3D reconstruction
Jorge Leandro, Ph.D.
 
Introductory Level of SLAM Seminar
Introductory Level of SLAM SeminarIntroductory Level of SLAM Seminar
Introductory Level of SLAM Seminar
Dong-Won Shin
 
Reading group - Week 2 - Trajectory Pooled Deep-Convolutional Descriptors (TDD)
Reading group - Week 2 - Trajectory Pooled Deep-Convolutional Descriptors (TDD)Reading group - Week 2 - Trajectory Pooled Deep-Convolutional Descriptors (TDD)
Reading group - Week 2 - Trajectory Pooled Deep-Convolutional Descriptors (TDD)
Saimunur Rahman
 
DTAM: Dense Tracking and Mapping in Real-Time, Robot vision Group
DTAM: Dense Tracking and Mapping in Real-Time, Robot vision GroupDTAM: Dense Tracking and Mapping in Real-Time, Robot vision Group
DTAM: Dense Tracking and Mapping in Real-Time, Robot vision Group
Lihang Li
 
Seed net automatic seed generation with deep reinforcement learning for robus...
Seed net automatic seed generation with deep reinforcement learning for robus...Seed net automatic seed generation with deep reinforcement learning for robus...
Seed net automatic seed generation with deep reinforcement learning for robus...
NAVER Engineering
 
Motion Estimation in h.264 encoder
Motion Estimation in h.264 encoderMotion Estimation in h.264 encoder
Motion Estimation in h.264 encoder
Talal Khaliq
 
Model compression
Model compressionModel compression
Model compression
Nanhee Kim
 
FastCampus 2018 SLAM Workshop
FastCampus 2018 SLAM WorkshopFastCampus 2018 SLAM Workshop
FastCampus 2018 SLAM Workshop
Dong-Won Shin
 
30th コンピュータビジョン勉強会@関東 DynamicFusion
30th コンピュータビジョン勉強会@関東 DynamicFusion30th コンピュータビジョン勉強会@関東 DynamicFusion
30th コンピュータビジョン勉強会@関東 DynamicFusion
Hiroki Mizuno
 
Single Image Depth Estimation using frequency domain analysis and Deep learning
Single Image Depth Estimation using frequency domain analysis and Deep learningSingle Image Depth Estimation using frequency domain analysis and Deep learning
Single Image Depth Estimation using frequency domain analysis and Deep learning
Ahan M R
 
A Novel Background Subtraction Algorithm for Dynamic Texture Scenes
A Novel Background Subtraction Algorithm for Dynamic Texture ScenesA Novel Background Subtraction Algorithm for Dynamic Texture Scenes
A Novel Background Subtraction Algorithm for Dynamic Texture Scenes
IJMER
 
Semantic segmentation with Convolutional Neural Network Approaches
Semantic segmentation with Convolutional Neural Network ApproachesSemantic segmentation with Convolutional Neural Network Approaches
Semantic segmentation with Convolutional Neural Network Approaches
UMBC
 
VJAI Paper Reading#3-KDD2019-ClusterGCN
VJAI Paper Reading#3-KDD2019-ClusterGCNVJAI Paper Reading#3-KDD2019-ClusterGCN
VJAI Paper Reading#3-KDD2019-ClusterGCN
Dat Nguyen
 
3D Reconstruction from Multiple uncalibrated 2D Images of an Object
3D Reconstruction from Multiple uncalibrated 2D Images of an Object3D Reconstruction from Multiple uncalibrated 2D Images of an Object
3D Reconstruction from Multiple uncalibrated 2D Images of an Object
Ankur Tyagi
 
Geometry Batching Using Texture-Arrays
Geometry Batching Using Texture-ArraysGeometry Batching Using Texture-Arrays
Geometry Batching Using Texture-Arrays
Matthias Trapp
 
Introductory Level of SLAM Seminar
Introductory Level of SLAM SeminarIntroductory Level of SLAM Seminar
Introductory Level of SLAM Seminar
Dong-Won Shin
 

Similar to ImageNet classification with deep convolutional neural networks(2012) (20)

Deep learning requirement and notes for novoice
Deep learning requirement and notes for novoiceDeep learning requirement and notes for novoice
Deep learning requirement and notes for novoice
AmmarAhmedSiddiqui2
 
Introduction to Neural Networks and Deep Learning
Introduction to Neural Networks and Deep LearningIntroduction to Neural Networks and Deep Learning
Introduction to Neural Networks and Deep Learning
Vahid Mirjalili
 
Lecture 6: Convolutional Neural Networks
Lecture 6: Convolutional Neural NetworksLecture 6: Convolutional Neural Networks
Lecture 6: Convolutional Neural Networks
Sang Jun Lee
 
Mask-RCNN for Instance Segmentation
Mask-RCNN for Instance SegmentationMask-RCNN for Instance Segmentation
Mask-RCNN for Instance Segmentation
Dat Nguyen
 
Hardware Acceleration for Machine Learning
Hardware Acceleration for Machine LearningHardware Acceleration for Machine Learning
Hardware Acceleration for Machine Learning
CastLabKAIST
 
Graph Regularised Hashing
Graph Regularised HashingGraph Regularised Hashing
Graph Regularised Hashing
Sean Moran
 
Network Deconvolution review [cdm]
Network Deconvolution review [cdm]Network Deconvolution review [cdm]
Network Deconvolution review [cdm]
Dongmin Choi
 
Efficient Variable Size Template Matching Using Fast Normalized Cross Correla...
Efficient Variable Size Template Matching Using Fast Normalized Cross Correla...Efficient Variable Size Template Matching Using Fast Normalized Cross Correla...
Efficient Variable Size Template Matching Using Fast Normalized Cross Correla...
Gurbinder Gill
 
FPL15 talk: Deep Convolutional Neural Network on FPGA
FPL15 talk: Deep Convolutional Neural Network on FPGAFPL15 talk: Deep Convolutional Neural Network on FPGA
FPL15 talk: Deep Convolutional Neural Network on FPGA
Hiroki Nakahara
 
Superpixel algorithms (whatershed, mean-shift, SLIC, BSLIC), Foolad
Superpixel algorithms (whatershed, mean-shift, SLIC, BSLIC), FooladSuperpixel algorithms (whatershed, mean-shift, SLIC, BSLIC), Foolad
Superpixel algorithms (whatershed, mean-shift, SLIC, BSLIC), Foolad
Shima Foolad
 
Hanjun Dai, PhD Student, School of Computational Science and Engineering, Geo...
Hanjun Dai, PhD Student, School of Computational Science and Engineering, Geo...Hanjun Dai, PhD Student, School of Computational Science and Engineering, Geo...
Hanjun Dai, PhD Student, School of Computational Science and Engineering, Geo...
MLconf
 
Lecture 5: Convolutional Neural Network Models
Lecture 5: Convolutional Neural Network ModelsLecture 5: Convolutional Neural Network Models
Lecture 5: Convolutional Neural Network Models
Mohamed Loey
 
B.tech_project_ppt.pptx
B.tech_project_ppt.pptxB.tech_project_ppt.pptx
B.tech_project_ppt.pptx
supratikmondal6
 
Cnn
CnnCnn
Cnn
Mehrnaz Faraz
 
Convolutional Neural Networks
Convolutional Neural NetworksConvolutional Neural Networks
Convolutional Neural Networks
milad abbasi
 
3D Brain Image Segmentation Model using Deep Learning and Hidden Markov Rando...
3D Brain Image Segmentation Model using Deep Learning and Hidden Markov Rando...3D Brain Image Segmentation Model using Deep Learning and Hidden Markov Rando...
3D Brain Image Segmentation Model using Deep Learning and Hidden Markov Rando...
EL-Hachemi Guerrout
 
NVIDIA 深度學習教育機構 (DLI): Image segmentation with tensorflow
NVIDIA 深度學習教育機構 (DLI): Image segmentation with tensorflowNVIDIA 深度學習教育機構 (DLI): Image segmentation with tensorflow
NVIDIA 深度學習教育機構 (DLI): Image segmentation with tensorflow
NVIDIA Taiwan
 
Convolutional neural network in deep learning
Convolutional neural network in deep learningConvolutional neural network in deep learning
Convolutional neural network in deep learning
priyasjcettnj
 
Convolutional neural network in deep learning
Convolutional neural network in deep learningConvolutional neural network in deep learning
Convolutional neural network in deep learning
priyasjcettnj
 
BLE Localiser for iOS Conf SG 2017
BLE Localiser for iOS Conf SG 2017BLE Localiser for iOS Conf SG 2017
BLE Localiser for iOS Conf SG 2017
yeokm1
 
Deep learning requirement and notes for novoice
Deep learning requirement and notes for novoiceDeep learning requirement and notes for novoice
Deep learning requirement and notes for novoice
AmmarAhmedSiddiqui2
 
Introduction to Neural Networks and Deep Learning
Introduction to Neural Networks and Deep LearningIntroduction to Neural Networks and Deep Learning
Introduction to Neural Networks and Deep Learning
Vahid Mirjalili
 
Lecture 6: Convolutional Neural Networks
Lecture 6: Convolutional Neural NetworksLecture 6: Convolutional Neural Networks
Lecture 6: Convolutional Neural Networks
Sang Jun Lee
 
Mask-RCNN for Instance Segmentation
Mask-RCNN for Instance SegmentationMask-RCNN for Instance Segmentation
Mask-RCNN for Instance Segmentation
Dat Nguyen
 
Hardware Acceleration for Machine Learning
Hardware Acceleration for Machine LearningHardware Acceleration for Machine Learning
Hardware Acceleration for Machine Learning
CastLabKAIST
 
Graph Regularised Hashing
Graph Regularised HashingGraph Regularised Hashing
Graph Regularised Hashing
Sean Moran
 
Network Deconvolution review [cdm]
Network Deconvolution review [cdm]Network Deconvolution review [cdm]
Network Deconvolution review [cdm]
Dongmin Choi
 
Efficient Variable Size Template Matching Using Fast Normalized Cross Correla...
Efficient Variable Size Template Matching Using Fast Normalized Cross Correla...Efficient Variable Size Template Matching Using Fast Normalized Cross Correla...
Efficient Variable Size Template Matching Using Fast Normalized Cross Correla...
Gurbinder Gill
 
FPL15 talk: Deep Convolutional Neural Network on FPGA
FPL15 talk: Deep Convolutional Neural Network on FPGAFPL15 talk: Deep Convolutional Neural Network on FPGA
FPL15 talk: Deep Convolutional Neural Network on FPGA
Hiroki Nakahara
 
Superpixel algorithms (whatershed, mean-shift, SLIC, BSLIC), Foolad
Superpixel algorithms (whatershed, mean-shift, SLIC, BSLIC), FooladSuperpixel algorithms (whatershed, mean-shift, SLIC, BSLIC), Foolad
Superpixel algorithms (whatershed, mean-shift, SLIC, BSLIC), Foolad
Shima Foolad
 
Hanjun Dai, PhD Student, School of Computational Science and Engineering, Geo...
Hanjun Dai, PhD Student, School of Computational Science and Engineering, Geo...Hanjun Dai, PhD Student, School of Computational Science and Engineering, Geo...
Hanjun Dai, PhD Student, School of Computational Science and Engineering, Geo...
MLconf
 
Lecture 5: Convolutional Neural Network Models
Lecture 5: Convolutional Neural Network ModelsLecture 5: Convolutional Neural Network Models
Lecture 5: Convolutional Neural Network Models
Mohamed Loey
 
Convolutional Neural Networks
Convolutional Neural NetworksConvolutional Neural Networks
Convolutional Neural Networks
milad abbasi
 
3D Brain Image Segmentation Model using Deep Learning and Hidden Markov Rando...
3D Brain Image Segmentation Model using Deep Learning and Hidden Markov Rando...3D Brain Image Segmentation Model using Deep Learning and Hidden Markov Rando...
3D Brain Image Segmentation Model using Deep Learning and Hidden Markov Rando...
EL-Hachemi Guerrout
 
NVIDIA 深度學習教育機構 (DLI): Image segmentation with tensorflow
NVIDIA 深度學習教育機構 (DLI): Image segmentation with tensorflowNVIDIA 深度學習教育機構 (DLI): Image segmentation with tensorflow
NVIDIA 深度學習教育機構 (DLI): Image segmentation with tensorflow
NVIDIA Taiwan
 
Convolutional neural network in deep learning
Convolutional neural network in deep learningConvolutional neural network in deep learning
Convolutional neural network in deep learning
priyasjcettnj
 
Convolutional neural network in deep learning
Convolutional neural network in deep learningConvolutional neural network in deep learning
Convolutional neural network in deep learning
priyasjcettnj
 
BLE Localiser for iOS Conf SG 2017
BLE Localiser for iOS Conf SG 2017BLE Localiser for iOS Conf SG 2017
BLE Localiser for iOS Conf SG 2017
yeokm1
 
Ad

Recently uploaded (20)

Using the Artificial Neural Network to Predict the Axial Strength and Strain ...
Using the Artificial Neural Network to Predict the Axial Strength and Strain ...Using the Artificial Neural Network to Predict the Axial Strength and Strain ...
Using the Artificial Neural Network to Predict the Axial Strength and Strain ...
Journal of Soft Computing in Civil Engineering
 
Environment .................................
Environment .................................Environment .................................
Environment .................................
shadyozq9
 
JRR Tolkien’s Lord of the Rings: Was It Influenced by Nordic Mythology, Homer...
JRR Tolkien’s Lord of the Rings: Was It Influenced by Nordic Mythology, Homer...JRR Tolkien’s Lord of the Rings: Was It Influenced by Nordic Mythology, Homer...
JRR Tolkien’s Lord of the Rings: Was It Influenced by Nordic Mythology, Homer...
Reflections on Morality, Philosophy, and History
 
sss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptx
sss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptx
sss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptx
ajayrm685
 
OPTIMIZING DATA INTEROPERABILITY IN AGILE ORGANIZATIONS: INTEGRATING NONAKA’S...
OPTIMIZING DATA INTEROPERABILITY IN AGILE ORGANIZATIONS: INTEGRATING NONAKA’S...OPTIMIZING DATA INTEROPERABILITY IN AGILE ORGANIZATIONS: INTEGRATING NONAKA’S...
OPTIMIZING DATA INTEROPERABILITY IN AGILE ORGANIZATIONS: INTEGRATING NONAKA’S...
ijdmsjournal
 
Frontend Architecture Diagram/Guide For Frontend Engineers
Frontend Architecture Diagram/Guide For Frontend EngineersFrontend Architecture Diagram/Guide For Frontend Engineers
Frontend Architecture Diagram/Guide For Frontend Engineers
Michael Hertzberg
 
Artificial intelligence and machine learning.pptx
Artificial intelligence and machine learning.pptxArtificial intelligence and machine learning.pptx
Artificial intelligence and machine learning.pptx
rakshanatarajan005
 
Applications of Centroid in Structural Engineering
Applications of Centroid in Structural EngineeringApplications of Centroid in Structural Engineering
Applications of Centroid in Structural Engineering
suvrojyotihalder2006
 
22PCOAM16 ML Unit 3 Full notes PDF & QB.pdf
22PCOAM16 ML Unit 3 Full notes PDF & QB.pdf22PCOAM16 ML Unit 3 Full notes PDF & QB.pdf
22PCOAM16 ML Unit 3 Full notes PDF & QB.pdf
Guru Nanak Technical Institutions
 
Little Known Ways To 3 Best sites to Buy Linkedin Accounts.pdf
Little Known Ways To 3 Best sites to Buy Linkedin Accounts.pdfLittle Known Ways To 3 Best sites to Buy Linkedin Accounts.pdf
Little Known Ways To 3 Best sites to Buy Linkedin Accounts.pdf
gori42199
 
ML_Unit_VI_DEEP LEARNING_Introduction to ANN.pdf
ML_Unit_VI_DEEP LEARNING_Introduction to ANN.pdfML_Unit_VI_DEEP LEARNING_Introduction to ANN.pdf
ML_Unit_VI_DEEP LEARNING_Introduction to ANN.pdf
rameshwarchintamani
 
Slide share PPT of SOx control technologies.pptx
Slide share PPT of SOx control technologies.pptxSlide share PPT of SOx control technologies.pptx
Slide share PPT of SOx control technologies.pptx
vvsasane
 
Control Methods of Noise Pollutions.pptx
Control Methods of Noise Pollutions.pptxControl Methods of Noise Pollutions.pptx
Control Methods of Noise Pollutions.pptx
vvsasane
 
Modeling the Influence of Environmental Factors on Concrete Evaporation Rate
Modeling the Influence of Environmental Factors on Concrete Evaporation RateModeling the Influence of Environmental Factors on Concrete Evaporation Rate
Modeling the Influence of Environmental Factors on Concrete Evaporation Rate
Journal of Soft Computing in Civil Engineering
 
vtc2018fall_otfs_tutorial_presentation_1.pdf
vtc2018fall_otfs_tutorial_presentation_1.pdfvtc2018fall_otfs_tutorial_presentation_1.pdf
vtc2018fall_otfs_tutorial_presentation_1.pdf
RaghavaGD1
 
Slide share PPT of NOx control technologies.pptx
Slide share PPT of  NOx control technologies.pptxSlide share PPT of  NOx control technologies.pptx
Slide share PPT of NOx control technologies.pptx
vvsasane
 
Generative AI & Large Language Models Agents
Generative AI & Large Language Models AgentsGenerative AI & Large Language Models Agents
Generative AI & Large Language Models Agents
aasgharbee22seecs
 
Agents chapter of Artificial intelligence
Agents chapter of Artificial intelligenceAgents chapter of Artificial intelligence
Agents chapter of Artificial intelligence
DebdeepMukherjee9
 
Water Industry Process Automation & Control Monthly May 2025
Water Industry Process Automation & Control Monthly May 2025Water Industry Process Automation & Control Monthly May 2025
Water Industry Process Automation & Control Monthly May 2025
Water Industry Process Automation & Control
 
2.3 Genetically Modified Organisms (1).ppt
2.3 Genetically Modified Organisms (1).ppt2.3 Genetically Modified Organisms (1).ppt
2.3 Genetically Modified Organisms (1).ppt
rakshaiya16
 
Environment .................................
Environment .................................Environment .................................
Environment .................................
shadyozq9
 
sss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptx
sss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptx
sss1.pptxsss1.pptxsss1.pptxsss1.pptxsss1.pptx
ajayrm685
 
OPTIMIZING DATA INTEROPERABILITY IN AGILE ORGANIZATIONS: INTEGRATING NONAKA’S...
OPTIMIZING DATA INTEROPERABILITY IN AGILE ORGANIZATIONS: INTEGRATING NONAKA’S...OPTIMIZING DATA INTEROPERABILITY IN AGILE ORGANIZATIONS: INTEGRATING NONAKA’S...
OPTIMIZING DATA INTEROPERABILITY IN AGILE ORGANIZATIONS: INTEGRATING NONAKA’S...
ijdmsjournal
 
Frontend Architecture Diagram/Guide For Frontend Engineers
Frontend Architecture Diagram/Guide For Frontend EngineersFrontend Architecture Diagram/Guide For Frontend Engineers
Frontend Architecture Diagram/Guide For Frontend Engineers
Michael Hertzberg
 
Artificial intelligence and machine learning.pptx
Artificial intelligence and machine learning.pptxArtificial intelligence and machine learning.pptx
Artificial intelligence and machine learning.pptx
rakshanatarajan005
 
Applications of Centroid in Structural Engineering
Applications of Centroid in Structural EngineeringApplications of Centroid in Structural Engineering
Applications of Centroid in Structural Engineering
suvrojyotihalder2006
 
Little Known Ways To 3 Best sites to Buy Linkedin Accounts.pdf
Little Known Ways To 3 Best sites to Buy Linkedin Accounts.pdfLittle Known Ways To 3 Best sites to Buy Linkedin Accounts.pdf
Little Known Ways To 3 Best sites to Buy Linkedin Accounts.pdf
gori42199
 
ML_Unit_VI_DEEP LEARNING_Introduction to ANN.pdf
ML_Unit_VI_DEEP LEARNING_Introduction to ANN.pdfML_Unit_VI_DEEP LEARNING_Introduction to ANN.pdf
ML_Unit_VI_DEEP LEARNING_Introduction to ANN.pdf
rameshwarchintamani
 
Slide share PPT of SOx control technologies.pptx
Slide share PPT of SOx control technologies.pptxSlide share PPT of SOx control technologies.pptx
Slide share PPT of SOx control technologies.pptx
vvsasane
 
Control Methods of Noise Pollutions.pptx
Control Methods of Noise Pollutions.pptxControl Methods of Noise Pollutions.pptx
Control Methods of Noise Pollutions.pptx
vvsasane
 
vtc2018fall_otfs_tutorial_presentation_1.pdf
vtc2018fall_otfs_tutorial_presentation_1.pdfvtc2018fall_otfs_tutorial_presentation_1.pdf
vtc2018fall_otfs_tutorial_presentation_1.pdf
RaghavaGD1
 
Slide share PPT of NOx control technologies.pptx
Slide share PPT of  NOx control technologies.pptxSlide share PPT of  NOx control technologies.pptx
Slide share PPT of NOx control technologies.pptx
vvsasane
 
Generative AI & Large Language Models Agents
Generative AI & Large Language Models AgentsGenerative AI & Large Language Models Agents
Generative AI & Large Language Models Agents
aasgharbee22seecs
 
Agents chapter of Artificial intelligence
Agents chapter of Artificial intelligenceAgents chapter of Artificial intelligence
Agents chapter of Artificial intelligence
DebdeepMukherjee9
 
2.3 Genetically Modified Organisms (1).ppt
2.3 Genetically Modified Organisms (1).ppt2.3 Genetically Modified Organisms (1).ppt
2.3 Genetically Modified Organisms (1).ppt
rakshaiya16
 
Ad

ImageNet classification with deep convolutional neural networks(2012)

  • 1. ImageNet Classification with Deep Convolutional Neural Networks 신우철
  • 2. Introduction 1. Trained one of the largest CNN on ImageNet data. The advantages of CNN are 1) CNN’s prior knowledge, which are stationarity of statistics and locality of pixel dependencies, 2) its easiness to be controlled, varying its depth and breath, contributing to fewer parameters and easier training. 2. Implemented highly-optimized GPU implementation to facilitate the training of large CNNs on high resolution images. 3. Introduced new features to improve performance, reduce training time, and prevent overfitting.
  • 3. Dataset • Down-sampled ImageNet images to 256 x 256. Trained on centered raw RGB values of pixels. 1) Rescaled the image such that the shorter side was of length 256 2) Cropped out the central 256 x 256 patch from the resulting image of 1). 3) Subtracted the mean activity over the training set from each pixel. cf) NORB MNIST LabelMe
  • 4. Architecture • 8 layers = 5 Convolutional + 3 Fully-connected • Newly introduced features 1) ReLU Nonlinearity • Much faster to train since it is non-saturating • Nonlinear |tanh(x)| function focuses on preventing overfitting, while ReLU focuses on fast learning of large models on large datasets
  • 5. Architecture 2) Training on two GPUs • Cross-GPU parallelization, while the GPUs only communicate in certain layers. This is to tune the amount of computation by communication.
  • 6. Architecture 3) Local Response Normalization • Since the architecture uses ReLU, one high activation value can affect adjacent activation values in convolution or pooling. Therefore, LRN is conducted. filter : the activity of a neuron computed by applying kernel i at position (x,y) : the response-normalized activity N : #all kernels n : adjacent #kernels at position (x,y)
  • 7. Architecture 3) Local Response Normalization filter 0 filter 1 filter 2 filter 3 1 2 3 1 2 1 2 1 2 4 2 1 4 5 6 2 3 2 3 2 3 5 2 1 7 8 9 3 4 3 4 3 4 2 2 4 0.50 0.25 0.30 0.17 0.22 0.07 0.10 0.11 0.33 0.20 0.40 0.20 0.20 0.15 0.15 0.07 0.08 0.04 0.08 0.12 0.21 0.15 0.25 0.10 0.12 0.10 0.10 0.04 0.04 0.03 0.14 0.10 0.10 0.10 0.15 0.13 k alpha beta n N 0 1 1 2 4 = 2 {0 + 1 x 12 + 22 + 42 }1
  • 8. Architecture 4) Overlapping Pooling • Overlapping reduces overfitting compared to non-overlapping pooling.
  • 10. Convolutional layer Kernel size = 11 Stride = 4 Filter = 96 Zero-padding = 0 (227 – 11) / 4 + 1 = 55 Maxpooling Kernel size(z) = 3 Stride(s) = 2 Convolutional layer Kernel size = 5 Stride = 1 Filter = 256 Zero-padding = 2 (55 – 3) / 2 + 1 = 27 (27 +2 * 2 – 5) / 1 + 1 = 27 Local response normalization
  • 11. Convolutional layer Kernel size = 3 Stride = 1 Filter = 384 Zero-padding = 1 (27 + 1 * 2– 3) / 1 + 1 = 27 Convolutional layer Kernel size = 3 Stride = 1 Filter = 384 Zero-padding = 1 (13 +1 * 2– 3) / 1 + 1 = 13 Maxpooling Kernel size(z) = 3 Stride = 2 (27 – 3) / 2 + 1 = 13 Local response normalization
  • 12. Convolutional layer Kernel size = 3 Stride = 1 Filter = 256 Zero-padding = 1 (13 +1 * 2– 3) / 1 + 1 = 13 Maxpooling Kernel size(z) = 3 Stride = 2 (13 – 3) / 2 + 1 = 6 Flatten 6 * 6 * 256 = 9216 Fully connected 4096 Fully connected 4096 Fully connected 1000(softmax)
  • 13. Reducing Overfitting 1) Data Augmentation (1) Image translations and horizontal reflections Train set • Image translations (x (256-224) * (256-224)) • Horizontal reflections (x 2) Total : (256-224) * (256-224) * 2 = 2048 Test set • Image translations (x 5) • Horizontal reflections (x 2) • Total: 5 * 2 = 10
  • 14. Reducing Overfitting 1) Data Augmentation (2) Altering intensity of RGB channels (performing PCA) 2) Dropout • Applied dropout on first two FC layers with p = 0.5 Pi : eigen vector  : eigen value  : random =
  • 15. Details of Learning • SGD with batch size of 128 examples • Momentum = 0.9 • Weight decay = 0.0005 • Weight initialization : N(0, 0.012 ) • Neuron biases initialization: Conv layers = 0 FC layers = 1 • Learning rate Initialized at 0.01 and reduced three times prior to termination. Reduction was done by dividing learning rate by 10 when the validation error rate stopped improving with the current learning rate.
  • 17. Results • Result of restricted connectivity between two GPUs result in specialization. Kernels on GPU 1 are largely color-agnostic, while kernels on GPU 2 are largely color-specific. • Perform kNN at the last 4096-dimensional hidden layer shows that images are semantically similar.
  翻译: