国科大中科院人工智能与机器学习12-CNN_Beyond42

上传人:ca****in 文档编号:123895385 上传时间:2022-07-23 格式:PPTX 页数:42 大小:3.14MB
返回 下载 相关 举报
国科大中科院人工智能与机器学习12-CNN_Beyond42_第1页
第1页 / 共42页
国科大中科院人工智能与机器学习12-CNN_Beyond42_第2页
第2页 / 共42页
国科大中科院人工智能与机器学习12-CNN_Beyond42_第3页
第3页 / 共42页
点击查看更多>>
资源描述
卷积卷积神经网络神经网络CNN及其变种及其变种山世光中科院计算所Institute of Computing Technology,Chinese Academy of SciencesCNN的早期历史的早期历史n卷积神经网络CNNK.Fukushima,“Neocognitron:A self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in position,”Biological Cybernetics,vol.36,pp.193202,1980Y.LeCun,B.Boser,J.S.Denker,D.Henderson,R.E.Howard,W.Hubbard,and L.D.Jackel,“Backpropagation applied to handwritten zip code recognition,”Neural Computation,vol.1,no.4,pp.541551,1989Y.Le Cun,L.Bottou,Y.Bengio,and P.Haffner,“Gradient-based learning applied to document recognition,”Proceedings of the IEEE,vol.86,no.11,pp.22782324,19982Institute of Computing Technology,Chinese Academy of SciencesDL时代的时代的CNN扩展扩展nA Krizhevsky,I Sutskever,GE Hinton.ImageNet classification with deep convolutional neural networks.NIPS2012nY.Jia et al.Caffe:Convolutional Architecture for Fast Feature Embedding.ACM MM2014nK.Simonyan,A.Zisserman.Very deep convolutional networks for large-scale image recognition.arXiv preprint arXiv:1409.1556,2014nC.Szegedy,W.Liu,Y.Jia,P.Sermanet,S.Reed,D.Anguelov,D.Erhan,V.Vanhoucke,A.Rabinovich.Going deeper with convolutions.CVPR2015(&arXiv:1409.4842,2014)3Institute of Computing Technology,Chinese Academy of Sciences卷积卷积示例示例4Institute of Computing Technology,Chinese Academy of Sciences卷积卷积形式化形式化n5Institute of Computing Technology,Chinese Academy of Sciences卷积卷积why?n1.sparse interactions有限连接,Kernel比输入小连接数少很多,学习难度小,计算复杂度低nm个节点与n个节点相连O(mn)n限定k(m)个节点与n个节点相连,则为O(kn)6Institute of Computing Technology,Chinese Academy of Sciences卷积卷积why?n1.sparse interactions有限连接,Kernel比输入小连接数少很多,学习难度小,计算复杂度低nm个节点与n个节点相连O(mn)n限定k(m)个节点与n个节点相连,则为O(kn)7Institute of Computing Technology,Chinese Academy of Sciences卷积卷积why?n1.sparse interactions有限(稀疏)连接nKernel比输入小n局部连接局部连接连接数少很多n学习难度小n计算复杂度低层级感受野(生物启发)n越高层的神经元,感受野越大8Institute of Computing Technology,Chinese Academy of Sciences卷积卷积why?n2.Parameter Sharing(参数共享)Tied weights进一步极大的缩减参数数量n3.Equivariant representations等变性配合Pooling可以获得平移不变性n对scale和rotation不具有此属性9Institute of Computing Technology,Chinese Academy of SciencesCNN的基本结构的基本结构n三个步骤卷积n突触前激活,net非线性激活nDetectorPoolingnLayer的两种定义复杂定义简单定义n有些层没有参数10Institute of Computing Technology,Chinese Academy of SciencesPooling11n定义(没有需要学习的参数)replaces the output of the net at a certain location with a summary statistic of the nearby outputsn种类max pooling(weighted)average poolingInstitute of Computing Technology,Chinese Academy of SciencesWhy Pooling?12n获取不变性小的平移不变性:有即可,不管在哪里n很强的先验假设The function the layer learns must be invariant to small translationsInstitute of Computing Technology,Chinese Academy of SciencesWhy Pooling?13n获取不变性小的平移不变性:有即可,不管在哪里旋转不变性?n9个不同朝向的kernels(模板)0.20.610.10.50.30.020.050.1Institute of Computing Technology,Chinese Academy of SciencesWhy Pooling?14n获取不变性小的平移不变性:有即可,不管在哪里旋转不变性?n9个不同朝向的kernels(模板)0.50.30.0210.40.30.60.30.1Institute of Computing Technology,Chinese Academy of SciencesPooling与下采样结合与下采样结合n更好的获取平移不变性n更高的计算效率(减少了神经元数)15Institute of Computing Technology,Chinese Academy of Sciences从全连接到有限连接从全连接到有限连接n部分链接权重被强制设置为0通常:非邻接神经元,仅保留相邻的神经元全连接网络的特例,大量连接权重为016Institute of Computing Technology,Chinese Academy of SciencesWhy Convolution&Pooling?na prior probability distribution over the parameters of a model that encodes our beliefs about what models are reasonable,before we have seen any data.n模型参数的先验概率分布(No free lunch)在见到任何数据之前,我们的信念(经验)告诉我们,什么样的模型参数是合理的nLocal connections;对平移的不变性;tied weigts来自生物神经系统的启发17Institute of Computing Technology,Chinese Academy of Sciences源源起:起:Neocognitron(1980)nSimplecomplexLower orderhigh order18K.Fukushima,“Neocognitron:A self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in position,”Biological Cybernetics,vol.36,pp.193202,1980Local ConnectionInstitute of Computing Technology,Chinese Academy of Sciences源起:源起:Neocognitron(1980)19Institute of Computing Technology,Chinese Academy of Sciences源起:源起:Neocognitron(1980)n训练方法分层自组织ncompetitive learning 无监督n输出层独立训练有监督20Institute of Computing Technology,Chinese Academy of SciencesLeCun-CNN1989用于字符识别用于字符识别n简化了Neocognitron的结构n训练方法监督训练BP算法n正切函数收敛更快,Sigmoid Loss,SGDn用于邮编识别大量应用21Institute of Computing Technology,Chinese Academy of SciencesLeCun-CNN1989用于字符识别用于字符识别n输入16x16图像nL1H112个5x5 kernel8x8个神经元nL2-H212个5x5x8 kernel4x4个神经元nL3H330个神经元nL4输出层10个神经元n总连接数5*5*12*64+5*5*8*12*16+192*30,约66,000个22Institute of Computing Technology,Chinese Academy of SciencesLeCun-CNN1989用于字符识别用于字符识别nTied weights对同一个feature map,kernel对不同位置是相同的!23Institute of Computing Technology,Chinese Academy of SciencesLeCun-CNN1989用于字符识别用于字符识别24Institute of Computing Technology,Chinese Academy of Sciences1998年年LeNet数字数字/字符识别字符识别nLeNet-5Feature mapna set of units whose weighs are constrained to be identical.25Institute of Computing Technology,Chinese Academy of Sciences1998年年LeNet数字数字/字符识别字符识别n 例如:C3层参数个数(3*6+4*9+6*1)*25+16=1516 26Institute of Computing Technology,Chinese Academy of Sciences后续:后续:CNN用于目标检测与识别用于目标检测与识别27Institute of Computing Technology,Chinese Academy of SciencesAlexNet for ImageNet(2012)n大规模CNN网络650K神经元60M参数使用了各种技巧nDropoutnData augmentnReLUnLocal Response NormalizationnContrast normalizationn.28Krizhevsky,Alex,Ilya Sutskever,and Geoffrey E.Hinton.Imagenet classification with deep convolutional neural networks.Advances in neural information processing systems.2012.Institute of Computing Technology,Chinese Academy of SciencesAlexNet for ImageNet(2012)nReLU激活函数29Institute of Computing Technology,Chinese Academy of SciencesAlexNet for ImageNet(2012)n实现2块GPU卡输入层150,528其它层253,440186,624 64,896 64,896 43,264 4096 4096 1000.30Krizhevsky,Alex,Ilya Sutskever,and Geoffrey E.Hinton.Imagenet classification with deep convolutional neural networks.Advances in neural information processing systems.2012.Institute of Computing Technology,Chinese Academy of SciencesAlexNet for ImageNet(2012)nImageNet物体分类任务上1000类,1,431,167幅图像31RankNameError rates(TOP5)Description1U.Toronto0.153Deep learning2U.Tokyo0.261Hand-crafted features and learning models.Bottleneck.3U.Oxford0.2704Xerox/INRIA0.271Krizhevsky,Alex,Ilya Sutskever,and Geoffrey E.Hinton.Imagenet classification with deep convolutional neural networks.Advances in neural information processing systems.2012.Institute of Computing Technology,Chinese Academy of SciencesAlexNet for ImageNetn深度的重要性32网络深度87664参数数量60M44M10M59M10M性能损失0%1.1%5.7%3.0%33.5%Krizhevsky,Alex,Ilya Sutskever,and Geoffrey E.Hinton.Imagenet classification with deep convolutional neural networks.Advances in neural information processing systems.2012.Institute of Computing Technology,Chinese Academy of SciencesVGG Net(2014)n多个stage每个stage多个卷积层n卷积采样间隔1x1n卷积核大小3x31个Pooling层(2x2)n16-19层n多尺度融合K.Simonyan,A.Zisserman.Very deep convolutional networks for large-scale image recognition.arXiv preprint arXiv:1409.1556,2014Institute of Computing Technology,Chinese Academy of SciencesVGG Net(2014)n几种配置Cov3-64:n3x3感受野n64个channel34K.Simonyan,A.Zisserman.Very deep convolutional networks for large-scale image recognition.arXiv preprint arXiv:1409.1556,2014Institute of Computing Technology,Chinese Academy of SciencesVGG Net(2014)35K.Simonyan,A.Zisserman.Very deep convolutional networks for large-scale image recognition.arXiv preprint arXiv:1409.1556,2014Institute of Computing Technology,Chinese Academy of SciencesGoogLeNet(2014)n超大规模22个卷积层的网络计算复杂度是AlexNet的4倍左右C.Szegedy,W.Liu,Y.Jia,P.Sermanet,S.Reed,D.Anguelov,D.Erhan,V.Vanhoucke,A.Rabinovich.Going deeper with convolutions.CVPR2015(&arXiv:1409.4842,2014)Institute of Computing Technology,Chinese Academy of SciencesGoogLeNet(2014)n超大规模24层网络nInception结构提取不同scale的特征然后串接起来1x1 convolutions3x3 convolutions5x5 convolutionsFilter concatenationPrevious layerC.Szegedy,W.Liu,Y.Jia,P.Sermanet,S.Reed,D.Anguelov,D.Erhan,V.Vanhoucke,A.Rabinovich.Going deeper with convolutions.CVPR2015(&arXiv:1409.4842,2014)Institute of Computing Technology,Chinese Academy of SciencesGoogLeNet(2014)n超大规模24层网络nInception结构提取不同scale的特征,然后串接起来增加1x1的卷积:把响应图的数量缩小了1x1 convolutions3x3 convolutions5x5 convolutionsFilter concatenationPrevious layer3x3 max pooling1x1 convolutions1x1 convolutions1x1 convolutionsC.Szegedy,W.Liu,Y.Jia,P.Sermanet,S.Reed,D.Anguelov,D.Erhan,V.Vanhoucke,A.Rabinovich.Going deeper with convolutions.CVPR2015(&arXiv:1409.4842,2014)Institute of Computing Technology,Chinese Academy of SciencesGoogLeNet(2014)39C.Szegedy,W.Liu,Y.Jia,P.Sermanet,S.Reed,D.Anguelov,D.Erhan,V.Vanhoucke,A.Rabinovich.Going deeper with convolutions.CVPR2015(&arXiv:1409.4842,2014)Institute of Computing Technology,Chinese Academy of SciencesGoogLeNet(2014)n在ImageNet上1000类物体分类上的性能40C.Szegedy,W.Liu,Y.Jia,P.Sermanet,S.Reed,D.Anguelov,D.Erhan,V.Vanhoucke,A.Rabinovich.Going deeper with convolutions.CVPR2015(&arXiv:1409.4842,2014)Institute of Computing Technology,Chinese Academy of Sciences41谢谢!谢谢!42
展开阅读全文
相关资源
正为您匹配相似的精品文档
相关搜索

最新文档


当前位置:首页 > 商业管理 > 销售管理


copyright@ 2023-2025  zhuangpeitu.com 装配图网版权所有   联系电话:18123376007

备案号:ICP2024067431-1 川公网安备51140202000466号


本站为文档C2C交易模式,即用户上传的文档直接被用户下载,本站只是中间服务平台,本站所有文档下载所得的收益归上传人(含作者)所有。装配图网仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对上载内容本身不做任何修改或编辑。若文档所含内容侵犯了您的版权或隐私,请立即通知装配图网,我们立即给予删除!