首页 | 官方网站   微博 | 高级检索  
     

激活函数在卷积神经网络中的对比研究
引用本文:田娟,李英祥,李彤岩.激活函数在卷积神经网络中的对比研究[J].计算机系统应用,2018,27(7):43-49.
作者姓名:田娟  李英祥  李彤岩
作者单位:成都信息工程大学 通信工程学院, 成都 610225,成都信息工程大学 通信工程学院, 成都 610225,成都信息工程大学 通信工程学院, 成都 610225
基金项目:四川省科技厅重点研发项目(2017FZ0100)
摘    要:近年,深度学习的快速发展致使越来越多的人从事相关的研究工作.但是,许多研究者在搭建深度神经网络模型时只是根据标准算法或改进算法直接搭建,而对算法本身及影响模型性能的因素不甚了解,致使在许多应用中或多或少存在盲目套用现象.通过研究深度神经网络,选择其中的重要影响因素激活函数进行深入研究.首先,分析了激活函数如何影响深度神经网络;接着对激活函数的发展现状及不同激活函数的原理性能进行了分析总结;最后,基于Caffe框架用CNN对Mnist数据集进行分类识别实验,对5种常用激活函数进行综合分析比较,为设计深度神经网络模型时选用激活函数提供参考.

关 键 词:卷积神经网络  激活函数  Caffe  梯度下降法  网络性能
收稿时间:2017/10/26 0:00:00
修稿时间:2017/11/14 0:00:00

Contrastive Study of Activation Function in Convolutional Neural Network
TIAN Juan,LI Ying-Xiang and LI Tong-Yan.Contrastive Study of Activation Function in Convolutional Neural Network[J].Computer Systems& Applications,2018,27(7):43-49.
Authors:TIAN Juan  LI Ying-Xiang and LI Tong-Yan
Affiliation:School of Communication Engineering, Chengdu University of Information Technology, Chengdu 610225, China,School of Communication Engineering, Chengdu University of Information Technology, Chengdu 610225, China and School of Communication Engineering, Chengdu University of Information Technology, Chengdu 610225, China
Abstract:In recent years, the rapid development of deep learning has led more and more people to engage in related research work. However, many researchers construct deep neural network models based on standard algorithms or improved algorithms, but do not understand the algorithm itself and the factors that affect the performance of the model, resulting in more or less blind application in many applications. By studying the deep neural network, the activation function of the important influencing factors was studied. First, the activation function is analyzed to influence the depth neural network. Then, the development of activation function and the principle and performance of different activation functions are analyzed and summarized. Finally, based on the Caffe framework, the CNN is used to classify and identify MNIST data sets. Five kinds of commonly used activation functions are analyzed and compared comprehensively to provide a reference for the selection of activation function in the design of deep neural network model.
Keywords:convolutional neural network  activation function  Caffe  gradient method  network performance
点击此处可从《计算机系统应用》浏览原始摘要信息
点击此处可从《计算机系统应用》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司    京ICP备09084417号-23

京公网安备 11010802026262号