首页 | 官方网站   微博 | 高级检索  
     

通过全局核降低高斯核的局部风险与基于遗传算法的两阶段模型选择
引用本文:常群,王晓龙,林沂蒙,陈清才.通过全局核降低高斯核的局部风险与基于遗传算法的两阶段模型选择[J].计算机研究与发展,2007,44(3):439-444.
作者姓名:常群  王晓龙  林沂蒙  陈清才
作者单位:哈尔滨工业大学计算机科学与技术学院 哈尔滨150001哈尔滨工业大学深圳研究生院媒体与生命科学计算实验室深圳518055(常群,王晓龙),哈尔滨工业大学深圳研究生院媒体与生命科学计算实验室 深圳518055(林沂蒙,陈清才),哈尔滨工业大学深圳研究生院媒体与生命科学计算实验室 深圳518055香港理工大学计算学系香港九龙(Daniel S.Yeung)
摘    要:支持向量分类时,由于样本分布的不均匀性,单宽度的高斯核会在空间的稠密区域产生过学习现象,在稀疏区域产生欠学习现象,即存在局部风险.针对于此,构造了一个全局性次核来降低高斯核产生的局部风险.形成的混合核称为主次核.利用幂级数构造性地给出并证明了主次核的正定性条件,进一步提出了基于遗传算法的两阶段模型选择算法来优化主次核的参数.实验验证了主次核和模型选择法的优越性.

关 键 词:支持向量机  主次核  高斯核  两阶段模型选择  高斯核  局部  风险  基于遗传算法  两阶段模型  选择法  Gaussian  Kernel  Genetic  Algorithms  Based  Model  Selection  Global  Risks  实验验证  参数  优化  选择算法  定性条件  构造性  幂级数  利用
修稿时间:11 3 2005 12:00AM

Reducing Gaussian Kernel's Local Risks by Global Kernel and Two-Stage Model Selection Based on Genetic Algorithms
Daniel S.Yeung,Chang Qun,Wang Xiaolong,Lin Yimeng,Daniel S.Yeung,Chen Qingcai.Reducing Gaussian Kernel''''s Local Risks by Global Kernel and Two-Stage Model Selection Based on Genetic Algorithms[J].Journal of Computer Research and Development,2007,44(3):439-444.
Authors:Daniel SYeung  Chang Qun  Wang Xiaolong  Lin Yimeng  Daniel SYeung  Chen Qingcai
Affiliation:School of Computer Science and Technology, Harbin Institute of Technology, Harbin 150001;Media Computation and Life Science Laboratory, Shenzhen Graduate School, Harbin Institute of Technology, Shenzhen 518055;Department of Computing, Hong Kong Polytechnic University, Kowloon, Hong Kong
Abstract:In classification by support vector machines with the Gaussian kernel, the kernel width defines the generalization scale in the pattern space or in the feature space. However, the Gaussian kernel with constant width is not well adaptive everywhere in the pattern space since the patterns are not evenly distributed. That is, the over-fitting learning will appear in the dense areas and otherwise the under-fitting learning in the sparse areas. To reduce such local risks, a secondary kernel with global character is introduced for the Gaussian kernel. Here the Gaussian kernel is regarded as the primary kernel. The constructed hybrid kernel is called the primary-secondary kernel (PSK). The positive definiteness of PSK with given constraints is proved by virtue of the power series. For support vector machines with PSK, the two-stage model selection based on genetic algorithms is proposed to tune the model parameters. That is, the algorithms firstly tune the model parameters with Gaussian kernel. Then the model parameters with the Gaussian kernel keep unchanged and the model parameters with the secondary kernel are further tuned. The two-stage model selection algorithms aim to overcome the problem of the optimization tendency embodied in the optimization algorithms. For the support vector machines with multiple parameters, the optimization tendency often causes the failure of the model selection. Finally, the experiments demonstrate that PSK performs better than the Gaussian kernel and also validate the efficiency of the proposed model selection algorithms.
Keywords:support vector machine  primary-secondary kernel  Gaussian kernel  two stage model selection
本文献已被 CNKI 维普 万方数据 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司    京ICP备09084417号-23

京公网安备 11010802026262号