一种处理不均衡多分类问题的特征选择集成方法
DOI:
作者:
作者单位:

江南大学

作者简介:

通讯作者:

中图分类号:

基金项目:

教育部-新华三集团“云数融合”基金项目(2017A13055)


An ensemble learning algorithm for feature selection based on solution to multi-class imbalance data classification
Author:
Affiliation:

1.JiangNan Universty;2.Jiangnan University

Fund Project:

Ministry of Education - Xinhua Group's "Cloud Number Integration" Fund Project (2017A13055)

  • 摘要
  • |
  • 图/表
  • |
  • 访问统计
  • |
  • 参考文献
  • |
  • 相似文献
  • |
  • 引证文献
  • |
  • 资源附件
  • |
  • 文章评论
    摘要:

    为解决不均衡多分类问题,提出一种特征选择和AdaBoost的集成方法。首先,数据进行预处理。利用WSPSO算法进行特征选择,根据特征重要性选取初始粒子构建初始种群,使得算法初期就可以沿着正确的搜索方向开展,减少不相关特征的影响。其次,利用AdaBoost算法对于样本权重较敏感的特点,增强对小类样本的关注度。并且利用AUCarea作为评价标准,相对于其他评价标准,AUCarea具有可视化的优点且对较差AUC更加敏感。最后,与其他几种不均衡分类算法在不平衡数据集上进行对比,结果证明该算法可有效的处理不均衡多分类问题。

    Abstract:

    Abstract: In order to solve the problem of unbalanced multi-classification, a feature selection and AdaBoost integration method are proposed. First, the data is preprocessed. The WSPSO algorithm is used to select features, and the initial population is constructed according to the importance of the feature. The initial algorithm can be carried out along the correct search direction to reduce the influence of incoherent features. Secondly, the AdaBoost algorithm is more sensitive to sample weights, and the attention to small samples is enhanced. And using AUCare as the evaluation standard, Compared to other evaluation criteria, AUCare has the advantage of visualization and is more sensitive to poor AUC. Finally, compared with several other unbalanced classification algorithms on the unbalanced data set, The result proves that the algorithm can effectively deal with the unbalanced multi-classification problem

    参考文献
    相似文献
    引证文献
引用本文
分享
文章指标
  • 点击次数:
  • 下载次数:
  • HTML阅读次数:
  • 引用次数:
历史
  • 收稿日期:2019-05-31
  • 最后修改日期:2019-12-06
  • 录用日期:2019-12-10
  • 在线发布日期:
  • 出版日期: