File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: ForesTexter: An efficient random forest algorithm for imbalanced text categorization

TitleForesTexter: An efficient random forest algorithm for imbalanced text categorization
Authors
KeywordsImbalanced classification
Text categorization
SVM
Stratified sampling
Random forests
Issue Date2014
Citation
Knowledge-Based Systems, 2014, v. 67, p. 105-116 How to Cite?
AbstractIn this paper, we propose a new random forest (RF) based ensemble method, ForesTexter, to solve the imbalanced text categorization problems. RF has shown great success in many real-world applications. However, the problem of learning from text data with class imbalance is a relatively new challenge that needs to be addressed. A RF algorithm tends to use a simple random sampling of features in building their decision trees. As a result, it selects many subspaces that contain few, if any, informative features for the minority class. Furthermore, the Gini measure for data splitting is considered to be skew sensitive and bias towards the majority class. Due to the inherent complex characteristics of imbalanced text datasets, learning RF from such data requires new approaches to overcome challenges related to feature subspace selection and cut-point choice while performing node splitting. To this end, we propose a new tree induction method that selects splits, both feature subspace selection and splitting criterion, for RF on imbalanced text data. The key idea is to stratify features into two groups and to generate effective term weighting for the features. One group contains positive features for the minority class and the other one contains the negative features for the majority class. Then, for feature subspace selection, we effectively select features from each group based on the term weights. The advantage of our approach is that each subspace contains adequate informative features for both minority and majority classes. One difference between our proposed tree induction method and the classical RF method is that our method uses Support Vector Machines (SVM) classifier to split the training data into smaller and more balance subsets at each tree node, and then successively retrains the SVM classifiers on the data partitions to refine the model while moving down the tree. In this way, we force the classifiers to learn from refined feature subspaces and data subsets to fit the imbalanced data better. Hence, the tree model becomes more robust for text categorization task with imbalanced dataset. Experimental results on various benchmark imbalanced text datasets (Reuters-21578, Ohsumed, and imbalanced 20 newsgroup) consistently demonstrate the effectiveness of our proposed ForesTexter method. The performance of our proposed approach is competitive against the standard random forest and different variants of SVM algorithms. © 2014 Elsevier B.V. All rights reserved.
Persistent Identifierhttp://hdl.handle.net/10722/276999
ISSN
2021 Impact Factor: 8.139
2020 SCImago Journal Rankings: 1.587
ISI Accession Number ID

 

DC FieldValueLanguage
dc.contributor.authorWu, Qingyao-
dc.contributor.authorYe, Yunming-
dc.contributor.authorZhang, Haijun-
dc.contributor.authorNg, Michael K.-
dc.contributor.authorHo, Shen Shyang-
dc.date.accessioned2019-09-18T08:35:18Z-
dc.date.available2019-09-18T08:35:18Z-
dc.date.issued2014-
dc.identifier.citationKnowledge-Based Systems, 2014, v. 67, p. 105-116-
dc.identifier.issn0950-7051-
dc.identifier.urihttp://hdl.handle.net/10722/276999-
dc.description.abstractIn this paper, we propose a new random forest (RF) based ensemble method, ForesTexter, to solve the imbalanced text categorization problems. RF has shown great success in many real-world applications. However, the problem of learning from text data with class imbalance is a relatively new challenge that needs to be addressed. A RF algorithm tends to use a simple random sampling of features in building their decision trees. As a result, it selects many subspaces that contain few, if any, informative features for the minority class. Furthermore, the Gini measure for data splitting is considered to be skew sensitive and bias towards the majority class. Due to the inherent complex characteristics of imbalanced text datasets, learning RF from such data requires new approaches to overcome challenges related to feature subspace selection and cut-point choice while performing node splitting. To this end, we propose a new tree induction method that selects splits, both feature subspace selection and splitting criterion, for RF on imbalanced text data. The key idea is to stratify features into two groups and to generate effective term weighting for the features. One group contains positive features for the minority class and the other one contains the negative features for the majority class. Then, for feature subspace selection, we effectively select features from each group based on the term weights. The advantage of our approach is that each subspace contains adequate informative features for both minority and majority classes. One difference between our proposed tree induction method and the classical RF method is that our method uses Support Vector Machines (SVM) classifier to split the training data into smaller and more balance subsets at each tree node, and then successively retrains the SVM classifiers on the data partitions to refine the model while moving down the tree. In this way, we force the classifiers to learn from refined feature subspaces and data subsets to fit the imbalanced data better. Hence, the tree model becomes more robust for text categorization task with imbalanced dataset. Experimental results on various benchmark imbalanced text datasets (Reuters-21578, Ohsumed, and imbalanced 20 newsgroup) consistently demonstrate the effectiveness of our proposed ForesTexter method. The performance of our proposed approach is competitive against the standard random forest and different variants of SVM algorithms. © 2014 Elsevier B.V. All rights reserved.-
dc.languageeng-
dc.relation.ispartofKnowledge-Based Systems-
dc.subjectImbalanced classification-
dc.subjectText categorization-
dc.subjectSVM-
dc.subjectStratified sampling-
dc.subjectRandom forests-
dc.titleForesTexter: An efficient random forest algorithm for imbalanced text categorization-
dc.typeArticle-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1016/j.knosys.2014.06.004-
dc.identifier.scopuseid_2-s2.0-84905560384-
dc.identifier.volume67-
dc.identifier.spage105-
dc.identifier.epage116-
dc.identifier.isiWOS:000340221600009-
dc.identifier.issnl0950-7051-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats