TY - GEN
T1 - Mutually beneficial learning with application to on-line news classification
AU - Wu, Lei
AU - Li, Zhiwei
AU - Li, Mingjing
AU - Ma, Wei-Ying
AU - Yu, Nenghai
N1 - Publication details (e.g. title, author(s), publication statuses and dates) are captured on an “AS IS” and “AS AVAILABLE” basis at the time of record harvesting from the data source. Suggestions for further amendments or supplementary information can be sent to [email protected].
PY - 2007
Y1 - 2007
N2 - There are three common challenges in real-world classification applications, i.e. how to use domain knowledge, how to resist noisy samples and how to use unlabeled data. To address these problems, a novel classification framework called Mutually Beneficial Learning (MBL) is proposed in this paper. MBL integrates two learning steps together. In the first step, the underlying local structures of feature space are discovered through a learning process. The result provides necessary capability to resist noisy samples and prepare better input for the second step where a consecutive classification process is further applied to the result. These two steps are iteratively performed until a stop condition is met. Different from traditional classifiers, the output of MBL consists of two components: a common classifier and a set of rules corresponding to local structures. In application, a test sample is first matched with the discovered rules. If a matched rule is found, the label of the rule is assigned to the sample; otherwise, the common classifier will be utilized to classify the sample. We applied the MBL to online news classification, and our experimental results showed that MBL is significantly better than Naïve Bayes and SVM, even when the data is noisy or partially labeled. © 2007 ACM.
AB - There are three common challenges in real-world classification applications, i.e. how to use domain knowledge, how to resist noisy samples and how to use unlabeled data. To address these problems, a novel classification framework called Mutually Beneficial Learning (MBL) is proposed in this paper. MBL integrates two learning steps together. In the first step, the underlying local structures of feature space are discovered through a learning process. The result provides necessary capability to resist noisy samples and prepare better input for the second step where a consecutive classification process is further applied to the result. These two steps are iteratively performed until a stop condition is met. Different from traditional classifiers, the output of MBL consists of two components: a common classifier and a set of rules corresponding to local structures. In application, a test sample is first matched with the discovered rules. If a matched rule is found, the label of the rule is assigned to the sample; otherwise, the common classifier will be utilized to classify the sample. We applied the MBL to online news classification, and our experimental results showed that MBL is significantly better than Naïve Bayes and SVM, even when the data is noisy or partially labeled. © 2007 ACM.
KW - implicit domain knowledge
KW - local structure
KW - mutually beneficial learning
KW - news classification
UR - http://www.scopus.com/inward/record.url?scp=79959640795&partnerID=8YFLogxK
UR - https://www.scopus.com/record/pubmetrics.uri?eid=2-s2.0-79959640795&origin=recordpage
U2 - 10.1145/1316874.1316889
DO - 10.1145/1316874.1316889
M3 - RGC 32 - Refereed conference paper (with host publication)
SN - 9781595938329
T3 - International Conference on Information and Knowledge Management, Proceedings
SP - 85
EP - 92
BT - CIKM 2007 Co-Located Workshops - Proceedings of PIKM 2007
PB - Association for Computing Machinery
T2 - 1st Ph.D. Workshop, PIKM 2007 - Co-Located with CIKM 2007
Y2 - 6 November 2007 through 9 November 2007
ER -