Learning ELM-Tree from big data based on uncertainty reduction

Ran Wang, Yu-Lin He*, Chi-Yin Chow, Fang-Fang Ou, Jian Zhang

*Corresponding author for this work

Research output: Journal Publications and ReviewsRGC 21 - Publication in refereed journalpeer-review

51 Citations (Scopus)

Abstract

A challenge in big data classification is the design of highly parallelized learning algorithms. One solution to this problem is applying parallel computation to different components of a learning model. In this paper, we first propose an extreme learning machine tree (ELM-Tree) model based on the heuristics of uncertainty reduction. In the ELM-Tree model, information entropy and ambiguity are used as the uncertainty measures for splitting decision tree (DT) nodes. Besides, in order to resolve the over-partitioning problem in the DT induction, ELMs are embedded as the leaf nodes when the gain ratios of all the available splits are smaller than a given threshold. Then, we apply parallel computation to five components of the ELM-Tree model, which effectively reduces the computational time for big data classification. Experimental studies demonstrate the effectiveness of the proposed method.
Original languageEnglish
Pages (from-to)79-100
JournalFuzzy Sets and Systems
Volume258
Online published14 May 2014
DOIs
Publication statusPublished - 1 Jan 2015

Research Keywords

  • Big data classification
  • Decision tree
  • ELM-Tree
  • Extreme learning machine
  • Uncertainty reduction

Fingerprint

Dive into the research topics of 'Learning ELM-Tree from big data based on uncertainty reduction'. Together they form a unique fingerprint.

Cite this