Publication Type

Conference Proceeding Article

Version

publishedVersion

Publication Date

9-2007

Abstract

The idea of using a team of weak learners to learn a dataset is a successful one in literature. In this paper, we explore a recursive incremental approach to ensemble learning. In this paper, patterns are clustered according to the output space of the problem, i.e., natural clusters are formed based on patterns belonging to each class. A combinatorial optimization problem is therefore formed, which is solved using evolutionary algorithms. The evolutionary algorithms identify the "easy" and the "difficult" clusters in the system. The removal of the easy patterns then gives way to the focused learning of the more complicated patterns. Incrementally, neural networks are added to the ensemble to focus on solving successively difficult examples. The problem therefore becomes recursively simpler. Over fitting is overcome by using a set of validation patterns along with a pattern distributor. An algorithm is also proposed to use the pattern distributor to determine the optimal number of recursions and hence the optimal number of weak learners for the problem. In this paper, we show that the generalization accuracy of the proposed algorithm is always better than that of the underlying weak learner. Empirical studies show generally good performance when compared to other state-of- the-art methods.

Discipline

Databases and Information Systems

Research Areas

Data Science and Engineering

Publication

Proceedings of the 2007 IEEE Congress on Evolutionary Computation, Singapore, September 25-28

First Page

1168

Last Page

1174

ISBN

9781424413393

Identifier

10.1109/CEC.2007.4424602

Publisher

IEEE

City or Country

Singapore

Additional URL

http://doi.org/10.1109/CEC.2007.4424602

Share

COinS