Selective negative correlation learning algorithm for incremental learning

Minlong LIN, Ke TANG, Xin YAO

Research output: Book Chapters | Papers in Conference ProceedingsConference paper (refereed)Researchpeer-review

5 Citations (Scopus)


Negative correlation learning (NCL) is a successful scheme for constructing neural network ensembles. In batch learning mode, NCL outperforms many other ensemble learning approaches. Recently, NCL is also shown to be a potentially powerful approach to incremental learning, while the advantage of NCL has not yet been fully exploited. In this paper, we propose a selective NCL approach for incremental learning. In the proposed approach, the previously trained ensemble is cloned when a new data set presents and the cloned ensemble is trained on the new data set. Then, the new ensemble is combined with the previous ensemble and a selection process is applied to prune the whole ensemble to a fixedsize. Simulation results on several benchmark datasets show that the proposed algorithm outperforms two recent incremental learning algorithms based on NCL. © 2008 IEEE.
Original languageEnglish
Title of host publicationProceedings of the International Joint Conference on Neural Networks
Number of pages6
Publication statusPublished - Jun 2008
Externally publishedYes


Dive into the research topics of 'Selective negative correlation learning algorithm for incremental learning'. Together they form a unique fingerprint.

Cite this