An analysis of multi-objective evolutionary algorithms for training ensemble models based on different performance measures in software effort estimation

Leandro L. MINKU, Xin YAO

Research output: Book Chapters | Papers in Conference ProceedingsConference paper (refereed)Researchpeer-review

24 Citations (Scopus)

Abstract

Background: Previous work showed that Multi-objective Evolutionary Algorithms (MOEAs) can be used for training ensembles of learning machines for Software Effort Estima-tion (SEE) by optimising different performance measures concurrently. Optimisation based on three measures (LSD, MMRE and PRED(25)) was analysed and led to promising results in terms of performance on these and other measures. Aims: (a) It is not known how well ensembles trained on other measures would behave for SEE, and whether train-ing on certain measures would improve performance partic-ularly on these measures. (b) It is also not known whether it is best to include all SEE models created by the MOEA into the ensemble, or solely the models with the best train-ing performance in terms of each measure being optimised. Investigating (a) and (b) is the aim of this work. Method: MOEAs were used to train ensembles by optimis-ing four different sets of performance measures, involving a total of nine different measures. The performance of all ensembles was then compared based on all these nine per-formance measures. Ensembles composed of different sets of models generated by the MOEAs were also compared. Results: (a) Ensembles trained on LSD, MMRE and PRED (25) obtained the best results in terms of most performance measures, being considered more successful than the others. Optimising certain performance measures did not necessar-ily lead to the best test performance on these particular mea-sures probably due to overfitting. (b) There was no inherent advantage in using ensembles composed of all the SEE mod-els generated by the MOEA in comparison to using solely the best SEE model according to each measure separately. Conclusions: Care must be taken to prevent overfitting on the performance measures being optimised. Our results sug-gest that concurrently optimising LSD, MMRE and PRED (25) promoted more ensemble diversity than other combina-tions of measures, and hence performed best. Low diversity is more likely to lead to overfitting.
Original languageEnglish
Title of host publicationPROMISE '13: Proceedings of the 9th International Conference on Predictive Models in Software Engineering
PublisherAssociation for Computing Machinery
Number of pages10
ISBN (Print)9781450320160
DOIs
Publication statusPublished - 9 Oct 2013
Externally publishedYes

Keywords

  • Ensembles of learning machines
  • Multi-objective evolutionary algorithms
  • Software effort estimation

Fingerprint

Dive into the research topics of 'An analysis of multi-objective evolutionary algorithms for training ensemble models based on different performance measures in software effort estimation'. Together they form a unique fingerprint.

Cite this