Get 20M+ Full-Text Papers For Less Than $1.50/day. Start a 14-Day Trial for You or Your Team.

Learn More →

FlexGP

FlexGP We describe FlexGP, the first Genetic Programming system to perform symbolic regression on large-scale datasets on the cloud via massive data-parallel ensemble learning. FlexGP provides a decentralized, fault tolerant parallelization framework that runs many copies of Multiple Regression Genetic Programming, a sophisticated symbolic regression algorithm, on the cloud. Each copy executes with a different sample of the data and different parameters. The framework can create a fused model or ensemble on demand as the individual GP learners are evolving. We demonstrate our framework by deploying 100 independent GP instances in a massive data-parallel manner to learn from a dataset composed of 515K exemplars and 90 features, and by generating a competitive fused model in less than 10 minutes. http://www.deepdyve.com/assets/images/DeepDyve-Logo-lg.png Journal of Grid Computing Springer Journals

Loading next page...
 
/lp/springer-journals/flexgp-SvO71XaCcP

References (21)

Publisher
Springer Journals
Copyright
Copyright © 2014 by Springer Science+Business Media Dordrecht
Subject
Computer Science; Processor Architectures; Management of Computing and Information Systems; User Interfaces and Human Computer Interaction
ISSN
1570-7873
eISSN
1572-9184
DOI
10.1007/s10723-014-9320-9
Publisher site
See Article on Publisher Site

Abstract

We describe FlexGP, the first Genetic Programming system to perform symbolic regression on large-scale datasets on the cloud via massive data-parallel ensemble learning. FlexGP provides a decentralized, fault tolerant parallelization framework that runs many copies of Multiple Regression Genetic Programming, a sophisticated symbolic regression algorithm, on the cloud. Each copy executes with a different sample of the data and different parameters. The framework can create a fused model or ensemble on demand as the individual GP learners are evolving. We demonstrate our framework by deploying 100 independent GP instances in a massive data-parallel manner to learn from a dataset composed of 515K exemplars and 90 features, and by generating a competitive fused model in less than 10 minutes.

Journal

Journal of Grid ComputingSpringer Journals

Published: Nov 18, 2014

There are no references for this article.