Traditionally the machine learning community has viewed the No Free Lunch (NFL) theorems for search and optimization as a limitation. I review, analyze, and unify the NFL theorem with the many frameworks to arrive at necessary conditions for improving black-box optimization, model selection, and machine learning in general. I review meta-learning literature to determine when and how meta-learning can benefit machine learning. We generalize meta-learning, in context of the NFL theorems, to arrive at a novel technique called Anti-Training with Sacrificial Data (ATSD). My technique applies at the meta level to arrive at domain specific algorithms and models. I also show how to generate sacrificial data. An extensive case study is presented along with simulated annealing results to demonstrate the efficacy of the ATSD method.
Identifer | oai:union.ndltd.org:arizona.edu/oai:arizona.openrepository.com:10150/605111 |
Date | January 2016 |
Creators | Valenzuela, Michael Lawrence |
Contributors | Rozenbilt, Jerzy W., Head, Kenneth L., Rozenbilt, Jerzy W., Head, Kenneth L., Lysecky, Roman L., Marcellin, Michael W. |
Publisher | The University of Arizona. |
Source Sets | University of Arizona |
Language | en_US |
Detected Language | English |
Type | text, Electronic Dissertation |
Rights | Copyright © is held by the author. Digital access to this material is made possible by the University Libraries, University of Arizona. Further transmission, reproduction or presentation (such as public display or performance) of protected items is prohibited except with permission of the author. |
Page generated in 0.002 seconds