Treffer: Meta-Learning from Learning Curves for Budget-Limited Algorithm Selection

Title:
Meta-Learning from Learning Curves for Budget-Limited Algorithm Selection
Contributors:
Chalearn, Laboratoire Interdisciplinaire des Sciences du Numérique (LISN), Institut National de Recherche en Informatique et en Automatique (Inria)-CentraleSupélec-Université Paris-Saclay-Centre National de la Recherche Scientifique (CNRS), TAckling the Underspecified (TAU), Institut National de Recherche en Informatique et en Automatique (Inria)-CentraleSupélec-Université Paris-Saclay-Centre National de la Recherche Scientifique (CNRS)-Institut National de Recherche en Informatique et en Automatique (Inria)-CentraleSupélec-Université Paris-Saclay-Centre National de la Recherche Scientifique (CNRS)-Centre Inria de l'Université Paris-Saclay, Centre Inria de Saclay, Institut National de Recherche en Informatique et en Automatique (Inria)-Institut National de Recherche en Informatique et en Automatique (Inria)-Centre Inria de Saclay, Institut National de Recherche en Informatique et en Automatique (Inria), ANR-19-CHIA-0022,HUMANIA,Intelligence Artificielle pour Tous(2019), European Project: 952215,H2020-ICT-2018-20,H2020-ICT-2019-3,TAILOR(2020)
Source:
Pattern Recognition Letters. 185:225-231
Publisher Information:
CCSD; Elsevier, 2024.
Publication Year:
2024
Collection:
collection:CNRS
collection:INRIA
collection:INRIA-SACLAY
collection:INRIA_TEST
collection:TESTALAIN1
collection:CENTRALESUPELEC
collection:INRIA2
collection:TDS-MACS
collection:UNIV-PARIS-SACLAY
collection:UNIVERSITE-PARIS-SACLAY
collection:ANR
collection:LISN
collection:GS-COMPUTER-SCIENCE
collection:LISN-AO
collection:INRIA-ETATSUNIS
collection:PSACLAY-TEST
collection:ANR-IA-19
collection:ANR-IA
Original Identifier:
ARXIV: 2410.07696
HAL: hal-04719035
Document Type:
Zeitschrift article<br />Journal articles
Language:
English
ISSN:
0167-8655
Relation:
info:eu-repo/semantics/altIdentifier/arxiv/2410.07696; info:eu-repo/semantics/altIdentifier/doi/10.1016/j.patrec.2024.08.010; info:eu-repo/grantAgreement//952215/EU/Foundations of Trustworthy AI - Integrating Reasoning, Learning and Optimization/TAILOR
DOI:
10.1016/j.patrec.2024.08.010
Rights:
info:eu-repo/semantics/OpenAccess
URL: http://creativecommons.org/licenses/by-nc/
Accession Number:
edshal.hal.04719035v1
Database:
HAL

Weitere Informationen

Training a large set of machine learning algorithms to convergence in order to select the best-performing algorithm for a dataset is computationally wasteful. Moreover, in a budget-limited scenario, it is crucial to carefully select an algorithm candidate and allocate a budget for training it, ensuring that the limited budget is optimally distributed to favor the most promising candidates. Casting this problem as a Markov Decision Process, we propose a novel framework in which an agent must select in the process of learning the most promising algorithm without waiting until it is fully trained. At each time step, given an observation of partial learning curves of algorithms, the agent must decide whether to allocate resources to further train the most promising algorithm (exploitation), to wake up another algorithm previously put to sleep, or to start training a new algorithm (exploration). In addition, our framework allows the agent to meta-learn from learning curves on past datasets along with dataset meta-features and algorithm hyperparameters. By incorporating meta-learning, we aim to avoid myopic decisions based solely on premature learning curves on the dataset at hand. We introduce two benchmarks of learning curves that served in international competitions at WCCI’22 and AutoML-conf’22, of which we analyze the results. Our findings show that both meta- learning and the progression of learning curves enhance the algorithm selection process, as evidenced by methods of winning teams and our DDQN baseline, compared to heuristic baselines or a random search. Interestingly, our cost-effective baseline, which selects the best-performing algorithm w.r.t. a small budget, can perform decently when learning curves do not intersect frequently.