Show simple item record

dc.contributor.authorJorgensen, Steven
dc.contributor.authorNadizar, Giorgia
dc.contributor.authorPietropolli, Gloria
dc.contributor.authorManzoni, Luca
dc.contributor.authorMedvet, Eric
dc.contributor.authorO'Reilly, Una-May
dc.contributor.authorHemberg, Erik
dc.date.accessioned2025-12-04T23:07:23Z
dc.date.available2025-12-04T23:07:23Z
dc.date.issued2025-10-31
dc.identifier.issn2688-3007
dc.identifier.urihttps://hdl.handle.net/1721.1/164207
dc.description.abstractCurriculum learning (CL) consists in using a diverse set of user-provided test cases, with varying levels of difficulty and organized in a suitable progression, for learning a policy. The quality of test cases is important to allow optimization techniques as genetic programming (GP) to solve policy search problems. In this work, we evaluate large language models (LLMs) as providers of test cases for GP-based policy search. We consider two policy search tasks, a single-player and a multi-player game, and four LLMs differing in complexity and specialization, which we prompt in order to generate suitable test cases for the two games. We experimentally assess the intrinsic quality of LLM-generated test cases and their utility when inserted in a curriculum consumed by a GP optimization. We evaluate the robustness of the approach with respect to the way cases are scheduled in curricula and with respect to the policy representation, for which we use both graphs and linear programs evolved by GP. We observe that the effectiveness of LLM-assisted CL depends on both the choice of LLM and the design of the prompting and scheduling strategies. These findings highlight important considerations for leveraging LLMs in automated curriculum design for GP-based optimization.en_US
dc.publisherACMen_US
dc.relation.isversionofhttp://dx.doi.org/10.1145/3772718en_US
dc.rightsArticle is made available in accordance with the publisher's policy and may be subject to US copyright law. Please refer to the publisher's site for terms of use.en_US
dc.sourceAssociation for Computing Machineryen_US
dc.titlePolicy Search through Genetic Programming and LLM-assisted Curriculum Learningen_US
dc.typeArticleen_US
dc.identifier.citationSteven Jorgensen, Giorgia Nadizar, Gloria Pietropolli, Luca Manzoni, Eric Medvet, Una-May O'Reilly, and Erik Hemberg. 2025. Policy Search through Genetic Programming and LLM-assisted Curriculum Learning. ACM Trans. Evol. Learn. Optim.en_US
dc.contributor.departmentMassachusetts Institute of Technology. Computer Science and Artificial Intelligence Laboratoryen_US
dc.relation.journalACM Transactions on Evolutionary Learning and Optimizationen_US
dc.identifier.mitlicensePUBLISHER_POLICY
dc.eprint.versionFinal published versionen_US
dc.type.urihttp://purl.org/eprint/type/JournalArticleen_US
eprint.statushttp://purl.org/eprint/status/PeerRevieweden_US
dc.date.updated2025-11-01T07:58:52Z
dc.language.rfc3066en
dc.rights.holderThe author(s)
dspace.date.submission2025-11-01T07:58:53Z
mit.licensePUBLISHER_POLICY
mit.metadata.statusAuthority Work and Publication Information Neededen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record