Design Principle Transfer in Neural Architecture Search via Large Language Models

Xun Zhou, Xingyu Wu*, Liang Feng*, Zhichao Lu, Kay Chen Tan

*Corresponding author for this work

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review

Abstract

Transferable neural architecture search (TNAS) has been introduced to design efficient neural architectures for multiple tasks, to enhance the practical applicability of NAS in real-world scenarios. In TNAS, architectural knowledge accumulated in previous search processes is reused to warm up the architecture search for new tasks. However, existing TNAS methods still search in an extensive search space, necessitating the evaluation of numerous architectures. To overcome this challenge, this work proposes a novel transfer paradigm, i.e., design principle transfer. In this work, the linguistic description of various structural components’ effects on architectural performance is termed design principles. They are learned from established architectures and then can be reused to reduce the search space by discarding unpromising architectures. Searching in the refined search space can boost both the search performance and efficiency for new NAS tasks. To this end, a large language model (LLM)-assisted design principle transfer (LAPT) framework is devised. In LAPT, LLM is applied to automatically reason the design principles from a set of given architectures, and then a principle adaptation method is applied to refine these principles progressively based on the new search results. Experimental results show that LAPT can beat the state-of-the-art TNAS methods on most tasks and achieve comparable performance on others. Copyright © 2025, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved.
Original languageEnglish
Title of host publicationProceedings of the 39th AAAI Conference on Artificial Intelligence
EditorsToby Walsh, Julie Shah, Zico Kolter
PublisherAAAI Press
Pages23000-23008
Volume39
ISBN (Print)1-57735-897-X, 978-1-57735-897-8
DOIs
Publication statusPublished - 2025
Event39th Annual AAAI Conference on Artificial Intelligence (AAAI 2025) - Pennsylvania Convention Center , Philadelphia, United States
Duration: 25 Feb 20254 Mar 2025
https://aaai.org/conference/aaai/aaai-25/

Publication series

NameProceedings of the AAAI Conference on Artificial Intelligence
PublisherAssociation for the Advancement of Artificial Intelligence
ISSN (Print)2159-5399

Conference

Conference39th Annual AAAI Conference on Artificial Intelligence (AAAI 2025)
Abbreviated titleAAAI-25
Country/TerritoryUnited States
CityPhiladelphia
Period25/02/254/03/25
Internet address

Funding

This work was supported in part by National Key R&D Program of China (2022YFC3801700); in part by the Research Grants Council of the Hong Kong SAR (Grant No. PolyU11211521, PolyU15218622, PolyU15215623, and C5052-23G), and the National Natural Science Foundation of China (Grant No. U21A20512).

Fingerprint

Dive into the research topics of 'Design Principle Transfer in Neural Architecture Search via Large Language Models'. Together they form a unique fingerprint.

Cite this