Skip to main navigation Skip to search Skip to main content

CLDYB: TOWARDS DYNAMIC BENCHMARKING FOR CONTINUAL LEARNING WITH PRE-TRAINED MODELS

Shengzhuang Chen, Yikai Liao, Xiaoxiao Sun, Kede Ma*, Ying Wei*

*Corresponding author for this work

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review

Abstract

The advent of the foundation model era has sparked significant research interest in leveraging pre-trained representations for continual learning (CL), yielding a series of top-performing CL methods on standard evaluation benchmarks. Nonetheless, there are growing concerns regarding potential data contamination during the pre-training stage. Furthermore, standard evaluation benchmarks, which are typically static, fail to capture the complexities of real-world CL scenarios, resulting in saturated performance. To address these issues, we describe CL on dynamic benchmarks (CLDyB), a general computational framework based on Markov decision processes for evaluating CL methods reliably. CLDyB dynamically identifies inherently difficult and algorithm-dependent tasks for the given CL methods, and determines challenging task orders using Monte Carlo tree search. Leveraging CLDyB, we first conduct a joint evaluation of multiple state-of-the-art CL methods, leading to a set of commonly challenging and generalizable task sequences where existing CL methods tend to perform poorly. We then conduct separate evaluations of individual CL methods using CLDyB, discovering their respective strengths and weaknesses. The source code and generated task sequences are publicly accessible at https://github.com/szc12153/CLDyB. © 2025 13th International Conference on Learning Representations, ICLR 2025. All rights reserved.
Original languageEnglish
Title of host publication13th International Conference on Learning Representations (ICLR 2025)
PublisherInternational Conference on Learning Representations, ICLR
Pages74731-74763
ISBN (Electronic)9798331320850
Publication statusPublished - 2025
Event13th International Conference on Learning Representations (ICLR 2025) - Singapore EXPO, Singapore, Singapore
Duration: 24 Apr 202528 Apr 2025
https://iclr.cc/Conferences/2025

Publication series

NameInternational Conference on Learning Representations, ICLR

Conference

Conference13th International Conference on Learning Representations (ICLR 2025)
Abbreviated titleICLR 2025
PlaceSingapore
CitySingapore
Period24/04/2528/04/25
Internet address

Funding

We would like to thank Xuelin Liu for assistance with the plots and diagrams. This work was supported in part by the Hong Kong RGC General Research Fund (11220224), the CityU Applied Research Grant (9667264), and an Industry Gift Fund (9229111).

RGC Funding Information

  • RGC-funded

Fingerprint

Dive into the research topics of 'CLDYB: TOWARDS DYNAMIC BENCHMARKING FOR CONTINUAL LEARNING WITH PRE-TRAINED MODELS'. Together they form a unique fingerprint.

Cite this