Delayed Bottlenecking: Alleviating Forgetting in Pre-trained Graph Neural Networks

Zhe Zhao, Pengkun Wang*, Xu Wang, Haibin Wen, Xiaolong Xie, Zhengyang Zhou, Qingfu Zhang, Yang Wang*

*Corresponding author for this work

Research output: Journal Publications and ReviewsRGC 21 - Publication in refereed journalpeer-review

1 Citation (Scopus)

Abstract

Pre-training GNNs to extract transferable knowledge and apply it to downstream tasks has become the de facto standard of graph representation learning. Recent works focused on designing self-supervised pre-training tasks to extract useful and universal transferable knowledge from large-scale unlabeled data. However, they have to face an inevitable question: traditional pre-training strategies that aim at extracting useful information about pre-training tasks, may not extract all useful information about the downstream task. In this paper, we reexamine the pre-training process within traditional pre-training and fine-tuning frameworks from the perspective of Information Bottleneck (IB) and confirm that the forgetting phenomenon in pre-training phase may cause detrimental effects on downstream tasks. Therefore, we propose a novel Delayed Bottlenecking Pre-training (DBP) framework which maintains as much as possible mutual information between latent representations and training data during pre-training phase by suppressing the compression operation and delays the compression operation to fine-tuning phase to make sure the compression can be guided with labeled fine-tuning data and downstream tasks. To achieve this, we design two information control objectives that can be directly optimized and further integrate them into the actual model design. Extensive experiments on both chemistry and biology domains demonstrate the effectiveness of DBP. © 2024 IEEE.
Original languageEnglish
Pages (from-to)1140-1153
JournalIEEE Transactions on Knowledge and Data Engineering
Volume37
Issue number3
Online published12 Dec 2024
DOIs
Publication statusPublished - Mar 2025

Funding

This work was supported in part by the Natural Science Foundation of China Youth Project under Grant 62402472, in part by the Natural Science Foundation of Jiangsu Province of China Youth Project under Grant BK20240461, and Grant BK20240460, in part by the Research Grants Council of the Hong Kong Special Administrative Region, China, under Grant CityU 11215723, in part by the National Natural Science Foundation of China under Grant 62072427, and Grant 12227901, in part by the Project of Stable Support for Youth Team in Basic Research Field, CAS, under Grant YSBR-005, in part by the Academic Leaders Cultivation Program, USTC, and in part by the the Key Basic Research Foundation of Shenzhen, China under Grant JCYJ20220818100005011.

Research Keywords

  • Forget
  • Graph neural networks
  • Information bottleneck
  • Pre-training

Fingerprint

Dive into the research topics of 'Delayed Bottlenecking: Alleviating Forgetting in Pre-trained Graph Neural Networks'. Together they form a unique fingerprint.

Cite this