Multi-Domain Riemannian Graph Gluing for Building Graph Foundation Models
Abstract
Multi-domain graph pre-training leverages Riemannian geometry to integrate knowledge across domains by merging graph datasets into a unified manifold, enabling systematic understanding of knowledge transfer through neural manifold gluing theory.
Multi-domain graph pre-training integrates knowledge from diverse domains to enhance performance in the target domains, which is crucial for building graph foundation models. Despite initial success, existing solutions often fall short of answering a fundamental question: how is knowledge integrated or transferred across domains? This theoretical limitation motivates us to rethink the consistency and transferability between model pre-training and domain adaptation. In this paper, we propose a fresh Riemannian geometry perspective, whose core idea is to merge any graph dataset into a unified, smooth Riemannian manifold, enabling a systematic understanding of knowledge integration and transfer. To achieve this, our key contribution is the theoretical establishment of neural manifold gluing, which first characterizes local geometry using an adaptive orthogonal frame and then "glues" the local pieces together into a coherent whole. Building on this theory, we present the GraphGlue framework, which supports batched pre-training with EMA prototyping and provides a transferability measure based on geometric consistence. Extensive experiments demonstrate its superior performance across diverse graph domains. Moreover, we empirically validated GraphGlue's geometric scaling law, showing that larger quantities of datasets improve model transferability by producing a smoother manifold. Codes are available at https://github.com/RiemannGraph/GraphGlue.
Community
Multi-domain graph pre-training integrates knowledge from diverse domains to enhance performance in the target domains, which is crucial for building graph foundation models. Despite initial success, existing solutions often fall short of answering a fundamental question: how is knowledge integrated or transferred across domains? This theoretical limitation motivates us to rethink the consistency and transferability between model pre-training and domain adaptation. In this paper, we propose a fresh Riemannian geometry perspective, whose core idea is to merge any graph dataset into a unified, smooth Riemannian manifold, enabling a systematic understanding of knowledge integration and transfer.
The key contributions are listed as follows.
- Problem. We investigate the theoretical underpinnings of multi-domain graph pre-training and study a foundational problem of how knowledge is integrated and transferred across different domains.
- Theory. We introduce a fresh differential-geometric perspective for systematically understanding knowledge transfer and propose the theory of neural manifold gluing, which consistently integrates multi-domain graphs into a unified, smooth Riemannian manifold via “gluing”.
- Methodology. We propose a GRAPHGLUE framework based on the above theory that supports batched pre-training for large-scale graphs and incorporates a natural metric to quantify transferability.
- Experiment. We evaluate GRAPHGLUEin cross-domain transfer learning and empirically demonstrate its geometric scaling law.
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper