Search papers, labs, and topics across Lattice.
This paper introduces GraphGlue, a framework for multi-domain graph pre-training that integrates knowledge by merging graph datasets into a unified Riemannian manifold. The core contribution is the theoretical establishment of neural manifold gluing, which uses adaptive orthogonal frames to characterize local geometry and then "glues" these local pieces together. Experiments across diverse graph domains demonstrate GraphGlue's superior performance and validate a geometric scaling law, showing that larger datasets improve model transferability by producing a smoother manifold.
GraphGlue offers a theoretically grounded method for multi-domain graph pre-training, showing how to merge diverse graph datasets into a unified Riemannian manifold for enhanced knowledge transfer.
Multi-domain graph pre-training integrates knowledge from diverse domains to enhance performance in the target domains, which is crucial for building graph foundation models. Despite initial success, existing solutions often fall short of answering a fundamental question: how is knowledge integrated or transferred across domains? This theoretical limitation motivates us to rethink the consistency and transferability between model pre-training and domain adaptation. In this paper, we propose a fresh Riemannian geometry perspective, whose core idea is to merge any graph dataset into a unified, smooth Riemannian manifold, enabling a systematic understanding of knowledge integration and transfer. To achieve this, our key contribution is the theoretical establishment of neural manifold gluing, which first characterizes local geometry using an adaptive orthogonal frame and then"glues"the local pieces together into a coherent whole. Building on this theory, we present the GraphGlue framework, which supports batched pre-training with EMA prototyping and provides a transferability measure based on geometric consistence. Extensive experiments demonstrate its superior performance across diverse graph domains. Moreover, we empirically validated GraphGlue's geometric scaling law, showing that larger quantities of datasets improve model transferability by producing a smoother manifold. Codes are available at https://github.com/RiemannGraph/GraphGlue.