Graph diffusion models have made significant progress in learning structured graph data and have demonstrated strong potential for predictive tasks. Existing approaches typically embed node, edge, and graph-level features into a unified latent space, modeling prediction tasks including classification and regression as a form of conditional generation. However, due to the non-Euclidean nature of graph data, features of different curvatures are entangled in the same latent space without releasing their geometric potential. To address this issue, we aim to construt an ideal Riemannian diffusion model to capture distinct manifold signatures of complex graph data and learn their distribution. This goal faces two challenges: numerical instability caused by exponential mapping during the encoding proces and manifold deviation during diffusion generation. To address these challenges, we propose GeoMancer: a novel Riemannian graph diffusion framework for both generation and prediction tasks. To mitigate numerical instability, we replace exponential mapping with an isometric-invariant Riemannian gyrokernel approach and decouple multi-level features onto their respective task-specific manifolds to learn optimal representations. To address manifold deviation, we introduce a manifold-constrained diffusion method and a self-guided strategy for unconditional generation, ensuring that the generated data remains aligned with the manifold signature. Extensive experiments validate the effectiveness of our approach, demonstrating superior performance across a variety of tasks.
翻译:图扩散模型在学习结构化图数据方面取得了显著进展,并在预测任务中展现出强大潜力。现有方法通常将节点、边和图级特征嵌入到统一的潜在空间中,将分类和回归等预测任务建模为条件生成的一种形式。然而,由于图数据的非欧几里得特性,不同曲率的特征在同一个潜在空间中纠缠,未能释放其几何潜力。为解决这一问题,我们旨在构建一个理想的黎曼扩散模型,以捕捉复杂图数据的多样流形特征并学习其分布。这一目标面临两个挑战:编码过程中指数映射引起的数值不稳定性,以及扩散生成过程中的流形偏差。为应对这些挑战,我们提出了GeoMancer:一种新颖的黎曼图扩散框架,适用于生成和预测任务。为缓解数值不稳定性,我们采用等距不变的黎曼陀螺核方法替代指数映射,并将多级特征解耦到各自任务特定的流形上以学习最优表示。为解决流形偏差,我们引入了流形约束扩散方法和用于无条件生成的自引导策略,确保生成的数据与流形特征保持一致。大量实验验证了我们方法的有效性,在多种任务中展现出优越性能。