Hi-GMAE: Hierarchical Graph Masked Autoencoders (2405.10642v1)
Abstract: Graph Masked Autoencoders (GMAEs) have emerged as a notable self-supervised learning approach for graph-structured data. Existing GMAE models primarily focus on reconstructing node-level information, categorizing them as single-scale GMAEs. This methodology, while effective in certain contexts, tends to overlook the complex hierarchical structures inherent in many real-world graphs. For instance, molecular graphs exhibit a clear hierarchical organization in the form of the atoms-functional groups-molecules structure. Hence, the inability of single-scale GMAE models to incorporate these hierarchical relationships often leads to their inadequate capture of crucial high-level graph information, resulting in a noticeable decline in performance. To address this limitation, we propose Hierarchical Graph Masked AutoEncoders (Hi-GMAE), a novel multi-scale GMAE framework designed to handle the hierarchical structures within graphs. First, Hi-GMAE constructs a multi-scale graph hierarchy through graph pooling, enabling the exploration of graph structures across different granularity levels. To ensure masking uniformity of subgraphs across these scales, we propose a novel coarse-to-fine strategy that initiates masking at the coarsest scale and progressively back-projects the mask to the finer scales. Furthermore, we integrate a gradual recovery strategy with the masking process to mitigate the learning challenges posed by completely masked subgraphs. Diverging from the standard graph neural network (GNN) used in GMAE models, Hi-GMAE modifies its encoder and decoder into hierarchical structures. This entails using GNN at the finer scales for detailed local graph analysis and employing a graph transformer at coarser scales to capture global information. Our experiments on 15 graph datasets consistently demonstrate that Hi-GMAE outperforms 17 state-of-the-art self-supervised competitors.
- Language models are few-shot learners. In NeurIPS.
- When to Pre-Train Graph Neural Networks? From Data Generation Perspective!. In SIGKDD.
- Chih-Chung Chang and Chih-Jen Lin. 2011. LIBSVM: a library for support vector machines. TIST (2011).
- BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In ACL.
- Graph Neural Networks with Learnable Structural and Positional Representations. In ICLR.
- Matthias Fey and Jan E. Lenssen. 2019. Fast Graph Representation Learning with PyTorch Geometric. In ICLR Workshop.
- Hongyang Gao and Shuiwang Ji. 2019. Graph u-nets. In ICML.
- Hierarchical graph learning for protein–protein interaction. Nature Communications (2023).
- Semi-Implicit Graph Variational Auto-Encoders. In NeurIPS.
- GraphMAE2: A Decoding-Enhanced Masked Self-Supervised Graph Learner. In WWW.
- GraphMAE: Self-Supervised Masked Graph Autoencoders. In SIGKDD.
- Strategies for Pre-training Graph Neural Networks. In ICLR.
- HGCVAE: Integrating Generative and Contrastive Learning for Heterogeneous Graph Learning. arXiv:2310.11102 (2023).
- Contrastive pre-training of GNNs on heterogeneous graphs. In CIKM. 803–812.
- George Karypis and Vipin Kumar. 1997. METIS: A software package for partitioning unstructured graphs, partitioning meshes, and computing fill-reducing orderings of sparse matrices. (1997).
- Diederik P Kingma and Jimmy Ba. 2014. Adam: A method for stochastic optimization. arXiv:1412.6980 (2014).
- Thomas N Kipf and Max Welling. 2016. Variational graph auto-encoders. arXiv:1611.07308 (2016).
- Self-attention graph pooling. In ICML.
- KPGT: Knowledge-Guided Pre-training of Graph Transformer for Molecular Property Prediction. In SIGKDD. 857–867.
- What’s Behind the Mask: Understanding Masked Graph Modeling for Graph Autoencoders. In SIGKDD.
- Let invariant rationale discovery inspire graph contrastive learning. In ICLR.
- Where to Mask: Structure-Guided Masking for Graph Masked Autoencoders. In IJCAI.
- Gapformer: Graph Transformer with Graph Pooling for Node Classification. In IJCAI.
- Graph Pooling for Graph Neural Networks: Progress, Challenges, and Opportunities. In IJCAI.
- Mask and Reason: Pre-Training Knowledge Graph Transformers for Complex Logical Queries. In SIGKDD.
- Rethinking Tokenizer and Decoder in Masked Graph Modeling for Molecules. NeurIPS (2023).
- Tudataset: A collection of benchmark datasets for learning with graphs. arXiv:2007.08663 (2020).
- Adversarially Regularized Graph Autoencoder for Graph Embedding. In IJCAI.
- GCC: Graph Contrastive Coding for Graph Neural Network Pre-Training. In SIGKDD.
- Recipe for a General, Powerful, Scalable Graph Transformer. In NeurIPS.
- GiGaMAE: Generalizable Graph Masked Autoencoder via Collaborative Latent Space Reconstruction. In CIKM.
- Teague Sterling and John J Irwin. 2015. ZINC 15–ligand discovery for everyone. Journal of chemical information and modeling (2015).
- InfoGraph: Unsupervised and Semi-supervised Graph-Level Representation Learning via Mutual Information Maximization. In ICLR.
- GPPT: Graph Pre-training and Prompt Tuning to Generalize Graph Neural Networks. In SIGKDD.
- S2GAE: Self-Supervised Graph Autoencoders Are Generalizable Learners with Graph Masking. In WSDM.
- Heterogeneous graph masked autoencoders. In AAAI.
- RARE: Robust Masked Graph Autoencoder. IEEE TKDE (2023).
- Deep Graph Infomax. In ICLR.
- Generative and Contrastive Paradigms Are Complementary for Graph Self-Supervised Learning. arXiv:2310.15523 (2023).
- MoleculeNet: a benchmark for molecular machine learning. Chem. Sci. (2018).
- SimGRACE: A Simple Framework for Graph Contrastive Learning without Data Augmentation. In WWW.
- Mole-BERT: Rethinking Pre-training Graph Neural Networks for Molecules. In ICLR.
- A survey of pretraining on graphs: Taxonomy, methods, and applications. arXiv:2202.07893 (2022).
- InfoGCL: Information-Aware Graph Contrastive Learning. In NeurIPS.
- How Powerful are Graph Neural Networks?. In ICLR.
- Self-supervised graph-level representation learning with local and global structure. In ICML.
- Skeletonmae: graph-based masked autoencoder for skeleton sequence pre-training. In ICCV.
- Do Transformers Really Perform Badly for Graph Representation?. In NeurIPS.
- Hierarchical Graph Representation Learning with Differentiable Pooling. In NeurIPS.
- Graph contrastive learning automated. In ICML.
- Graph Contrastive Learning with Augmentations. In NeurIPS.
- ProteinMAE: Masked Autoencoder for Protein Surface Self-supervised Learning. Bioinformatics (2023).
- Hierarchical molecular graph self-supervised learning for property prediction. Communications Chemistry (2023).
- Contrastive Cross-scale Graph Knowledge Synergy. In SIGKDD.
- COSTA: Covariance-Preserving Feature Augmentation for Graph Contrastive Learning. In SIGKDD.
- Hierarchical Graph Transformer with Adaptive Node Sampling. In NeurIPS.
- Masked Graph Autoencoder with Non-discrete Bandwidths. In WWW.
- Unified 2d and 3d pre-training of molecular representations. In SIGKDD.
- Hierarchical Transformer for Scalable Graph Learning. In IJCAI.