Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Similarity-aware Positive Instance Sampling for Graph Contrastive Pre-training (2206.11959v1)

Published 23 Jun 2022 in cs.LG and cs.AI

Abstract: Graph instance contrastive learning has been proved as an effective task for Graph Neural Network (GNN) pre-training. However, one key issue may seriously impede the representative power in existing works: Positive instances created by current methods often miss crucial information of graphs or even yield illegal instances (such as non-chemically-aware graphs in molecular generation). To remedy this issue, we propose to select positive graph instances directly from existing graphs in the training set, which ultimately maintains the legality and similarity to the target graphs. Our selection is based on certain domain-specific pair-wise similarity measurements as well as sampling from a hierarchical graph encoding similarity relations among graphs. Besides, we develop an adaptive node-level pre-training method to dynamically mask nodes to distribute them evenly in the graph. We conduct extensive experiments on $13$ graph classification and node classification benchmark datasets from various domains. The results demonstrate that the GNN models pre-trained by our strategies can outperform those trained-from-scratch models as well as the variants obtained by existing methods.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Xueyi Liu (14 papers)
  2. Yu Rong (146 papers)
  3. Tingyang Xu (55 papers)
  4. Fuchun Sun (127 papers)
  5. Wenbing Huang (95 papers)
  6. Junzhou Huang (137 papers)

Summary

We haven't generated a summary for this paper yet.