Abstract

Recent years have witnessed the emerging success of graph neural networks (GNNs) for modeling structured data. However, most GNNs are designed for homogeneous graphs, in which all nodes and edges belong to the same types, making it infeasible to represent heterogeneous structures. In this paper, we present the Heterogeneous Graph Transformer (HGT) architecture for modeling Web-scale heterogeneous graphs. To model heterogeneity, we design node- and edge-type dependent parameters to characterize the heterogeneous attention over each edge, empowering HGT to maintain dedicated representations for different types of nodes and edges. To handle Web-scale graph data, we design the heterogeneous mini-batch graph sampling algorithm—HGSampling—for efficient and scalable training. Extensive experiments on the Open Academic Graph of 179 million nodes and 2 billion edges show that the proposed HGT model consistently outperforms all the state-of-the-art GNN baselines by 9–21 on various downstream tasks. The dataset and source code of HGT are publicly available at https://github.com/acbull/pyHGT.

Keywords

Computer scienceScalabilityTheoretical computer scienceGraphHomogeneousTransformerDistributed computingData miningDatabaseMathematics

Affiliated Institutions

Related Publications

Publication Info

Year
2020
Type
article
Pages
2704-2710
Citations
1160
Access
Closed

External Links

Social Impact

Altmetric

Social media, news, blog, policy document mentions

Citation Metrics

1160
OpenAlex

Cite This

Ziniu Hu, Yuxiao Dong, Kuansan Wang et al. (2020). Heterogeneous Graph Transformer. , 2704-2710. https://doi.org/10.1145/3366423.3380027

Identifiers

DOI
10.1145/3366423.3380027