Recent years have witnessed the emerging success of graph neural networks
(GNNs) for modeling structured data. However, most GNNs are designed for
homogeneous graphs, in which all nodes and edges belong to the same types,
making them infeasible to represent heterogeneous structures. In this paper, we
present the Heterogeneous Graph Transformer (HGT) architecture for modeling
Web-scale heterogeneous graphs. To model heterogeneity, we design node- and
edge-type dependent parameters to characterize the heterogeneous attention over
each edge, empowering HGT to maintain dedicated representations for different
types of nodes and edges. To handle dynamic heterogeneous graphs, we introduce
the relative temporal encoding technique into HGT, which is able to capture the
dynamic structural dependency with arbitrary durations. To handle Web-scale
graph data, we design the heterogeneous mini-batch graph sampling
algorithm---HGSampling---for efficient and scalable training. Extensive
experiments on the Open Academic Graph of 179 million nodes and 2 billion edges
show that the proposed HGT model consistently outperforms all the
state-of-the-art GNN baselines by 9%--21% on various downstream tasks.
%0 Generic
%1 hu2020heterogeneous
%A Hu, Ziniu
%A Dong, Yuxiao
%A Wang, Kuansan
%A Sun, Yizhou
%D 2020
%K #dl #embedding #ml
%T Heterogeneous Graph Transformer
%U http://arxiv.org/abs/2003.01332
%X Recent years have witnessed the emerging success of graph neural networks
(GNNs) for modeling structured data. However, most GNNs are designed for
homogeneous graphs, in which all nodes and edges belong to the same types,
making them infeasible to represent heterogeneous structures. In this paper, we
present the Heterogeneous Graph Transformer (HGT) architecture for modeling
Web-scale heterogeneous graphs. To model heterogeneity, we design node- and
edge-type dependent parameters to characterize the heterogeneous attention over
each edge, empowering HGT to maintain dedicated representations for different
types of nodes and edges. To handle dynamic heterogeneous graphs, we introduce
the relative temporal encoding technique into HGT, which is able to capture the
dynamic structural dependency with arbitrary durations. To handle Web-scale
graph data, we design the heterogeneous mini-batch graph sampling
algorithm---HGSampling---for efficient and scalable training. Extensive
experiments on the Open Academic Graph of 179 million nodes and 2 billion edges
show that the proposed HGT model consistently outperforms all the
state-of-the-art GNN baselines by 9%--21% on various downstream tasks.
@misc{hu2020heterogeneous,
abstract = {Recent years have witnessed the emerging success of graph neural networks
(GNNs) for modeling structured data. However, most GNNs are designed for
homogeneous graphs, in which all nodes and edges belong to the same types,
making them infeasible to represent heterogeneous structures. In this paper, we
present the Heterogeneous Graph Transformer (HGT) architecture for modeling
Web-scale heterogeneous graphs. To model heterogeneity, we design node- and
edge-type dependent parameters to characterize the heterogeneous attention over
each edge, empowering HGT to maintain dedicated representations for different
types of nodes and edges. To handle dynamic heterogeneous graphs, we introduce
the relative temporal encoding technique into HGT, which is able to capture the
dynamic structural dependency with arbitrary durations. To handle Web-scale
graph data, we design the heterogeneous mini-batch graph sampling
algorithm---HGSampling---for efficient and scalable training. Extensive
experiments on the Open Academic Graph of 179 million nodes and 2 billion edges
show that the proposed HGT model consistently outperforms all the
state-of-the-art GNN baselines by 9%--21% on various downstream tasks.},
added-at = {2020-08-27T23:30:32.000+0200},
author = {Hu, Ziniu and Dong, Yuxiao and Wang, Kuansan and Sun, Yizhou},
biburl = {https://www.bibsonomy.org/bibtex/20c4c17852fb0490cdd9d4299bac7e04e/yihong-liu},
description = {Heterogeneous Graph Transformer},
interhash = {a898c0dc2305e22504e4213ec94b3040},
intrahash = {0c4c17852fb0490cdd9d4299bac7e04e},
keywords = {#dl #embedding #ml},
note = {cite arxiv:2003.01332Comment: Published on WWW 2020},
timestamp = {2020-08-27T23:30:32.000+0200},
title = {Heterogeneous Graph Transformer},
url = {http://arxiv.org/abs/2003.01332},
year = 2020
}