Abstract
Heterogeneous Graph Neural Networks(HGNNs), as an effective tool for mining heterogeneous graphs, have achieved remarkable performance on series of real-world applications. Yet, HGNNs are limited in their mining power as they require all nodes to have complete and reliable attributes. It is usually unrealistic since the attributes of many nodes in reality are inevitably missing or noisy. Existing methods usually take imputation schemes to complete missing attributes, in which topology information is ignored, leading to suboptimal performance. In this work, we study the consistency-based augmentation on heterogeneous graphs, completing the missing attributes and improving original attributes simultaneously, and propose a novel generic architecture−Learning on Heterogeneous Graph Neural Networks with Consistency-based Augmentation(CAHGNN), including random sampling, attribute augmentation and consistency training. In graph augmentation, to ensure attributes sensible and accurate, the attention mechanism is adopted to complete attributes under the guidance of the topological relationship between nodes. Extensive experiments on three benchmark datasets demonstrate the superior performance of CAHGNN over state-of-the-art baselines on semi-supervised node classification.








Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Data Availability
The authors confirm that the data supporting the findings of this study are available within the article.
References
Yang C, Xiao Y, Zhang Y, Sun Y, Han J (2022) Heterogeneous network representation learning: A unified framework with survey and benchmark. IEEE Trans Knowl Data Eng 34(10):4854–4873
Shi C, Li Y, Zhang J, Sun Y, Philip SY (2016) A survey of heterogeneous information network analysis. IEEE Trans Knowl Data Eng 29(1):17–37
Shen W, Han J, Wang J, Yuan X, Yang Z (2017) Shine+: A general framework for domain-specific entity linking with heterogeneous information networks. IEEE Trans Knowl Data Eng 30(2):353–366
Wang X, Ji H, Shi C, Wang B, Ye Y, Cui P, Yu PS (2019) Heterogeneous graph attention network. In: The world wide web conference, pp 2022–2032
Yun S, Jeong M, Kim R, Kang J, Kim HJ (2019) Graph transformer networks. Advances in neural information processing systems 32
Fu X, Zhang J, Meng Z, King I (2020) Magnn: Metapath aggregated graph neural network for heterogeneous graph embedding. Proceedings of the web conference 2020:2331–2341
Zhang C, Song D, Huang C, Swami A, Chawla NV (2019) Heterogeneous graph neural network. In: Proceedings of the 25th ACM SIGKDD International conference on knowledge discovery & data mining, pp 793–803
Fan S, Zhu J, Han X, Shi C, Hu L, Ma B, Li Y (2019) Metapath-guided heterogeneous graph neural network for intent recommendation. In: Proceedings of the 25th ACM SIGKDD International conference on knowledge discovery & data mining, pp 2478–2486
Li A, Qin Z, Liu R, Yang Y, Li D (2019) Spam review detection with graph convolutional networks. In: Proceedings of the 28th ACM International conference on information and knowledge management, pp 2703–2711
Yang T, Hu L, Shi C, Ji H, Li X, Nie L (2021) Hgat: Heterogeneous graph attention networks for semi-supervised short text classification. ACM Transactions on Information Systems (TOIS) 39(3):1–29
Sun Y, Han J (2013) Mining heterogeneous information networks: a structural analysis approach. Acm Sigkdd Explorations Newsletter 14(2):20–28
Jin D, Huo C, Liang C, Yang L (2021) Heterogeneous graph neural network via attribute completion. Proceedings of the web conference 2021:391–400
He D, Liang C, Huo C, Feng Z, Jin D, Yang L, Zhang W (2022) Analyzing heterogeneous networks with missing attributes by unsupervised contrastive learning. IEEE Transactions on Neural Networks and Learning Systems
Xie Q, Dai Z, Hovy E, Luong T, Le Q (2020) Unsupervised data augmentation for consistency training. Adv Neural Inf Process Syst 33:6256–6268
Dong Y, Chawla NV, Swami A (2017) metapath2vec: Scalable representation learning for heterogeneous networks. In: Proceedings of the 23rd ACM SIGKDD International conference on knowledge discovery and data mining, pp 135–144
Fu T-y, Lee W-C, Lei Z (2017) Hin2vec: Explore meta-paths in heterogeneous information networks for representation learning. In: Proceedings of the 2017 ACM on Conference on information and knowledge management, pp 1797–1806
Shang J, Qu M, Liu J, Kaplan LM, Han J, Peng J (2016) Meta-path guided embedding for similarity search in large-scale heterogeneous information networks. arXiv:1610.09769
Wang X, Zhang Y, Shi C (2019) Hyperbolic heterogeneous information network embedding. Proceedings of the AAAI Conference on artificial intelligence 33:5337–5344
Hu Z, Dong Y, Wang K, Sun Y (2020) Heterogeneous graph transformer. Proceedings of the web conference 2020:2704–2710
You J, Ma X, Ding Y, Kochenderfer MJ, Leskovec J (2020) Handling missing data with graph representation learning. Adv Neural Inf Process Syst 33:19075–19087
Chen X, Chen S, Yao J, Zheng H, Zhang Y, Tsang IW (2020) Learning on attribute-missing graphs. IEEE transactions on pattern analysis and machine intelligence
Wang Y, Wang W, Liang Y, Cai Y, Liu J, Hooi B (2020) Nodeaug: Semi-supervised node classification with data augmentation. In: Proceedings of the 26th ACM SIGKDD International conference on knowledge discovery & data mining, pp 207–217
Feng W, Zhang J, Dong Y, Han Y, Luan H, Xu Q, Yang Q, Kharlamov E, Tang J (2020) Graph random neural networks for semi-supervised learning on graphs. Adv Neural Inf Process Syst 33:22092–22103
Liu S, Ying R, Dong H, Li L, Xu T, Rong Y, Zhao P, Huang J, Wu D (2022) Local augmentation for graph neural networks. In: Proceedings of the 39th International conference on machine learning, vol 162, pp 14054–14072
Zhao T, Liu Y, Neves L, Woodford O, Jiang M, Shah N (2021) Data augmentation for graph neural networks. Proceedings of the AAAI Conference on artificial intelligence 35:11015–11023
Ding K, Xu Z, Tong H, Liu H (2022) Data augmentation for deep graph learning: A survey. In: Proceedings of the 28th ACM SIGKDD International conference on knowledge discovery & data mining, vol 24, pp 61–77
Xu Z, Du B, Tong H (2022) Graph sanitation with application to node classification. Proc ACM Web conference 2022:1136–1147
Taguchi H, Liu X, Murata T (2021) Graph convolutional networks for graphs containing missing features. Future Gener Comput Syst 117:155–168
Rossi E, Kenlay H, Gorinova MI, Chamberlain BP, Dong X, Bronstein MM (2022) On the unreasonable effectiveness of feature propagation in learning on graphs with missing node features. In: Proceedings of the 1st Learning on graphs conference, vol 198, pp 11–11116
Bahdanau D, Cho K, Bengio Y (2015) Neural machine translation by jointly learning to align and translate. In: International conference on learning representations
Lee J, Lee I, Kang J (2019) Self-attention graph pooling. In: International conference on machine learning, pp 3734–3743
Veličković P, Cucurull G, Casanova A, Romero A, Liò P, Bengio Y (2018) Graph attention networks. In: International conference on learning representations
McPherson M, Smith-Lovin L, Cook JM (2001) Birds of a feather: Homophily in social networks. Annual review of sociology, pp 415–444
Pei H, Wei B, Chang KC, Lei Y, Yang B (2020) Geom-gcn: Geometric graph convolutional networks. In: International conference on learning representations
Schlichtkrull M, Kipf TN, Bloem P, Berg Rvd, Titov I, Welling M (2018) Modeling relational data with graph convolutional networks. In: European semantic web conference, pp 593–607 . Springer
Huang X, Song Q, Li Y, Hu X (2019) Graph recurrent networks with attributed random walks. In: Proceedings of the 25th ACM SIGKDD International conference on knowledge discovery & data mining, pp 732–740
Mikolov T, Chen K, Corrado GS, Dean J (2013) Efficient estimation of word representations in vector space. In: International conference on learning representations
Berthelot D, Carlini N, Goodfellow I, Papernot N, Oliver A, Raffel CA (2019) Mixmatch: A holistic approach to semi-supervised learning. Advances in neural information processing systems 32
Goodfellow I, Bengio Y, Courville A (2016) Deep learning. MIT Press
Kipf TN, Welling M (2017) Semi-supervised classification with graph convolutional networks. In: International conference on learning representations
Kingma DP, Ba J (2015) Adam: A method for stochastic optimization. In: International conference on learning representations
Suykens JA (2001) Support vector machines: a nonlinear modelling and control perspective. Eur J Control 7(2–3):311–327
Van der Maaten L, Hinton G (2008) Visualizing data using t-sne. Journal of machine learning research 9(11)
Acknowledgements
This work is supported by the Fundamental Research Funds for the Central Universities, China under Grant 2021III030JC.
Funding
This work is supported by the Fundamental Research Funds for the Central Universities, China under Grant 2021III030JC.
Author information
Authors and Affiliations
Contributions
All authors contributed to the study conception and design. Material preparation, data collection and analysis were performed by Liang Yixuan. The first draft of the manuscript was written by Liang Yixuan, Wan Yuan commented on previous versions of the manuscript and critically revised the work. All authors read and approved the final manuscript.
Corresponding author
Ethics declarations
Conflicts of interest
The authors have no competing interests to declare that are relevant to the content of this article.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Liang, Y., Wan, Y. Learning on heterogeneous graph neural networks with consistency-based augmentation. Appl Intell 53, 27624–27636 (2023). https://doi.org/10.1007/s10489-023-04995-6
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-023-04995-6