Author:
Shi Shuhao,Qiao Kai,Yang Shuai,Wang Linyuan,Chen Jian,Yan Bin
Abstract
The graph neural network (GNN) has been widely used for graph data representation. However, the existing researches only consider the ideal balanced dataset, and the imbalanced dataset is rarely considered. Traditional methods such as resampling, reweighting, and synthetic samples that deal with imbalanced datasets are no longer applicable in GNN. This study proposes an ensemble model called Boosting-GNN, which uses GNNs as the base classifiers during boosting. In Boosting-GNN, higher weights are set for the training samples that are not correctly classified by the previous classifiers, thus achieving higher classification accuracy and better reliability. Besides, transfer learning is used to reduce computational cost and increase fitting ability. Experimental results indicate that the proposed Boosting-GNN model achieves better performance than graph convolutional network (GCN), GraphSAGE, graph attention network (GAT), simplifying graph convolutional networks (SGC), multi-scale graph convolution networks (N-GCN), and most advanced reweighting and resampling methods on synthetic imbalanced datasets, with an average performance improvement of 4.5%.
Subject
Artificial Intelligence,Biomedical Engineering
Reference52 articles.
1. “N-gcn: Multi-scale graph convolution for semi-supervised node classification,”;Abu-El-Haija;UAI,2019
2. Adaptive graph convolutional recurrent network for traffic forecasting;Bai;ArXiv, abs/2007.02842,2020
3. Neighbourhood sampling in bagging for imbalanced data;Blaszczynski;Neurocomputing,2015
4. A systematic study of the class imbalance problem in convolutional neural networks;Buda;Neural Netw,2018
5. “What is the effect of importance weighting in deep learning?”;Byrd;ICML,2019
Cited by
16 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献