Online Adversarial Knowledge Distillation for Graph Neural Networks

Abstract

Knowledge distillation, a technique recently gaining popularity for enhancing model generalization in Convolutional Neural Networks (CNNs), operates under the assumption that both teacher and student models are trained on identical data distributions. However, its effect on Graph Neural Networks (GNNs) is less than satisfactory since the graph topology and node attributes are prone to evolve, thereby leading to the issue of distribution shift. In this paper, we tackle this challenge by simultaneously training a group of graph neural networks in an online distillation fashion, where the group knowledge plays a role as a dynamic virtual teacher and the structure changes in graph neural networks are effectively captured. To improve the distillation performance, two types of knowledge are transferred among the students to enhance each other, local knowledge reflecting information in the graph topology and node attributes, and global knowledge reflecting the prediction over classes. We transfer the global knowledge with KL-divergence as the vanilla knowledge distillation does, while exploiting the complicated structure of the local knowledge with an efficient adversarial cyclic learning framework. Extensive experiments verified the effectiveness of our proposed online adversarial distillation approach. The code is published at https://github.com/wangz3066/OnlineDistillGCN.

Publication
In Expert Systems with Applications
Can Wang
Can Wang
王灿 教授
Zhe Wang
Zhe Wang
Student

I am a fourth-year Ph.D. student, and my supervisors are Prof. Chun Chen and Prof. Can Wang.

Defang Chen
Defang Chen
陈德仿 博士后
Yan Feng
Yan Feng
冯雁 副教授
Chun Chen
Chun Chen
陈纯 院士