Home > Papers

 
 
Session-based Recommendation with Self-Distillation Graph Neural Networks
Yuming Wang,Siyang Zhang *
School of Information and Telecommunication Engineering, Beijing University of Posts and Telecommunications ,Beijing 100876;School of Information and Telecommunication Engineering, Beijing University of Posts and Telecommunications ,Beijing 100876
*Correspondence author
#Submitted by
Subject:
Funding: none
Opened online:22 March 2022
Accepted by: none
Citation: Yuming Wang,Siyang Zhang.Session-based Recommendation with Self-Distillation Graph Neural Networks[OL]. [22 March 2022] http://en.paper.edu.cn/en_releasepaper/content/4756481
 
 
Recommendation systems have become fundamental in e-commerce scenarios, and session-based recommendation plays an increasingly significant role in recommendation systems because of its flexibility and highly practical value. Although there have been some promising results in previous works, they are still insufficient to achieve superior recommendation performance due to the limited even noisy information involved in the next click in each session. To obtain more accurate predictive vectors without the misleading of potential noisy information, we propose Self-Distillation Graph Neural Networks to make full use of the valuable information in a session, which is termed as SD-GNN for brevity. Specifically, we employ the well-evaluated and flexible deep ensemble in deep learning as the teacher model, which assembles multiple randomly initialized GNNs in a simple way. Furthermore, we leverage the soft target distribution produced by the teacher model to train each GNN in the ensemble to achieve self-knowledge distillation. Our whole method is easily implementable and scalable due to the proposed Self-Distillation technique. Extensive experiments on two benchmark datasets verify that the proposed method (SD-GNN) significantly outperforms state-of-the-art baselines and shows powerful performance in the session-based recommendation.
Keywords:Graph Neural Networks.; Session-based Recommendation ; Knowledge Distillation
 
 
 

For this paper

  • PDF (0B)
  • ● Revision 0   
  • ● Print this paper
  • ● Recommend this paper to a friend
  • ● Add to my favorite list

    Saved Papers

    Please enter a name for this paper to be shown in your personalized Saved Papers list

Tags

Add yours

Related Papers

Statistics

PDF Downloaded 11
Bookmarked 0
Recommend 0
Comments Array
Submit your papers