• Complex
  • Title
  • Keyword
  • Abstract
  • Scholars
  • Journal
  • ISSN
  • Conference
搜索

Author:

He, Ming (He, Ming.) | An, Boyang (An, Boyang.) | Wang, Jiwen (Wang, Jiwen.) | Wen, Hao (Wen, Hao.)

Indexed by:

EI Scopus

Abstract:

Explanations can substantially enhance users’ trust and satisfaction with recommender systems. Counterfactual explanations have demonstrated remarkable effectiveness in enhancing the performance of explainable sequential recommendation. However, existing counterfactual explanation models for sequential recommendation ignore temporal dependencies in a user’s historical behavior sequence. Moreover, counterfactual histories must be as close as possible to the real history; otherwise, they will violate the user’s real behavioral preferences. In this paper, we propose Counterfactual Explanations with Temporal Dependencies (CETD), a counterfactual explanation model based on a Variational Autoencoder (VAE) for sequential recommendation that handles temporal dependencies. When generating counterfactual histories, CETD uses a Recurrent Neural Network (RNN) to capture both long-term preferences and short-term behavior in the user’s real behavioral history, which can enhance explainability. Meanwhile, CETD fits the distribution of reconstructed data in a latent space, and then uses the variance obtained from learning to make counterfactual sequences closer to the original sequence, which will reduce the proximity of counterfactual histories. Extensive experiments on two real-world datasets show that the proposed CETD consistently outperforms state-of-the-art methods. © 2023, The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

Keyword:

Recommender systems Recurrent neural networks Behavioral research

Author Community:

  • [ 1 ] [He, Ming]Beijing University of Technology, Beijing, China
  • [ 2 ] [An, Boyang]Beijing University of Technology, Beijing, China
  • [ 3 ] [Wang, Jiwen]Beijing University of Technology, Beijing, China
  • [ 4 ] [Wen, Hao]Beijing University of Technology, Beijing, China

Reprint Author's Address:

Email:

Show more details

Related Keywords:

Related Article:

Source :

ISSN: 0302-9743

Year: 2023

Volume: 14306 LNCS

Page: 533-543

Language: English

Cited Count:

WoS CC Cited Count: 0

SCOPUS Cited Count: 1

ESI Highly Cited Papers on the List: 0 Unfold All

WanFang Cited Count:

Chinese Cited Count:

30 Days PV: 13

Affiliated Colleges:

Online/Total:457/10601767
Address:BJUT Library(100 Pingleyuan,Chaoyang District,Beijing 100124, China Post Code:100124) Contact Us:010-67392185
Copyright:BJUT Library Technical Support:Beijing Aegean Software Co., Ltd.