Decoupled progressive distillation for sequential prediction with interaction dynamics

Kaixi HU, Lin LI, Qing XIE, Jianquan LIU, Xiaohui TAO, Guandong XU

Research output: Contribution to journalArticlespeer-review

1 Citation (Scopus)

Abstract

Sequential prediction has great value for resource allocation due to its capability in analyzing intents for next prediction. A fundamental challenge arises from real-world interaction dynamics where similar sequences involving multiple intents may exhibit different next items. More importantly, the character of volume candidate items in sequential prediction may amplify such dynamics, making deep networks hard to capture comprehensive intents. This article presents a sequential prediction framework with Decoupled Progressive Distillation (DePoD), drawing on the progressive nature of human cognition. We redefine target and non-target item distillation according to their different effects in the decoupled formulation. This can be achieved through two aspects: (1) Regarding how to learn, our target item distillation with progressive difficulty increases the contribution of low-confidence samples in the later training phase while keeping high-confidence samples in the earlier phase. And, the non-target item distillation starts from a small subset of non-target items from which size increases according to the item frequency. (2) Regarding whom to learn from, a difference evaluator is utilized to progressively select an expert that provides informative knowledge among items from the cohort of peers. Extensive experiments on four public datasets show DePoD outperforms state-of-the-art methods in terms of accuracy-based metrics. Copyright © 2023 held by the owner/author(s).

Original languageEnglish
Article number72
JournalACM Transactions on Information Systems
Volume42
Issue number3
DOIs
Publication statusPublished - Dec 2023

Citation

Hu, K., Li, L., Xie, Q., Liu, J., Tao, X., & Xu, G. (2023). Decoupled progressive distillation for sequential prediction with interaction dynamics. ACMTransactions on Information Systems, 42(3), Article 72. https://doi.org/10.1145/3632403

Keywords

  • Sequential prediction
  • Representation learning
  • Interaction dynamics
  • Knowledge distillation

Fingerprint

Dive into the research topics of 'Decoupled progressive distillation for sequential prediction with interaction dynamics'. Together they form a unique fingerprint.