Abstract
Sequential prediction has great value for resource allocation due to its capability in analyzing intents for next prediction. A fundamental challenge arises from real-world interaction dynamics where similar sequences involving multiple intents may exhibit different next items. More importantly, the character of volume candidate items in sequential prediction may amplify such dynamics, making deep networks hard to capture comprehensive intents. This article presents a sequential prediction framework with Decoupled Progressive Distillation (DePoD), drawing on the progressive nature of human cognition. We redefine target and non-target item distillation according to their different effects in the decoupled formulation. This can be achieved through two aspects: (1) Regarding how to learn, our target item distillation with progressive difficulty increases the contribution of low-confidence samples in the later training phase while keeping high-confidence samples in the earlier phase. And, the non-target item distillation starts from a small subset of non-target items from which size increases according to the item frequency. (2) Regarding whom to learn from, a difference evaluator is utilized to progressively select an expert that provides informative knowledge among items from the cohort of peers. Extensive experiments on four public datasets show DePoD outperforms state-of-the-art methods in terms of accuracy-based metrics. Copyright © 2023 held by the owner/author(s).
Original language | English |
---|---|
Article number | 72 |
Journal | ACM Transactions on Information Systems |
Volume | 42 |
Issue number | 3 |
DOIs | |
Publication status | Published - Dec 2023 |
Citation
Hu, K., Li, L., Xie, Q., Liu, J., Tao, X., & Xu, G. (2023). Decoupled progressive distillation for sequential prediction with interaction dynamics. ACMTransactions on Information Systems, 42(3), Article 72. https://doi.org/10.1145/3632403Keywords
- Sequential prediction
- Representation learning
- Interaction dynamics
- Knowledge distillation