Submitted:
19 May 2023
Posted:
22 May 2023
You are already at the latest version
Abstract

Keywords:
1. Introduction
- Q1.
- Why train the long and short interest models separately?
- Q2.
- How the integration module corrects for interest bias?
- Q3.
- What is the versatility of our approach?
- A plug-and-play user long and short interest fusion module is designed, which can effectively and quickly fuse long and short interests using the shared attention mechanism, thus improving the model accuracy;
- The sources of interest bias are analyzed experimentally, and an improved ternary contrast loss function is introduced to accelerate the convergence of the model by using the bias between features as the index of the loss function;
- The effectiveness and generality of our proposed method is demonstrated by combining and experimenting several different long and short interest models on several different domain datasets with data of different sequence lengths as input.
2. Related Work
2.1. Entanglement Training
2.2. Disentanglement Training
2.3. Feature Fusion
3. OUR APPROACH
3.1. LS Interest Modeling
3.2. Fusion Module

3.3. Dynamic Decouple Loss
4. Experiments
- RQ1:How our modules perform in practice?
- RQ2:What is the individual contribution of each component in our model?
- RQ3:Can our Module effectively handle the complexity of sequence data with varying lengths across different scenarios?
4.1. Datasets and Experimental Setup
| Datasets1 | Users | Items | Average click sequence |
Average click categories |
Average click time span |
|---|---|---|---|---|---|
| Amazon | 19240 | 63001 | 8.780 | 8.780 | 1.050 |
| Taobao | 104693 | 1592919 | 102.170 | 24.385 | 0.009 |
| Douabn | 52539 | 140502 | 6.321 | 2.726 | 0.418 |
| Yelp | 1542656 | 209393 | 3.956 | 3.025 | 0.324 |
4.2. Competitors
4.3. Overall Performance Comparison(RQ1)
- The overall performance of the short-term interest model is better. From the overall results, the short-term interest model is better overall than the long-term interest model because it can capture the actual by-sequence information of user interaction well. The results from DIN, PACA, and RCNN show that the length interest features are more informative, and if more effective methods can be used to fit the long-term interest features, the improvement of the models’ effectiveness is significant.
- The long-term interest model has the advantage of playing in two situations: a large variety of products and a long time span of user clicks. Although the short-term interest model is generally better, the short-term model does not necessarily outperform the long-term model for the two cases of large variety of items and long click time span. The RCNN and CASER models also use CNN networks, but CASER is slightly less effective than the RCNN model, but there is a large gap between them and FMLP, which indicates that the user’s pre-click data helps the model to capture long-term interest, but the short-term interest weight is generally larger than the long-term interest weight. The cold start problem is a difficult problem for both models, and the best results for the Yelp dataset are lower than the remaining three datasets, and there is a large gap between the long-term and short-term models, which verifies that fitting to the length of the sequence data and effective extraction of the sequence data are the keys to improve the performance.
- Joint modeling of long and short interests is a generally effective approach. Joint modeling of models somewhat alleviates the poor performance of independent models in cold starts, large span of user clicks, and many types of user clicks, but it is not always effective. NARM, LUTUTR, and SLIREC are trained by entangling long and short interests with each other, which somewhat increases the redundancy of the models , and the performance of SLIREC in ACC and F1 in which inadequate. In contrast, CLSR decouples the calculation of long and short interests, and the adaptive weighting fuses the long and short interest weights, which alleviates the training volume of the model, which makes up for the lack of realizations of SLIREC on Taobao.
- Contrast learning and feature enhancement can effectively improve model performance. Our model differs little from the best comparison model CLSR on the data-rich dataset Taobao, but improves AUC by almost 0.01 on Douabn, which has a much smaller variety of products, and for the cold-start case, our model leads CLSR in all metrics, and has into 1.3% improvement on the long time span Taobao dataset. These validate the effectiveness of contrast loss and feature enhancement.
| Dataset | Amazon | Taobao | Douban | Yelp | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Category | Model | ACC | AUC | F1 | ACC | AUC | F1 | ACC | AUC | F1 | ACC | AUC | F1 |
| Long-term | DIN | 0.7148 | 0.8095 | 0.7167 | 0.6895 | 0.7624 | 0.6941 | 0.8549 | 0.8974 | 0.8577 | 0.6566 | 0.7035 | 0.6589 |
| PACA | 0.7057 | 0.8154 | 0.7096 | 0.7033 | 0.7761 | 0.7097 | 0.8440 | 0.8838 | 0.8369 | 0.6610 | 0.7271 | 0.6683 | |
| NARM | 0.7364 | 0.8340 | 0.7310 | 0.7021 | 0.7723 | 0.7029 | 0.8699 | 0.9174 | 0.8787 | 0.6834 | 0.7319 | 0.6803 | |
| RCNN | 0.7698 | 0.8465 | 0.7668 | 0.7174 | 0.8084 | 0.7218 | 0.8740 | 0.9281 | 0.8817 | 0.7093 | 0.7504 | 0.7017 | |
| Short-term | CASER | 0.7665 | 0.8415 | 0.7633 | 0.7122 | 0.7096 | 0.7145 | 0.8710 | 0.9204 | 0.8763 | 0.7307 | 0.7809 | 0.7393 |
| GRU4REC | 0.7747 | 0.8574 | 0.7789 | 0.7189 | 0.8087 | 0.7187 | 0.8811 | 0.9168 | 0.8837 | 0.7399 | 0.7807 | 0.7271 | |
| DIEN | 0.7805 | 0.8636 | 0.7774 | 0.7296 | 0.8390 | 0.7279 | 0.8951 | 0.9286 | 0.8938 | 0.7524 | 0.8036 | 0.7513 | |
| FMLP | 0.7881 | 0.8716 | 0.7830 | 0.7374 | 0.8389 | 0.7448 | 0.8941 | 0.9378 | 0.8896 | 0.7580 | 0.8073 | 0.7515 | |
| LS-term | LUTUR | 0.7924 | 0.8786 | 0.7911 | 0.7561 | 0.8391 | 0.7719 | 0.9066 | 0.9454 | 0.9010 | 0.7630 | 0.8098 | 0.7641 |
| SLIREC | 0.8002 | 0.8773 | 0.7973 | 0.7543 | 0.8318 | 0.7693 | 0.9018 | 0.9463 | 0.9086 | 0.7719 | 0.8122 | 0.7783 | |
| CLSR | 0.8046 | 0.8857 | 0.8077 | 0.7607 | 0.8388 | 0.7691 | 0.9132 | 0.9481 | 0.9047 | 0.7818 | 0.8164 | 0.7730 | |
| CDF-LS | 0.8014 | 0.8824 | 0.8096 | 0.7724 | 0.8392 | 0.7710 | 0.9134 | 0.9567 | 0.9079 | 0.7907 | 0.8204 | 0.7943 | |

4.4. Results of Ablation Experiments(RQ2)
4.4.1. Contrast Loss
| DIN Model | Length | 10 | 20 | 40 | 50 | ||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| User histories | Thread | class | popu | cross | class | popu | cross | class | popu | cross | class | popu | cross |
| Sequences | <=5 | 31 | 0.016 | 0.027 | 67 | 0.010 | 0.027 | 145 | 0.014 | 0.064 | 189 | 0.018 | 0.025 |
| >=50 | 34 | 0.005 | 0.333 | 87 | 0.008 | 0.276 | 165 | 0.005 | 0.202 | 98 | 0.001 | 0.330 | |
| Categories | <=2 | 2 | 0.000 | 0.384 | 52 | 0.004 | 0.014 | 164 | 0.007 | 0.014 | 202 | 0.008 | 0.011 |
| >=20 | 67 | 0.008 | 0.176 | 41 | 0.016 | 0.208 | 144 | 0.010 | 0.096 | 89 | 0.010 | 0.173 | |
| Time span | <=1 | 41 | 0.010 | 0.015 | 47 | 0.032 | 0.040 | 201 | 0.003 | 0.026 | 141 | 0.009 | 0.028 |
| >=3300 | 39 | 0.005 | 0.034 | 64 | 0.012 | 0.120 | 193 | 0.005 | 0.082 | 133 | 0.012 | 0.085 | |
| Model | Contrast | Weights | Amazon | Yelp | ||
|---|---|---|---|---|---|---|
| AUC | F1 | AUC | F1 | |||
| LUTUR | ✔ | 0.8804+0.0018 | 0.8013+0.0102 | 0.8149+0.0051 | 0.7736+0.0095 | |
| ✔ | ✔ | 0.8820+0.0034 | 0.8049+0.0136 | 0.8157+0.0059 | 0.7746+0.0105 | |
| SLIREC | ✔ | 0.8853+0.0080 | 0.8051+0.0078 | 0.8166+0.0044 | 0.7804+0.0021 | |
| ✔ | ✔ | 0.8861+0.0088 | 0.8064+0.0091 | 0.8171+0.0049 | 0.7847+0.0063 | |
| CLSR | ✔ | ✔ | 0.8859+0.0002 | 0.8101+0.0024 | 0.8217+0.0053 | 0.7881+0.0151 |
4.4.2. Feature Fusion
| Model | Fusion | Amazon | Yelp | ||
|---|---|---|---|---|---|
| AUC | F1 | AUC | F1 | ||
| DIN+CASER | ✗ | 0.8640+0.0225 | 0.7811+0.0178 | 0.7988+0.0179 | 0.7439+0.0046 |
| ✔ | 0.8715+0.0300 | 0.7866+0.0233 | 0.8123+0.0314 | 0.7501+0.0118 | |
| NARM+DIEN | ✗ | 0.8695+0.0059 | 0.7854+0.0008 | 0.8101+0.0065 | 0.7613+0.0100 |
| ✔ | 0.8763+0.0127 | 0.7878+0.0104 | 0.8160+0.0124 | 0.7729+0.0216 | |
| PACA+GRU4REC | ✗ | 0.8641+0.0067 | 0.7811+0.0022 | 0.8010+0.0203 | 0.7684+0.0413 |
| ✔ | 0.8720+0.0146 | 0.7869+0.0080 | 0.8103+0.0296 | 0.7700+0.0429 | |
| RCNN+FMLP | ✗ | 0.8740+0.0024 | 0.7846+0.0016 | 0.8074+0.0001 | 0.7517+0.0002 |
| ✔ | 0.8809+0.0093 | 0.7903+0.0073 | 0.8166+0.0093 | 0.7624+0.0109 | |
4.5. Robustness Test Experimental Results(RQ3)
| Model | Sequences | Amazon | Yelp | |||||
|---|---|---|---|---|---|---|---|---|
| 20 | 30 | 40 | 50 | AUC | F1 | AUC | F1 | |
| DIN + CASER | ✔ | ✔ | 0.8427 | 0.7696 | 0.7909 | 0.7415 | ||
| ✔ | ✔ | 0.8715 | 0.7866 | 0.8123 | 0.7501 | |||
| ✔ | ✔ | 0.8753 | 0.7917 | 0.8149 | 0.7553 | |||
| ✔ | ✔ | 0.8631 | 0.7898 | 0.8077 | 0.7430 | |||
| ✔ | ✔ | 0.8649 | 0.7910 | 0.8083 | 0.7431 | |||
| ✔ | ✔ | 0.8548 | 0.7704 | 0.7905 | 0.7257 | |||
| FMLP + RCNN | ✔ | ✔ | 0.8701 | 0.7792 | 0.8075 | 0.7547 | ||
| ✔ | ✔ | 0.8809 | 0.7903 | 0.8166 | 0.7724 | |||
| ✔ | ✔ | 0.8847 | 0.7953 | 0.8189 | 0.7764 | |||
| ✔ | ✔ | 0.8703 | 0.7765 | 0.8086 | 0.7561 | |||
| ✔ | ✔ | 0.8761 | 0.7846 | 0.8137 | 0.7704 | |||
| ✔ | ✔ | 0.8602 | 0.7739 | 0.8060 | 0.7433 | |||
5. Conclusion
Author Contributions
Funding
References
- An, M.; Wu, F.; Wu, C.; Zhang, K.; Liu, Z.; Xie, X. Neural news recommendation with long-and short-term user representations. In Proceedings of the Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, 2019, pp. 336–345.
- Zhou, G.; Zhu, X.; Song, C.; Fan, Y.; Zhu, H.; Ma, X.; Yan, Y.; Jin, J.; Li, H.; Gai, K. Deep interest network for click-through rate prediction. In Proceedings of the Proceedings of the 24th ACM SIGKDD international conference on knowledge discovery & data mining, 2018, pp. 1059–1068.
- Zhou, G.; Mou, N.; Fan, Y.; Pi, Q.; Bian, W.; Zhou, C.; Zhu, X.; Gai, K. Deep interest evolution network for click-through rate prediction. In Proceedings of the Proceedings of the AAAI conference on artificial intelligence, 2019, Vol. 33, pp. 5941–5948.
- Covington, P.; Adams, J.; Sargin, E. Deep neural networks for youtube recommendations. In Proceedings of the Proceedings of the 10th ACM conference on recommender systems, 2016, pp. 191–198.
- He, X.; Liao, L.; Zhang, H.; Nie, L.; Hu, X.; Chua, T.S. Neural collaborative filtering. In Proceedings of the Proceedings of the 26th international conference on world wide web, 2017, pp. 173–182.
- Sarwar, B.; Karypis, G.; Konstan, J.; Riedl, J. Item-based collaborative filtering recommendation algorithms. In Proceedings of the Proceedings of the 10th international conference on World Wide Web, 2001, pp. 285–295.
- Xu, G.; Wu, Z.; Zhang, Y.; Cao, J. Social networking meets recommender systems: survey. International Journal of Social Network Mining 2015, 2, 64–100.
- Koren, Y.; Bell, R.; Volinsky, C. Matrix factorization techniques for recommender systems. Computer 2009, 42, 30–37.
- Song, Y.; Xin, R.; Chen, P.; Zhang, R.; Chen, J.; Zhao, Z. Identifying performance anomalies in fluctuating cloud environments: A robust correlative-GNN-based explainable approach. Future Generation Computer Systems 2023.
- Chen, P.; Liu, H.; Xin, R.; Carval, T.; Zhao, J.; Xia, Y.; Zhao, Z. Effectively detecting operational anomalies in large-scale IoT data infrastructures by using a gan-based predictive model. The Computer Journal 2022, 65, 2909–2925.
- Wu, L.; Wu, D.A.; Liu, M.; Wang, X.M.; Gong, H.G. Periodic intermittently connected-based data delivery in opportunistic netw. Ruanjian Xuebao/Journal of Software 2013, 24, 507–525.
- Liu, M.; Deng, J.; Yang, M.; Cheng, X.; Liu, N.; Liu, M.; Wang, X. Cost Ensemble with Gradient Selecting for GANs. In Proceedings of the Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, IJCAI-22; Raedt, L.D., Ed. International Joint Conferences on Artificial Intelligence Organization, 2022, pp. 1194–1200. Main Track. [CrossRef]
- Lu, H.; Cheng, X.; Xia, W.; Deng, P.; Liu, M.; Xie, T.; Wang, X.; Liu, M. CyclicShift: A Data Augmentation Method For Enriching Data Patterns. In Proceedings of the Proceedings of the 30th ACM International Conference on Multimedia, 2022, pp. 4921–4929.
- Xie, T.; Cheng, X.; Wang, X.; Liu, M.; Deng, J.; Zhou, T.; Liu, M. Cut-thumbnail: A novel data augmentation for convolutional neural network. In Proceedings of the Proceedings of the 29th ACM International Conference on Multimedia, 2021, pp. 1627–1635.
- Liu, M.; Deng, J.; Yang, M.; Cheng, X.; Liu, N.; Liu, M.; Wang, X. Cost Ensemble with Gradient Selecting for GANs. In Proceedings of the Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, IJCAI 2022, Vienna, Austria, 23-29 July 2022; Raedt, L.D., Ed. ijcai.org, 2022, pp. 1194–1200. [CrossRef]
- Li, N.; Liu, Y.; Wu, Y.; Liu, S.; Zhao, S.; Liu, M. Robutrans: A robust transformer-based text-to-speech model. In Proceedings of the Proceedings of the AAAI Conference on Artificial Intelligence, 2020, Vol. 34, pp. 8228–8235.
- Lv, F.; Jin, T.; Yu, C.; Sun, F.; Lin, Q.; Yang, K.; Ng, W. SDM: Sequential deep matching model for online large-scale recommender system. In Proceedings of the Proceedings of the 28th ACM International Conference on Information and Knowledge Management, 2019, pp. 2635–2643.
- Li, J.; Ren, P.; Chen, Z.; Ren, Z.; Lian, T.; Ma, J. Neural attentive session-based recommendation. In Proceedings of the Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, 2017, pp. 1419–1428.
- Hidasi, B.; Karatzoglou, A.; Baltrunas, L.; Tikk, D. Session-based recommendations with recurrent neural networks. arXiv preprint arXiv:1511.06939 2015.
- Pi, Q.; Bian, W.; Zhou, G.; Zhu, X.; Gai, K. Practice on long sequential user behavior modeling for click-through rate prediction. In Proceedings of the Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, 2019, pp. 2671–2679.
- Pi, Q.; Zhou, G.; Zhang, Y.; Wang, Z.; Ren, L.; Fan, Y.; Zhu, X.; Gai, K. Search-based user interest modeling with lifelong sequential behavior data for click-through rate prediction. In Proceedings of the Proceedings of the 29th ACM International Conference on Information & Knowledge Management, 2020, pp. 2685–2692.
- Chang, J.; Gao, C.; Zheng, Y.; Hui, Y.; Niu, Y.; Song, Y.; Jin, D.; Li, Y. Sequential recommendation with graph neural networks. In Proceedings of the Proceedings of the 44th international ACM SIGIR conference on research and development in information retrieval, 2021, pp. 378–387.
- Xu, C.; Zhao, P.; Liu, Y.; Xu, J.; S. Sheng, V.S.S.; Cui, Z.; Zhou, X.; Xiong, H. Recurrent convolutional neural network for sequential recommendation. In Proceedings of the The world wide web conference, 2019, pp. 3398–3404.
- Zhao, W.; Wang, B.; Ye, J.; Gao, Y.; Yang, M.; Chen, X. PLASTIC: Prioritize Long and Short-term Information in Top-n Recommendation using Adversarial Training. In Proceedings of the Ijcai, 2018, pp. 3676–3682.
- Zheng, Y.; Gao, C.; Li, X.; He, X.; Li, Y.; Jin, D. Disentangling user interest and conformity for recommendation with causal embedding. In Proceedings of the Proceedings of the Web Conference 2021, 2021, pp. 2980–2991.
- Dong, D.; Zheng, X.; Zhang, R.; Wang, Y. Recurrent Collaborative Filtering for Unifying General and Sequential Recommender. In Proceedings of the IJCAI, 2018, pp. 3350–3356.
- Bai, T.; Du, P.; Zhao, W.X.; Wen, J.R.; Nie, J.Y. A long-short demands-aware model for next-item recommendation. arXiv preprint arXiv:1903.00066 2019.
- Hu, L.; Li, C.; Shi, C.; Yang, C.; Shao, C. Graph neural news recommendation with long-term and short-term interest modeling. Information Processing & Management 2020, 57, 102142.
- Bian, Z.; Zhou, S.; Fu, H.; Yang, Q.; Sun, Z.; Tang, J.; Liu, G.; Liu, K.; Li, X. Denoising User-aware Memory Network for Recommendation. In Proceedings of the Fifteenth ACM Conference on Recommender Systems, 2021, pp. 400–410.
- Kang, W.C.; McAuley, J. Self-attentive sequential recommendation. In Proceedings of the 2018 IEEE International Conference on Data Mining (ICDM). IEEE, 2018, pp. 197–206.
- Ma, C.; Ma, L.; Zhang, Y.; Sun, J.; Liu, X.; Coates, M. Memory augmented graph neural networks for sequential recommendation. In Proceedings of the Proceedings of the AAAI conference on artificial intelligence, 2020, Vol. 34, pp. 5045–5052.
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778.
- Srivastava, N.; Salakhutdinov, R.R. Multimodal learning with deep boltzmann machines. Advances in neural information processing systems 2012, 25.
- Volpi, R.; Morerio, P.; Savarese, S.; Murino, V. Adversarial feature augmentation for unsupervised domain adaptation. In Proceedings of the Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 5495–5504.
- Feng, Y.; Lv, F.; Shen, W.; Wang, M.; Sun, F.; Zhu, Y.; Yang, K. Deep session interest network for click-through rate prediction. arXiv preprint arXiv:1905.06482 2019.
- Zhang, J.; Ma, C.; Mu, X.; Zhao, P.; Zhong, C.; Ruhan, A. Recurrent convolutional neural network for session-based recommendation. Neurocomputing 2021, 437, 157–167.
- Yu, Z.; Lian, J.; Mahmoody, A.; Liu, G.; Xie, X. Adaptive User Modeling with Long and Short-Term Preferences for Personalized Recommendation. In Proceedings of the IJCAI, 2019, pp. 4213–4219.
- Zhou, K.; Yu, H.; Zhao, W.X.; Wen, J.R. Filter-enhanced MLP is all you need for sequential recommendation. In Proceedings of the Proceedings of the ACM Web Conference 2022, 2022, pp. 2388–2399.
- Tang, J.; Wang, K. Personalized top-n sequential recommendation via convolutional sequence embedding. In Proceedings of the Proceedings of the eleventh ACM international conference on web search and data mining, 2018, pp. 565–573.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).