咨询与建议

看过本文的还看了

相关文献

该作者的其他文献

文献详情 >One Person, One Model-Learning... 收藏
arXiv

One Person, One Model-Learning Compound Router for Sequential Recommendation

作     者:Liu, Zhiding Cheng, Mingyue Li, Zhi Liu, Qi Chen, Enhong 

作者机构:Anhui Province Key Laboratory of Big Data Analysis and Application University of Science and Technology of China Hefei China State Key Laboratory of Cognitive Intelligence Hefei China Shenzhen International Graduate School Tsinghua University Shenzhen China 

出 版 物:《arXiv》 (arXiv)

年 卷 期:2022年

核心收录:

主  题:Network architecture 

摘      要:Deep learning has brought significant breakthroughs in sequential recommendation (SR) for capturing dynamic user interests. A series of recent research revealed that models with more parameters usually achieve optimal performance for SR tasks, inevitably resulting in great challenges for deploying them in real systems. Following the simple assumption that light networks might already suffice for certain users, in this work, we propose CANet, a conceptually simple yet very scalable framework for assigning adaptive network architecture in an input-dependent manner to reduce unnecessary computation. The core idea of CANet is to route the input user behaviors with a light-weighted router module. Specifically, we first construct the routing space with various submodels parameterized in terms of multiple model dimensions such as the number of layers, hidden size and embedding size. To avoid extra storage overhead of the routing space, we employ a weight-slicing schema to maintain all the submodels in exactly one network. Furthermore, we leverage several solutions to solve the discrete optimization issues caused by the router module. Thanks to them, CANet could adaptively adjust its network architecture for each input in an end-to-end manner, in which the user preference can be effectively captured. To evaluate our work, we conduct extensive experiments on benchmark datasets. Experimental results show that CANet reduces computation by 55 ∼ 65% while preserving the accuracy of the original model. Our codes are available at https://***/icantnamemyself/CANet. Copyright © 2022, The Authors. All rights reserved.

读者评论 与其他读者分享你的观点

用户名:未登录
我的评分