咨询与建议

看过本文的还看了

相关文献

该作者的其他文献

文献详情 >Generalized Kullback-Leibler D... 收藏
arXiv

Generalized Kullback-Leibler Divergence Loss

作     者:Cui, Jiequan Zhu, Beier Xu, Qingshan Tian, Zhuotao Qi, Xiaojuan Yu, Bei Zhang, Hanwang Hong, Richang 

作者机构:College of Computing & Data Science Nanyang Technological University Singapore Department of Computer Science & Engineering The Chinese University of Hong Kong Sha Tin Hong Kong The University of Hong Kong Hong Kong Harbin Institution of Technology China Hefei University of Technology China 

出 版 物:《arXiv》 (arXiv)

年 卷 期:2025年

核心收录:

主  题:Mean square error 

摘      要:In this paper, we delve deeper into the Kullback–Leibler (KL) Divergence loss and mathematically prove that it is equivalent to the Decoupled Kullback-Leibler (DKL) Divergence loss that consists of 1) a weighted Mean Square Error (wMSE) loss and 2) a Cross-Entropy loss incorporating soft labels. Thanks to the decoupled structure of DKL loss, we have identified two areas for improvement. Firstly, we address the limitation of KL loss in scenarios like knowledge distillation by breaking its asymmetric optimization property along with a smoother weight function. This modification effectively alleviates convergence challenges in optimization, particularly for classes with high predicted scores in soft labels. Secondly, we introduce class-wise global information into KL/DKL to reduce bias arising from individual samples. With these two enhancements, we derive the Generalized Kullback–Leibler (GKL) Divergence loss and evaluate its effectiveness by conducting experiments on CIFAR-10/100, ImageNet, and vision-language datasets, focusing on adversarial training, and knowledge distillation tasks. Specifically, we achieve new state-of-the-art adversarial robustness on the public leaderboard — RobustBench and competitive knowledge distillation performance across CIFAR/ImageNet models and CLIP models, demonstrating the substantial practical merits. Our code is available at https://***/jiequancui/DKL. © 2025, CC BY-NC-SA.

读者评论 与其他读者分享你的观点

用户名:未登录
我的评分