咨询与建议

看过本文的还看了

相关文献

该作者的其他文献

文献详情 >InstKD: Towards Lightweight 3D... 收藏
IEEE Transactions on Intelligent Vehicles

InstKD: Towards Lightweight 3D Object Detection With Instance-Aware Knowledge Distillation

作     者:Zhang, Haonan Liu, Longjun Huang, Yuqi Lei, Xinyu Tong, Lei Wen, Bihan 

作者机构:National Key Laboratory of Human-Machine Hybrid Augmented Intelligence National Engineering Research Center for Visual Information and Applications and Institute of Artificial Intelligence and Robotics Xi'an Jiaotong University Xi'an China State Key Laboratory of Advanced Rail Autonomous Operation Beijing Jiaotong University Beijing China School of Electrical and Electronic Engineering Nanyang Technological University Singapore 

出 版 物:《IEEE Transactions on Intelligent Vehicles》 (IEEE Trans. Intell. Veh.)

年 卷 期:2024年

页      面:1-13页

核心收录:

学科分类:0710[理学-生物学] 0808[工学-电气工程] 0817[工学-化学工程与技术] 08[工学] 0703[理学-化学] 0835[工学-软件工程] 0836[工学-生物工程] 0811[工学-控制科学与工程] 0812[工学-计算机科学与技术(可授工学、理学学位)] 0823[工学-交通运输工程] 

主  题:Object detection 

摘      要:Deep neural network (DNN) is extensively explored for LiDAR-based 3D object detection, a crucial perception task in the field of autonomous driving. However, the presence of redundant parameters and complex computations pose challenges for the practical deployment of DNNs. Despite knowledge distillation (KD) is an effective approach for accelerating models, extremely small number of efforts explore its potential on LiDARbased 3D detectors. Besides, existing studies neglect to elaborately investigate 3D voxel-wise features for compression. To this end, we propose instance-aware knowledge distillation (InstKD) for 3D detector compression. The proposed method conducts KD by fully excavating two types of knowledge related to 3D voxelwise features. Firstly, the 3D voxel-wise feature of teacher is transferred to teach the student. In order to prioritize the knowledge with strong guiding capacity, we introduce expanded bounding box (E-Bbox) to distinguish and balance the foreground and background regions. Besides, we generate contribution map (CM) by calculating the gap between the classification response of teacher and student models to further dynamically balance individual instance for distillation. Secondly, we also align the relation-based knowledge of 3D voxel-wise features between the distillation pairs. To avoid incalculable relation on a massive number of 3D voxel-wise features, we distill the relation among instances selected by E-Bboxes, where the intra-relation of homogeneous instances and inter-relation of heterogeneous instances are transferred in a dual-pathway manner. In the experiments, we compress different models on benchmarks with varying scales. The results demonstrate that our method achieves the lightweight 3D detector with slight performance drop. For example, on KITTI dataset, our 2× compressed SECOND (75.5% parameters and 74.5% FLOPs reduction) achieves 66.83% mAP, surpassing its teacher model. The key code is available at https://***/zhnxj

读者评论 与其他读者分享你的观点

用户名:未登录
我的评分