版权所有:内蒙古大学图书馆 技术提供:维普资讯• 智图
内蒙古自治区呼和浩特市赛罕区大学西街235号 邮编: 010021
作者机构:Univ Warwick WMG Coventry CV4 7AL England Onsemi Greenwood House Bracknell RG12 2AA England
出 版 物:《IEEE ACCESS》 (IEEE Access)
年 卷 期:2025年第13卷
页 面:21695-21706页
核心收录:
基 金:European Union Innovate UK Royal Academy of Engineering under the Industrial Fellowships Scheme Horizon Europe - Pillar II Funding Source: Horizon Europe - Pillar II
主 题:Image color analysis Artificial neural networks Accuracy Pedestrians Cameras Automotive engineering Machine learning Pipelines Feature extraction Europe Bayer data object detection perception sensors assisted and automated driving intelligent vehicles
摘 要:Whilst Deep Neural Networks (DNNs) have been developing swiftly, most of the research has been focused on videos based on RGB frames. RGB data has been traditionally optimised for human vision and is a highly re-elaborated and interpolated version of the collected raw data. In fact, the sensor collects the light intensity value per pixel, but an RGB frame contains 3 values, for red, green, and blue colour channels. This conversion to RGB requires computational resource, time, power, and increases by a factor of three the amount of output data. This work investigates DNN based detection using (for training and evaluation) Bayer frames, generated from a benchmarking automotive dataset (i.e. KITTI dataset). A Deep Neural Network (DNN) is deployed in an unmodified form, and also modified to accept only single channel frames, such as Bayer frames. Eleven different re-trained versions of the DNN are produced, and cross-evaluated across different data formats. The results demonstrate that the selected DNN has the same accuracy when evaluating RGB or Bayer data, without significant degradation in the perception (the variation of the Average Precision is 1%). Moreover, the colour filter array position and the colour correction matrix do not seem to contribute significantly to the DNN performance. This work demonstrates that Bayer data can be used for object detection in automotive without significant perception performance loss, allowing for more efficient sensing-perception systems.