A novel real-time infrared pedestrian detection algorithm is introduced in this study. The proposed approach leverages re-parameterized convolution and channel-spatial location fusion attention to tackle the difficult...
详细信息
A novel real-time infrared pedestrian detection algorithm is introduced in this study. The proposed approach leverages re-parameterized convolution and channel-spatial location fusion attention to tackle the difficulties presented by low-resolution, partial occlusion, and environmental interference in infrared pedestrian images. These factors have historically hindered the accurate detection of pedestrians using traditional algorithms. First, to tackle the problem of weak featurerepresentation of infrared pedestrian targets caused by low resolution and partial occlusion, a new attention module that integrates channel and spatial is devised and introduced to CSPDarkNet53 to design a new backbone CSLF-DarkNet53. The designed attention model can enhance the feature expression ability of pedestrian targets and make pedestrian targets more prominent in complex backgrounds. Second, to enhance the efficiency of detection and accelerate convergence, a multi-branch decoupled detector head is designed to operate the classification and location of infrared pedestrians separately. Finally, to improve poor real-time without losing precision, we introduce the re-parameterized convolution (repconv) using parameter identity transformation to decouple the training process and detection process. During the training procedure, to enhance the fitting ability of small convolution kernels, a multi-branch structure with convolution kernels of different scales is designed. Compared with the nice classical detection algorithms, the results of the experiment show that the proposed RCSLFNet not only detects partial occlusion infrared pedestrians in complex environments accurately but also has better real-time performance on the KAIST dataset. The mAP@0.5 reaches 86% and the detection time is 0.0081 s, 2.9% higher than the baseline.
Ships are important targets for marine surveillance in both military and civilian domains. Since the rise of deep learning, ship detection in synthetic aperture radar (SAR) images has achieved significant progress. Ho...
详细信息
Ships are important targets for marine surveillance in both military and civilian domains. Since the rise of deep learning, ship detection in synthetic aperture radar (SAR) images has achieved significant progress. However, the variability in ship size and resolution, especially the widespread presence of numerous small-sized ships, continues to pose challenges for effective ship detection in SAR images. To address the challenges posed by small ship targets, we propose an enhanced YOLO network to improve the detection accuracy of small targets. Firstly, we propose a Shuffle re-parameterization (SR) module as a replacement for the C2f module in the original YOLOv8 network. The SR module employs re-parameterized convolution along with channel shuffle operations to improve feature extraction capabilities. Secondly, we employ the space-to-depth (SPD) module to perform down-sampling operations within the backbone network, thereby reducing the information loss associated with pooling operations. Thirdly, we incorporate a Hybrid Attention (HA) module into the neck network to enhance the featurerepresentation of small ship targets while mitigating the interference caused by surrounding sea clutter and speckle noise. Finally, we add the shape-NWD loss to the regression loss, which emphasizes the shape and scale of the bounding box and mitigates the sensitivity of Intersection over Union (IoU) to positional deviations in small ship targets. Extensive experiments were carried out on three publicly available datasets-namely, LS-SSDD, HRSID, and iVision-MRSSD-to demonstrate the effectiveness and reliability of the proposed method. In the small ship dataset LS-SSDD, the proposed method exhibits a notable improvement in average precision at an IoU threshold of 0.5 (AP50), surpassing the baseline network by over 4%, and achieving an AP50 of 77.2%. In the HRSID and iVision-MRSSD datasets, AP50 reaches 91% and 95%, respectively. Additionally, the average precision for small targets
This paper proposes an efficient cloud detection algorithm for Sustainable Development Scientific Satellite (SDGSAT-1) data. The core work includes the following: (1) constructing a SDGSAT-1 cloud detection dataset co...
详细信息
This paper proposes an efficient cloud detection algorithm for Sustainable Development Scientific Satellite (SDGSAT-1) data. The core work includes the following: (1) constructing a SDGSAT-1 cloud detection dataset containing five types of elements: clouds, cloud shadow, snow, water body, and land, with a total of 15,000 samples;(2) designing a multi-scale convolutional attention unit (RDE-MSCA) based on a gated linear unit (GLU), with parallel re-parameterized convolution (repConv) and detail-enhanced convolution (DEConv). This design focuses on improving the featurerepresentation and edge detail capture capabilities of targets such as clouds, cloud shadow, and snow. Specifically, the repConv branch focuses on learning a new global representation, reconstructing the original multi-branch deep convolution into a single-branch structure that can efficiently fuse channel features, reducing computational and memory overhead. The DEConv branch, on the other hand, uses differential convolution to enhance the extraction of high-frequency information, and is equivalent to a normal convolution in the form of re-parameterization during the inference stage without additional overhead;GLU then realizes adaptive channel-level information regulation during the multi-branch fusion process, which further enhances the model's discriminative power for easily confused objects. It is integrated into the SegNeXt architecture based on RDE-MSCA and proposed as RDE-SegNeXt. Experiments show that this model can achieve 71.85% mIoU on the SDGSAT-1 dataset with only about 1/12 the computational complexity of the Swin-L model (a 2.71% improvement over Swin-L and a 5.26% improvement over the benchmark SegNeXt-T). It also significantly improves the detection of clouds, cloud shadow, and snow. It achieved competitive results on both the 38-Cloud and LoveDA public datasets, verifying its effectiveness and versatility.
A forest fire is a natural disaster characterized by rapid spread, difficulty in extinguishing, and widespread destruction, which requires an efficient response. Existing detection methods fail to balance global and l...
详细信息
Satellite cloud imagery is pivotal for meteorologists in characterizing weather patterns, detecting climate anomaly regions, and predicting rain effects. The task of satellite cloud image forecasting is crucial, and w...
详细信息
Satellite cloud imagery is pivotal for meteorologists in characterizing weather patterns, detecting climate anomaly regions, and predicting rain effects. The task of satellite cloud image forecasting is crucial, and while deep learning models have shown promise in predicting spatio-temporal data, traditional methods face challenges with extracting long-term spatio-temporal features and high computation costs. To address these issues, we propose the re -parameterized Sequence -to -Sequence Satellite Cloud Imagery Prediction Network (repSSCIPN). rep-SSCIPN utilizes rep -convolution layers to reduce inference -time cost and memory consumption, enhancing efficiency by converting re -parameterized blocks into a single convolution layer during inference. The sequence normalization attention mechanism in rep-SSCIPN highlights crucial feature sequences and establishes their inter -dependencies. We validate our novel method using a real -world satellite cloud image dataset from the meteorological satellite "Himawari." Experimental results showcase significant improvements in prediction accuracy and reconstruction quality compared to ConvLSTM, PredRNN, FCLSTM, LMC, SimVP and SCSTque models. The efficiency gains make rep-SSCIPN a promising advancement for satellite cloud image prediction. ARTICLE INFO.
暂无评论