Synthetic aperture radar (SAR) image target detection and recognition (SAR-TDR) tasks have become research hot spots in the remotesensing application. These targets include ships, vehicles, aircraft, oil tanks, bridg...
详细信息
Synthetic aperture radar (SAR) image target detection and recognition (SAR-TDR) tasks have become research hot spots in the remotesensing application. These targets include ships, vehicles, aircraft, oil tanks, bridges, and so on. However, with the rapid development of SAR technology and increasingly complex electromagnetic environment, complex characteristics of SAR images bring severe challenges to the accurate SAR-TDR via traditional physical models or manual feature-extraction-based machine learning methods. In recent years, deep learning (DL), as a powerful automatic feature extraction algorithm, has been widely used in the computer vision domain. More specifically, DL has also been introduced into the SAR-TDR tasks and has effectively achieved good performance in terms of accuracy, real-time processing, etc. With the rapid development of DL, SAR imageprocessing, and practical requirements of SAR-TDR in civilian and military domains, it is crucial to conduct a systematic survey on SAR-TDR in the past few years. In this survey article, we mainly conduct a systematic overview of DL-based SAR-TDR literature on two tasks, i.e., target recognition (e.g., ground vehicles, ships, and aircraft) and target detection (e.g., ships, aircraft, change detection, sea surface oil spills, and oil tanks). More specifically, our related works about these topics are also presented to verify the effectiveness of DL-based methods. First, several DL methods (e.g., convolutional neural networks, recurrent neural networks, autoencoders, and generative adversarial networks), commonly used in SAR-TDR, are briefly introduced. Then, a systematic review of DL-based SAR-TDR (including our related works) is presented. Finally, the current challenges and future possible research directions are deeply analyzed and discussed.
To address the challenges posed by complex backgrounds and target scale variations in remotesensing aircraft detection, we propose an improved YOLOv7-tiny network designed to boost both detection accuracy and the mod...
详细信息
To address the challenges posed by complex backgrounds and target scale variations in remotesensing aircraft detection, we propose an improved YOLOv7-tiny network designed to boost both detection accuracy and the model's ability to generalize across varying conditions. Complex backgrounds hinder effective target feature extraction, while variations in target scale complicate the real-time detection of objects at different sizes. To address these issues, we introduce a multi-head dynamic joint self-attention mechanism that combines self-attention, channel attention, and spatial attention, thereby enhancing the model's ability to capture crucial target features. Furthermore, we design an enhanced feature extraction module that expands the receptive field and improves the model's ability to extract features from multi-scale targets. In addition, we constructed diverse remotesensing aircraft detection datasets and applied a range of data augmentation techniques to improve the robustness and adaptability of the model in real-world scenarios. The experimental results demonstrate that the enhanced YOLOv7-tiny model achieves a mAP of 93.41%, which is a 2.98% improvement over the original model. The model also achieves precision and recall rates of 90.44% and 87.81%, respectively, with a FPS rate of 23.31. These results highlight the model's excellence in both precision and real-time performance, making it highly effective for remotesensing aircraft detection and fine-grained feature recognition, particularly in complex environments where real-time processing is crucial.
With the widespread application and rapid development of remotesensing technology, the quality requirements for remotesensingimages are gradually improving. Currently, relying solely on one sensor is difficult to e...
详细信息
Convolutional neural networks have demonstrated remarkable capability in extracting deep semantic features from images, leading to significant advancements in various imageprocessing tasks. This success has also open...
详细信息
Convolutional neural networks have demonstrated remarkable capability in extracting deep semantic features from images, leading to significant advancements in various imageprocessing tasks. This success has also opened up new possibilities for change detection (CD) in remotesensing applications. But unlike the conventional imagerecognition tasks, the performance of AI models in CD heavily relies on the method used to fuse the features from two different phases of the image. The existing deep-learning-based methods for CD typically fuse features of bitemporal images using difference or concatenation techniques. However, these approaches often fail tails to prioritize potential change areas adequately and neglect the rich contextual information essential for discerning subtle changes, potentially leading to slower convergence speed and reduced accuracy. To tackle this challenge, we propose a novel feature fusion approach called feature-difference attention-based feature fusion CD network. This method aims to enhance feature fusion by incorporating a feature-difference attention-based feature fusion module, enabling a more focused analysis of change areas. Additionally, a deep-supervised attention module is implemented to leverage the deep surveillance module for cascading refinement of change areas. Furthermore, an atrous spatial pyramid pooling fast is employed to efficiently acquire multiscale object information. The proposed method is evaluated on two publicly available datasets, namely the WHU-CD and LEVIR-CD datasets. Compared with the state-of-the-art CD methods, the proposed method outperforms in all metrics, with an intersection over union of 92.49% and 85.56%, respectively.
Existing super-resolution reconstruction algorithms for remotesensingimages often struggle to fully extract and utilize features in complex scenes, and the reconstruction results are not optimal due to the influence...
详细信息
remotesensing scene classification is a critical task in the processing and analysis of remotesensingimages. Traditional methods typically use standard convolutional kernels to extract feature information. Although...
详细信息
remotesensing scene classification is a critical task in the processing and analysis of remotesensingimages. Traditional methods typically use standard convolutional kernels to extract feature information. Although these methods have seen improvements, they still struggle to fully capture unique local details, thus affecting classification accuracy. Each category within remotesensing scenes has its unique local details, such as the rectangular features of buildings in schools or industrial areas, as well as bridges and roads in parks or squares. The most important features are often these rectangular structures and their spatial positions, which standard convolutional kernels find challenging to capture *** address this issue, we propose a remotesensing scene classification method based on a Rectangle Convolution Self-Attention Fusion Network (RCSFN) architecture. In the RCSFN network, the Rectangle Convolution Maximum Fusion (RCMF) module operates in parallel with the first 4 x 4 convolutional layer of VanillaNet-5. The RCMF module uses two different rectangular convolutional kernels to extract different receptive fields, enhancing the extraction of shallow local features through addition and fusion. This process, combined with the concatenation of the original input features, results in richer local detail ***, we introduce an Area Selection (AS) module that focuses on selecting feature information within local regions. The Sequential Polarisation Self-Attention (SPS) mechanism, integrated with the Mini Region Convolution (MRC) module through feature multiplication, enhances important features and improves spatial positional relationships, thereby increasing the accuracy of recognising categories with rectangular or elongated features. Experiments were carried out on AID and NWPU-RESISC45 data sets, and the overall classification accuracy was 96.56% and 92.46%, respectively. This shows that the RCSFN network model proposed in t
With the rapid development of remotesensing technology, remotesensingimages play an important role in the agricultural field, geological field, and natural disaster detection. The size of aircraft in complex scenes...
详细信息
Underwater pile foundation detection is crucial for environmental monitoring and marine engineering. Traditional methods for detecting underwater pile foundations are labor-intensive and inefficient. Deep learning-bas...
详细信息
Underwater pile foundation detection is crucial for environmental monitoring and marine engineering. Traditional methods for detecting underwater pile foundations are labor-intensive and inefficient. Deep learning-based imageprocessing has revolutionized detection, enabling identification through sonar imagery analysis. This study proposes an innovative methodology, named the AquaPile-YOLO algorithm, for underwater pile foundation detection. Our approach significantly enhances detection accuracy and robustness by integrating multi-scale feature fusion, improved attention mechanisms, and advanced data augmentation techniques. Trained on 4000 sonar images, the model excels in delineating pile structures and effectively identifying underwater targets. Experimental data show that the model can achieve good target identification results in similar experimental scenarios, with a 96.89% accuracy rate for underwater target recognition.
This paper addresses the problem of blind deblurring of single remotesensing (RS) images with deep neural networks. Most existing deep learning-based methods are migrated from natural image deblurring models, disrega...
详细信息
This paper addresses the problem of blind deblurring of single remotesensing (RS) images with deep neural networks. Most existing deep learning-based methods are migrated from natural image deblurring models, disregarding the domain gap to remotesensingimages. Besides, the image deblurring problem is typically considered as an independent low-level image pre-processing, taking no account of downstream tasks, such as classification and segmentation. In this paper, we first present a novel decoder with a par-allel fusion stream for fusing multi-scale RS features and expanding the receptive field. Then, to generate high-quality sharp RS images, we propose to calculate the perceptual loss on an RS-pre-trained network instead of computing on VGG19 pre-trained on natural images i.e. imageNet. For bridging the RS im-age deblurring results to the downstream recognition tasks, we further propose a semantic loss, which is calculated on the last-layer feature map of an RS segmentation network. With extensive experiments con-ducted on public RS image datasets, we demonstrate that the proposed method improves results for RS image deblurring and achieves competitive performance both qualitatively and quantitatively. Moreover, downstream recognition experiments validate the superior quality of the recovered images over existing methods.& COPY;2023 Elsevier B.V. All rights reserved.
High-resolution wide swath (HRWS) synthetic aperture radar (SAR) systems are normally designed to have identical antenna patterns in each receive channel. Nevertheless, due to different factors, this condition might n...
详细信息
High-resolution wide swath (HRWS) synthetic aperture radar (SAR) systems are normally designed to have identical antenna patterns in each receive channel. Nevertheless, due to different factors, this condition might not be satisfied, resulting in a multichannel radar system with different antenna patterns. This letter studies the impact of these relative antenna differences on the performance of a state-of-the-art azimuth motion-adaptive image reconstruction for a real airborne SAR sensor with multiple azimuth channels on receive. The performance of the reconstruction algorithm is evaluated both when these relative differences are neglected and when they are accounted for in the multichannel reconstruction process. Additionally, the range dependence of the antenna pattern as a function of wavenumber and the use of range block processing to minimize the impact of this dependence are analyzed. The results confirm the relevance of including these additional steps in the reconstruction, extending the understanding of SAR systems using digital beamforming (DBF) in azimuth.
暂无评论