In the field of remotesensing, the acquired images are often severely degraded due to adverse weather conditions, such as haze and raindrops, posing significant challenges for subsequent visual tasks. Although CNN an...
详细信息
In the field of remotesensing, the acquired images are often severely degraded due to adverse weather conditions, such as haze and raindrops, posing significant challenges for subsequent visual tasks. Although CNN and Transformer have been widely applied to address these issues, they struggle to balance the relationship between global scene recovery, local detail preservation, and computational efficiency, leading to an imbalance between model performance and efficiency. To this end, we propose a lightweight and efficient visual state space model for remotesensingimage restoration. Specifically, we propose the Efficient Vision Mamba Block as the core component of the model, incorporating the State Space Model to leverage its linear complexity for modeling long-range dependencies. Furthermore, we design a multi-router scanning strategy to perform global modeling of remotesensingimages, capturing large spatial features from different routes and directions. Compared with existing methods that employ fixed-direction scanning, our approach avoids information redundancy caused by repeated scanning, making the model better adaptable to the complex and changeable weather conditions. Extensive experiments validate the superiority of our proposed model, outperforming state-of-the-art methods on both the StateHaze1k and UAV-Rain1k datasets.
remotesensing object detection, with large differences in object size, arbitrary orientation and tight arrangement, leads to difficulties in object recognition and localization. Therefore, a remotesensingimage obje...
详细信息
remotesensing object detection, with large differences in object size, arbitrary orientation and tight arrangement, leads to difficulties in object recognition and localization. Therefore, a remotesensingimage object Detector (BAIDet) based on Background and Angle Information is proposed in this paper. Firstly, a large convolutional kernel global attention module is designed to fully utilize the global information of remotesensingimages by expanding the receptive field. And obtain the edge information of ground objects through deformable convolution. Secondly, an angle-sensitive probabilistic intersection-over-union loss function (AS-ProbIoU Loss) is developed for bounding box regression for oriented object detection. Finally, experimental results on four remotesensingimage datasets, DOTA, HRSC 2016, UCAS-AOD, and DIOR-R, demonstrated the effectiveness of this method.
remotesensingimage denoising tasks are challenged by complex noise distributions and multiple noise types, including a mixture of additive Gaussian white noise (AWGN) and impulse noise (IN). For better image recover...
详细信息
remotesensingimage denoising tasks are challenged by complex noise distributions and multiple noise types, including a mixture of additive Gaussian white noise (AWGN) and impulse noise (IN). For better image recovery, complex contextual information needs to be balanced while maintaining spatial details. In this paper, a denoising model based on multilevel progressive image recovery is proposed to address the problem of remotesensingimage denoising. In our model, the deformable convolution improves spatial feature sampling to effectively capture image details. Meanwhile, attention-guided filtering is used to pass the output images from the first and second stages to the third stage in order to prevent information loss and optimize the image recovery effect. The experimental results show that under the mixed noise scene of Gaussian and pepper noise, our proposed model shows superior performance relative to several existing methods in terms of both visual effect and objective evaluation indexes. Our model can effectively reduce the influence of image noise and recover more realistic image details.
The proceedings contain 33 papers. The topics discussed include: fast on-board ship detection in panchromatic images based on micro-nano satellite;star centroid extraction in the image sequences of a discrete detector...
ISBN:
(纸本)9781510655379
The proceedings contain 33 papers. The topics discussed include: fast on-board ship detection in panchromatic images based on micro-nano satellite;star centroid extraction in the image sequences of a discrete detector array on geostationary orbit;two-step registration of near-space remotesensingimages via deep neural networks;accuracy improvement with input image upscaling for in-orbit object detection;an automatic approach for ice-block falls detection in the Martian north polar region;compressed domain classification of remotesensing scene images based on sub-band data fusion;accelerating a two-stage object detector for high quality in-orbit remotesensing;fire image detection based on clustering data mining techniques;fusion of optical and SAR satellite data for environmental monitoring: assessment of damages and disturbances originated by forest fires;and multi-temporal change detection based on China’s domestic hyperspectral remotesensing satellite images.
Combining convolutional neural networks (CNNs) and transformers is a crucial direction in remotesensingimage semantic segmentation. However, due to differences in the spatial information focus and feature extraction...
详细信息
Combining convolutional neural networks (CNNs) and transformers is a crucial direction in remotesensingimage semantic segmentation. However, due to differences in the spatial information focus and feature extraction methods, existing feature transfer and fusion strategies do not effectively integrate the advantages of both approaches. To address these issues, we propose a CNN-transformer hybrid network for precise remotesensingimage semantic segmentation. We propose a novel Swin Transformer block to optimize feature extraction and enable the model to handle remotesensingimages of arbitrary sizes. Additionally, we design an Edge Spatial Attention module to focus attention on local edge structures, effectively integrating global features and local details. This facilitates efficient information flow between the Transformer encoder and CNN decoder. Finally, a multi-scale convolutional decoder is employed to fully leverage both global information from the Transformer and local features from the CNN, leading to accurate segmentation results. Our network achieved state-of-the-art performance on the Vaihingen and Potsdam datasets, reaching mIoU and F1 scores of 67.37% and 79.82%, as well as 72.39% and 83.68%, respectively.
作者:
Liu, JingyiYang, XiaominSichuan Univ
Coll Elect & Informat Engn Chengdu 610065 Sichuan Peoples R China Hubei Minzu Univ
Coll Intelligent Syst Sci & Engn Enshi 445000 Hubei Peoples R China Hubei Minzu Univ
Key Lab Green Mfg Superlight Elastomer Mat State Ethn Affairs Commiss Enshi 445000 Hubei Peoples R China
remotesensing super-resolution remains a research hotspot in the field of remotesensing. Unlike natural images, remotesensingimages typically possess rich scenes and complex spatial structures, making it challengi...
详细信息
remotesensing super-resolution remains a research hotspot in the field of remotesensing. Unlike natural images, remotesensingimages typically possess rich scenes and complex spatial structures, making it challenging to enhance their resolution. This paper introduces a hybrid attention module that integrates CNNs, Transformers, and spatial attention to effectively extract multi-level features from remotesensingimages. The module seamlessly integrates the early feature projection processes of CNNs and Transformers, harmoniously blending CNNs' strength in local modeling with Transformers' expertise in global modeling, thereby simplifying the complexity of fusion. To accurately preserve the intricate details of the input images, we construct a structure enhancement module that focuses on extracting edge and linear structural details to retain the complex features of the images. Furthermore, we build a multi-level remotesensing super-resolution network. A series of experiments conducted on the AID dataset demonstrate its excellent effectiveness and generalization capabilities.
The objective of super-resolution in remotesensingimagery is to enhance low-resolution images to recover high-quality details. With the rapid progress of deep learning technology, the deep learning-based super-resol...
详细信息
The objective of super-resolution in remotesensingimagery is to enhance low-resolution images to recover high-quality details. With the rapid progress of deep learning technology, the deep learning-based super-resolution technology for remotesensingimages has also made remarkable achievements. However, these methods encounter several challenges. They often struggle with processing long-range spatial information that encompasses complex scene changes, adversely affecting the image's coherence and accuracy. Furthermore, the lack of connectivity in feature extraction blocks hinders effective feature utilization in deeper network layers, leading to issues such as gradient vanishing and exploding. Additionally, constraints in the spatial domain of previous methods frequently result in severe shape distortion and blurring. To address these issues, this study proposes the CFFormer, a new super-resolution framework that employs the Swin Transformer as its core architecture and incorporates the Channel Fourier Block (CFB) to refine features in the frequency domain. The Global Attention Block (GAB) is also integrated to enhance global information capture, thereby improving the extraction of spatial features. To increase model stability and feature utilization efficiency, a Jump-Joint Fusion Mechanism is designed, culminating in a Residual Fusion Swin Transformer Block (RFSTB) that alleviates the gradient vanishing issue and optimizes feature reuse. Experimental results confirm the CFFormer's superior performance in remotesensingimage reconstruction, demonstrating outstanding perceptual quality and reliability. Notably, the CFFormer achieves a Peak signal-to-Noise Ratio (PSNR) of 29.83 dB on the UcMercedx4 dataset, surpassing the SwinIR method by approximately 0.5 dB, indicating a substantial enhancement.
Haze causes information loss and quality degradation in remotesensingimages. Unsupervised learning-based dehazing methods aim to reduce reliance on paired hazy images and their labels. However, complex mapping relat...
详细信息
ISBN:
(纸本)9798350344868;9798350344851
Haze causes information loss and quality degradation in remotesensingimages. Unsupervised learning-based dehazing methods aim to reduce reliance on paired hazy images and their labels. However, complex mapping relationships often increase the difficulty in network convergence, resulting in color distortion and loss of texture details in remotesensingimages. To address these issues, we propose an unsupervised haze removal method based on saliency-guided transmission refinement for remotesensingimages. Firstly, we introduce a saliency-guided transmission refinement method, which decomposes and recombines two transmission maps obtained under different conditions, guided by saliency information. Secondly, we propose a loss function comprising energy loss and texture loss. The energy loss provides an energy reference based on the coarse transmission estimation, while the texture loss enhances the preservation of texture details. Experimental results demonstrate that our method achieves comparable performance to several supervised methods.
Interferometric synthetic aperture radar (InSAR) denoising is an essential processing step in deformation measurement and topography reconstruction. A noisy InSAR phase image gives rise to the phase unwrapping difficu...
详细信息
Interferometric synthetic aperture radar (InSAR) denoising is an essential processing step in deformation measurement and topography reconstruction. A noisy InSAR phase image gives rise to the phase unwrapping difficulties and even results in the degradation of various final products of InSAR. To address this issue, we develop a compressive sensing (CS)-based InSAR phase denoising technique in this article. Since the spectrum of the InSAR phase image is usually sparse in the 2-D frequency domain, the estimation of sensing dictionary matrix of the linear system between the InSAR phase signal and its spectrum in the pursuit of sparsity is considered for InSAR phase denoising. The optimization problem derived by the signal parameterization approach is effectively carried out by estimating the basis function that is closely analogous to the strongest signal component in the spectrum of the InSAR phase image. The proposed method is effectively capable of eliminating noise and preserving detailed fringe information of InSAR. In the end, simulations and experimental results demonstrate that the proposed scheme outperforms other conventional InSAR phase denoising methods.
As the annotation of remotesensingimages requires domain expertise, it is difficult to construct a large-scale and accurate annotated dataset. image-level annotation data learning has become a research hotspot. In a...
详细信息
ISBN:
(纸本)9798350344868;9798350344851
As the annotation of remotesensingimages requires domain expertise, it is difficult to construct a large-scale and accurate annotated dataset. image-level annotation data learning has become a research hotspot. In addition, due to the difficulty in avoiding mislabeling, label noise cleaning is also a concern. In this paper, a semantic segmentation method for remotesensingimages based on uncertainty perception with noisy labels is proposed. The main contributions are three-fold. First, a label cleaning method based on iterative learning is presented to handle noise labels such as missing or incorrect annotations. Second, a two-stage semantic segmentation model is proposed for image-level annotation, which eliminates the need for post-processing steps during testing. Lastly, a complementary uncertainty perception function is introduced to improve the utilization of dataset features and enhance the accuracy of segmentation. The effectiveness of this method was verified through comprehensive evaluation with 7 models on four datasets.
暂无评论