Investigating ground objects widely distributed in geography and large in scale is one of the primary missions for satellite sensors. On the other hand, recognizing objects from images is one of the classic tasks for ...
详细信息
Investigating ground objects widely distributed in geography and large in scale is one of the primary missions for satellite sensors. On the other hand, recognizing objects from images is one of the classic tasks for convolutional neural networks (CNNs), currently the most popular computer vision technique. Data processing, such as data augmentation, channel selection, and image fusion, can be essential when applying CNNs to satellite images. With a case study of recognizing solar panels from satellite images using CNN, the related data processing issues are discussed, and an approach to embed channel fusion methods into CNN is established. As a result, the following findings are concluded from our case study: (1) not all channels in satellite images contribute to specific object recognition, and thus channel selection is necessary in applying CNN on satellite images;(2) fine-tuning the fusion method embedded in CNN improves the model stability;and (3) transfer learning is outperformed by CNN models trained with augmented data for object recognition from satellite images.
Building structural type (BST) information is vital for seismic risk and vulnerability modeling. However, obtaining this kind of information is not a trivial task. The conventional method involves a labor-intensive an...
详细信息
Building structural type (BST) information is vital for seismic risk and vulnerability modeling. However, obtaining this kind of information is not a trivial task. The conventional method involves a labor-intensive and inefficient manual inspection process for each building. Nowadays, a few methods have explored to use remotesensingimages and some building-related knowledge (BRK) to realize automated BST recognition. However, these methods have many limitations, such as insufficient mining of multimodal information and difficulty obtaining BRK, which hinders their promotion and practical use. To alleviate the shortcomings above, we propose a deep multimodal fusion model, which combines satellite optical remotesensingimage, aerial synthetic aperture radar image, and BRK (roof type, color, and group pattern) obtained by domain experts to achieve accurate automatic reasoning of BSTs. Specifically, first, we use a pseudo-siamese network to extract the image feature. Second, a knowledge graph (KG) based on the BRK is constructed, and then, we use a graph attention network to extract the semantic feature from the KG. Third, we propose a novel multistage gated fusion mechanism to fuse the image and semantic feature. Our method's best overall accuracy and kappa coefficient on the dataset collected in the study area are 90.35% and 0.83, which outperforms a series of existing methods. Through our model, high-precision BST information can be obtained for earthquake disaster prevention, reduction, and emergency decision making.
Guizhou Province, situated in the southwest of China, boasts diverse and complex geographical environments and abundant forest resources. However, it faces threats from natural disasters like forest fires. Accurate es...
详细信息
Most existing change detection (CD) methods target homogeneous images. However, in real-world scenarios like disaster management, where CD is urgent and pre-changed and post-changed images are typical of different mod...
详细信息
Most existing change detection (CD) methods target homogeneous images. However, in real-world scenarios like disaster management, where CD is urgent and pre-changed and post-changed images are typical of different modalities, significant challenges arise for multimodal change detection (MCD). One challenge is that bitemporal image pairs, sourced from distinct sensors, may cause an image domain gap. Another issue surfaces when multimodal bi-temporal image pairs require collaborative input from domain experts who are specialised among different image fields for pixel-level annotation, resulting in scarce annotated samples. To address these challenges, this paper proposes a novel self-supervised difference contrast learning framework (Self-DCF). This framework facilitates networks training without labelled samples by automatically exploiting the feature information inherent in bi-temporal imagery to supervise each other mutually. Additionally, a Unified Mapping Unit reduces the domain gap between different modal images. The efficiency and robustness of Self-DCF are validated on five popular datasets, outperforming state-of-the-art algorithms.
Generative adversarial networks (GANs) have demonstrated their remarkable capacity to learn the training data distribution and produce high-quality synthetic images, which have been widely adopted in imagerecognition...
详细信息
Generative adversarial networks (GANs) have demonstrated their remarkable capacity to learn the training data distribution and produce high-quality synthetic images, which have been widely adopted in imagerecognition tasks in remotesensing (RS) research communities. However, previous work has shown that using GANs does not preserve privacy, e.g., being susceptible to membership attacks, while sensitive information is vulnerable to nefarious activities. This drawback is considered severe in RS communities, in which critical researches highly value the security and privacy of the image content. Thus, to publicly share sensitive data for supporting critical researches and, in the meantime, guarantee the model accuracy trained from privacy-preserving data, this work develops GANs within the differential privacy (DP) framework and proposes an RS differentially private generative adversarial network (RS-DPGAN) for both privacy-preserving synthetic image generation and classification. Our RS-DPGAN is capable of releasing safe version of synthetic data while obtaining favorable classification results, which gives rigorous guarantees for the privacy of sensitive data and balance between the model accuracy and privacy-preserving degree. Both extensive empirical and statistical results confirm the effectiveness of our framework.
Extracting agricultural parcel boundaries from remotesensingimages based on deep learning methods is currently the most promising method. Due to the diversity of agricultural types and limitations of deep learning n...
详细信息
Extracting agricultural parcel boundaries from remotesensingimages based on deep learning methods is currently the most promising method. Due to the diversity of agricultural types and limitations of deep learning networks, parcel boundaries extracted by edge detection deep learning networks lack target specificity, and parcel boundaries extracted by semantic segmentation deep learning networks generally causes a loss of precise boundary information. Based on the principles of multi-task deep learning networks Psi-Net, BsiNet, and ResUNet-a, we constructed a parallel multi-task deep learning network M_ResUnet. The M_ResUnet is capable of performing edge detection and semantic segmentation simultaneously, and it is designed for extracting agricultural parcel boundaries from high-resolution remotesensingimages. To improve the effectiveness of parcel boundary recognition, the edge enhancement concept from the context-aware tracing strategy (CATS) is introduced into the M_ResUnet, which enhanced the continuity and effectiveness of parcel boundary recognition and accelerated network training convergence. Finally, we conducted experiments on three different study areas with different remotesensing data sources and agricultural parcel types. The experimental results demonstrate that the method we proposed achieved better continuity in identifying parcel boundaries, and it also improved the recognition of boundaries between neighboring parcels with strong similarities remotesensing features such as texture and spectral characteristics.
The increasing volume and heterogeneous nature of remotesensing data necessitate efficient and accurate image retrieval systems, particularly for land use and land cover (LULC) mapping. This paper presents an advance...
详细信息
The increasing volume and heterogeneous nature of remotesensing data necessitate efficient and accurate image retrieval systems, particularly for land use and land cover (LULC) mapping. This paper presents an advanced framework that integrates an optimized label propagation network (OLPNet) with a two-stage hybrid hierarchical classification (TS-H2C) approach for efficient remotesensingimage retrieval (RSIR). The TS-H2C-OLPNet framework employs two sparse kernel learning machines- the relevance vector machine (RVM) and support vector machine (SVM), to enhance label distribution in complex, high-dimensional feature spaces. The framework utilizes a reconstruction-based relational autoencoder (RAE) to extract robust deep features with reduced dimensionality. In the Stage-1 hierarchy, the RVM generates confidence scores to determine super-class labels, from there the OLPNet propagates respective sub-class labels to the Stage-2 SVM. This approach efficiently manages the search space, speeds up retrieval, and effectively handles highly overlapping LULC classes, improving the recognition of unseen categories. Extensive experiments on benchmark RSI datasets with challenging scene categories demonstrate that the TS-H2C-OLPNet framework achieve SOTA Precision, Recall, and F1-score performance. Results ensure that integrating optimized label propagation and hybrid hierarchical classification can offer a robust solution for large-scale retrieval using complex LULC data.
Labeling of the connected components is the key operation of the target recognition and segmentation in remotesensing *** conventional connected-component labeling(CCL) algorithms for ordinary optical images are cons...
详细信息
Labeling of the connected components is the key operation of the target recognition and segmentation in remotesensing *** conventional connected-component labeling(CCL) algorithms for ordinary optical images are considered time-consuming in processing the remotesensingimages because of the larger size.A dynamic run-length based CCL algorithm(Dy RLC) is proposed in this paper for the large size,big granularity sparse remotesensingimage,such as space debris images and ship *** addition,the equivalence matrix method is proposed to help design the pre-processing method to accelerate the equivalence labels *** result shows our algorithm outperforms 22.86% on execution time than the other algorithms in space debris image *** proposed algorithm also can be implemented on the field programming logical array(FPGA) to enable the realization of the real-time processing on-board.
In order to realize the accurate recognition of landslides in remotesensingimages, an improved DeepLabv3+ landslide extraction model is proposed in this paper.(1) Hybrid Module and Attention Module based CSPNet (HA-...
详细信息
Convolutional neural networks (CNNs) are the mainstream model for extracting rich features in deep learning-driven studies on cloud detection for remotesensingimages. However, due to the limitation of receptive fiel...
详细信息
暂无评论