With the frequent occurrence of geological disasters, landslide identification has become an important research problem. In recent years, with the application and research of deep learning in the fields of computer vi...
With the frequent occurrence of geological disasters, landslide identification has become an important research problem. In recent years, with the application and research of deep learning in the fields of computer vision and remotesensingimage analysis, deep learning has become a popular method for landslide identification research because of its excellent feature extraction and patternrecognition capabilities. In this paper, we add new bands reflecting vegetation and water of landslide factors to the original bands of remotesensingimage samples by linear combination of bands, and conduct two sets of comparison experiments on Unet and Swin-Unet to verify that adding band features can help improve landslide identification accuracy. The experimental results show that the experimental group with additional band features has a numerical improvement in landslide identification accuracy, with F1-score as the evaluation index, the Unet experimental group has a 2.29% improvement and the Swin-Unet experimental group has a 1.78% improvement. The results of this paper have implications for the subsequent application of bands combination of landslide factor features in remotesensing landslide identification driven by deep learning methods.
Earth Observation imagery can capture rare and unusual events, such as disasters and major landscape changes, whose visual appearance contrasts with the usual observations. Deep models trained on common remotesensing...
详细信息
ISBN:
(数字)9798350365474
ISBN:
(纸本)9798350365481
Earth Observation imagery can capture rare and unusual events, such as disasters and major landscape changes, whose visual appearance contrasts with the usual observations. Deep models trained on common remotesensing data will output drastically different features for these out-of-distribution samples, compared to those closer to their training dataset. Detecting them could therefore help anticipate changes in the observations, either geographical or environmental. In this work, we show that the reconstruction error of diffusion models can effectively serve as unsupervised out-of-distribution detectors for remotesensingimages., using as a plausibility score. Moreover, we introduce ODEED, a novel reconstruction-based scorer using the probability-flow ODE of diffusion models. We validate it experimentally on SpaceNet 8 with various scenarios, such as classical OOD detection with geographical shift and near-OOD setups: pre/post-flood and non-flooded/flooded imagerecognition. We show that our ODEED scorer significantly outperforms other diffusion-based and discriminative baselines on the more challenging near-OOD scenarios of flood image detection, where OOD images are close to the distribution tail. We aim to pave the way towards better use of generative models for anomaly detection in remotesensing.
Although deep learning has revolutionized remotesensing (RS) image scene classification, current deep learningbased approaches highly depend on the massive supervision of predetermined scene categories and have disap...
详细信息
Although deep learning has revolutionized remotesensing (RS) image scene classification, current deep learningbased approaches highly depend on the massive supervision of predetermined scene categories and have disappointingly poor performance on new categories that go beyond predetermined scene categories. In reality, the classification task often has to be extended along with the emergence of new applications that inevitably involve new categories of RS image scenes, so how to make the deep learning model own the inference ability to recognize the RS image scenes from unseen categories, which do not overlap the predetermined scene categories in the training stage, becomes incredibly important. By fully exploiting the RS domain characteristics, this paper constructs a new remotesensing knowledge graph (RSKG) from scratch to support the inference recognition of unseen RS image scenes. To improve the semantic representation ability of RS-oriented scene categories, this paper proposes to generate a Semantic Representation of scene categories by representation learning of RSKG (SR-RSKG). To pursue robust cross-modal matching between visual features and semantic representations, this paper proposes a novel deep alignment network (DAN) with a series of well-designed optimization constraints, which can simultaneously address zero-shot and generalized zero-shot RS image scene classification. Extensive experiments on one merged RS image scene dataset, which is the integration of multiple publicly open datasets, show that the recommended SR-RSKG obviously outperforms the traditional knowledge types (e.g., natural language processing models and manually annotated attribute vectors), and our proposed DAN shows better performance compared with the state-of-the-art methods under both the zero-shot and generalized zero-shot RS image scene classification settings. The constructed RSKG will be made publicly available along with this paper (https://***/kdy2021/SR-RSKG).
remotesensingimage fusion, i.e., fusing remotesensingimages from different sensors or different time into a comprehensive image, can integrate image information for all kinds of image tasks, such as object detecti...
remotesensingimage fusion, i.e., fusing remotesensingimages from different sensors or different time into a comprehensive image, can integrate image information for all kinds of image tasks, such as object detection, patternrecognition, and so on. In this paper, we focus on the image fusion of optical image and synthetic aperture radar (SAR) image, where the traditional methods, such as sparse representation and image decomposition, usually fail to improve the image quality and enhance object features' level. Inspired by the powerful generation model named diffusion denoising probabilistic model (DDPM), we propose a novel method based on diffusion posterior sampling for image fusion of optical image and SAR image. Starting from the variational model of image fusion and the total variation constraint, we approximate the posterior sampling process of image fusion by a closed-form analytic solution. After that, DDPM can be used to generate the fused image with high image quality and enhanced object features. The feasibility and the superiority of the proposed method is validated in numerical experiments.
Semantic segmentation is in-demand in High Resolution remotesensing (HRRS) imageprocessing. Unlike natural images, HRRS images usually provide channels such as Near Infrared (NIR) in addition to RGB channels. Howeve...
详细信息
ISBN:
(纸本)9783030880071;9783030880064
Semantic segmentation is in-demand in High Resolution remotesensing (HRRS) imageprocessing. Unlike natural images, HRRS images usually provide channels such as Near Infrared (NIR) in addition to RGB channels. However, in order to make use of the pre-trained model, the current semantic segmentation methods in remotesensing field usually only use the RGB channel and discard the information of other channels. In this paper, to make full use of the HRRS image information, a dual-stream fusion network is proposed to fuse the information of different channel combinations through a Feature Pyramid Network (FPN), then a Stage Pyramid Pooling (SPP) module is used to integrate the features of different scales and produce the final segmentation results. Experiments on the RSCUP competition dataset show that the proposed approach can effectively improve the segmentation performance.
The land-covers within an observed remotesensing scene are usually of different scales;therefore, the ensemble of multi-scale information is a commonly used strategy to achieve more accurate scene inter-pretation;how...
详细信息
The land-covers within an observed remotesensing scene are usually of different scales;therefore, the ensemble of multi-scale information is a commonly used strategy to achieve more accurate scene inter-pretation;however, this process suffers from being time-consuming. In terms of this issue, this paper proposes a scale distillation network to explore the possibility that single-scale classification network can achieve the same (or even better) classification performance compared with multi-scale one. The pro-posed scale distillation network consists of a cumbersome multi-scale teacher network and a lightweight single-scale student network. The former is trained for multi-scale information learning, and the latter improves the classification accuracy by accepting the knowledge from the multi-scale teacher network and its true label. The experimental results show the advantages of scale distillation on hyperspectral image classification. The single-scale student network can even achieve higher evaluation accuracy than the multi-scale teacher network. In addition, a faithful explainable scale network is designed to visually explain the trained scale distillation network. The traditional deep neural network is a black-box and lacks interpretability. The explanation of the trained network can explore more hidden information from the predictions. We visually explain the prediction results of scale distillation network, and the results show that the explainable scale network can more precisely analyze the relationship between the learned scale features and the land-cover categories. Moreover, the possible application of the explainable scale network on classification is further discussed in this study. (c) 2021 Elsevier Ltd. All rights reserved.
Transformer architecture has attained noteworthy performance achievements in recent image super-resolution research. However, current transformer-based methods still expose limitations in fully harnessing domain-speci...
详细信息
Segment-Anything model (SAM) is a foundation segmentation model published in April 2023. Trained on an unprecedented 11 million annotated images, the model can generate segmented masks bearing clear-cut contours by in...
详细信息
Fire detection based on computer vision technology can avoid many flaws in conventional methods. However, existing methods fail to achieve a good trade-off in accuracy, model size, speed, and cost. This paper presents...
详细信息
Fire detection based on computer vision technology can avoid many flaws in conventional methods. However, existing methods fail to achieve a good trade-off in accuracy, model size, speed, and cost. This paper presents a high-performance forest fire recognition algorithm to solve the current problems in forest fire monitoring. Firstly, visual saliency areas in motion images are extracted to improve detection efficiency. Secondly, transfer learning techniques are employed to improve the generalization performance of the constructed deep learning classification model. Finally, fire detection is realized based on C++ deployment algorithms Compared with the existing forest fire detection methods, the proposed method has higher classification accuracy and speed, with a more comprehensive application range and lower cost. The performance of our method can meet the accuracy and speed requirements of real-time fire detection, and it can be deployed and practiced on multiple platforms.
Digital imageprocessing technology has gone through rapid development and is extensively applied in daily life and production, with the rapid development of modern information technology. It plays an inestimable role...
详细信息
暂无评论