Machine learning applications in remotesensing often require a labour-intensive feature engineering step, if only a small number of samples is available and transfer learning is not applicable. Here, we are introduci...
详细信息
ISBN:
(数字)9789819703760
ISBN:
(纸本)9789819703753;9789819703760
Machine learning applications in remotesensing often require a labour-intensive feature engineering step, if only a small number of samples is available and transfer learning is not applicable. Here, we are introducing the concept of Spatial Variation Sequences, which allows to apply methodologies from automated time-series feature engineering to remotesensing applications of static images. The presented example application detects swimming pools from four-channel satellite images with an F-1-score of 0.95, by generating spatial variation sequences from a modified swimming pool index. The automated feature engineering approach reduced the dimensionality of the classification problem by 99.7%. A more traditional approach using transfer learning on pretrained Convolutional Neural Networks (CNN) was evaluated in parallel for comparison. The CNN approach boasted a higher performance of F-1-score of 0.98 but required the use of pre-trained weights. The comparable performance of the FE and CNN approach demonstrates that time-series feature extraction is a valuable alternative to traditional remotesensing methods in the presence of data scarcity or the need of significant dimensionality reduction.
License plate detection is an important task in Intelligent Transportation Systems (ITS) and has a wide range of applications in vehicle management, traffic control, and public safety. In order to improve the accuracy...
详细信息
License plate detection is an important task in Intelligent Transportation Systems (ITS) and has a wide range of applications in vehicle management, traffic control, and public safety. In order to improve the accuracy and speed of mobile recognition, an improved lightweight YOLOv5s model is proposed for license plate detection. First, an improved Stemblock network is used to replace the original Focus layer in the network, which ensures strong feature expression capability and reduces a large number of parameters to lower the computational complexity;then, an improved lightweight network, ShuffleNetv2, is used to replace the backbone network of the YOLOv5s, which makes the model lighter and ensures the detection accuracy at the same time. Then, a feature enhancement module is designed to reduce the information loss caused by the rearrangement of the backbone network channels, which facilitates the information interaction in the feature fusion process;finally, the low-, medium- and high-level features in the Shufflenetv2 network structure are fused to form the final high-level output features. Experimental results on the CCPD dataset show that compared to other methods this paper obtains better performance and faster speed in the license plate detection task, in which the average precision mean value reaches 96.6%, and can achieve a detection speed of 43.86 frame/s, and the parameter volume is reduced to 5.07 M. 1. Improved Stemblock structure to reduce the number of parameters.2. Improve Shufflunetv2 as the backbone network and reduce the number of parameters.3. The attention mechanism is added to the network to improve the detection ***
Deep learning-based image super-resolution (SR) technology has gained extensive attention in the remotesensing community, which aims to reconstruct the abundant details of target images. However, the practical applic...
详细信息
Urban vegetation recognition based on remotesensing data is highly affected by the complexities of urban areas due to the existence of various kinds of objects and their relations. Spectral similarities between tree ...
详细信息
Urban vegetation recognition based on remotesensing data is highly affected by the complexities of urban areas due to the existence of various kinds of objects and their relations. Spectral similarities between tree canopies and types of grass lands, spatial adjacency between houses and tall trees, shadow and occluded areas make some difficulties for recognition of the plant species. In this research, the capabilities of multi-agent systems are utilized for feature fusion of hyperspectral imagery and lidar data for improving the vegetation recognition results in urban areas. The proposed algorithm has two main steps composed of generating a knowledge base containing spectral and height features extracted from input hyperspectral and Lidar data, respectively, and performing the hierarchical classification process to generate vegetation classification map based on parallel processing by object recognition agents. Evaluation of the capabilities of the proposed methodology is performed on hyperspectral and lidar DSM over Houston University and its surrounding areas. According to the obtained results, fusion of hyperspectral and Lidar DSM with the capabilities of multi-agent processing can improve the overall accuracy of vegetation recognition results for about 15.53% and 6.58% comparing with performing multi-agent and maximum likelihood classifier only on hyperspectral image, respectively.
Sea observation through remotesensing technologies plays an essential role in understanding the health status of the marine coastal environment, its fauna species and their future behavior. Accurate knowledge of the ...
详细信息
Sea observation through remotesensing technologies plays an essential role in understanding the health status of the marine coastal environment, its fauna species and their future behavior. Accurate knowledge of the marine habitat and the factors affecting faunal variations allows us to perform predictions and adopt proper decisions. This paper concerns the proposal of a classification system devoted to recognizing marine mesoscale events. These phenomena are studied and monitored by analyzing sea surface temperature imagery. Currently, the standard way to perform such analysis relies on experts manually visualizing, analyzing, and tagging large imagery datasets. Nowadays, the availability of remotesensing data has increased so much that it is desirable to replace the labor-intensive, time-consuming, and subjective manual interpretation with automated analysis tools. The results presented in this work have been obtained by applying the proposed approach to images captured over the southwestern region of the Iberian Peninsula.
With the rapid development of artificial neural network (ANN), the field of synthetic aperture radar (SAR) target recognition has witnessed significant progress. However, due to the poor interpretability and ease of b...
详细信息
Weakly supervised image semantic segmentation has become the most popular method in recent years because of its low cost and has been widely used in medical image segmentation, automatic driving, remotesensingimage ...
详细信息
ISBN:
(纸本)9789819784929;9789819784936
Weakly supervised image semantic segmentation has become the most popular method in recent years because of its low cost and has been widely used in medical image segmentation, automatic driving, remotesensingimage analysis and other fields. However, the current weakly supervised semantic segmentation based on transformer has some problems, such as focusing on the whole, ignoring local details and confusing different categories. To solve these problems, we come up with a token-guided single stage weakly supervised image semantic segmentation algorithm. First of all, in order to solve the problem of insufficient attention to details, we proposed an optimization clipping method, which realized the selection of uncertain regions as much as possible and the fine marking of uncertain regions. Then, the single-class token to multiple class tokens method is purposed to obtain multiple class tokens for fine guidance. In particular, we designed a multiple class tokens guide method to complete the function of classifying uncertain regions and correctly activating them. The quantitative and qualitative results of the public dataset PASCAL VOC 2012 validate the effectiveness of the method.
The rotation of targets in high-resolution synthetic aperture radar (SAR) images results in complex and nonlinear micro-Doppler modulation of SAR returns, exhibiting distinct and diverse imaging characteristics under ...
详细信息
The rotation of targets in high-resolution synthetic aperture radar (SAR) images results in complex and nonlinear micro-Doppler modulation of SAR returns, exhibiting distinct and diverse imaging characteristics under different conditions. This poses challenges to image interpretation, target detection and recognition. This article establishes signal models for single and multiple scattering of rotating targets in high-resolution spaceborne SAR and, in conjunction with the processing procedures of range Doppler algorithm (RDA), derives the azimuth bandwidth and the conditions for azimuth frequency aliasing caused by rotation. Quantitative analysis of the errors introduced by target rotation in various signal domains is performed, taking into account the stationary phase point offset, the rotation-induced range migration correction error, and the change of azimuth phase in detail. In addition, representative signatures of these influences in SAR images are simulated and interpreted and more accurate theoretical models of imaging results are derived. The processing and analysis of both simulation and real measured data validate the correctness of the theoretical derivation in this study and provide support for the interpretation and information extraction of high-resolution SAR rotating targets.
We present a simple and efficient method to leverage emerging text-to-image generative models in creating large-scale synthetic supervision for the task of damage assessment from aerial images. While significant recen...
详细信息
ISBN:
(纸本)9798350365474
We present a simple and efficient method to leverage emerging text-to-image generative models in creating large-scale synthetic supervision for the task of damage assessment from aerial images. While significant recent advances have resulted in improved techniques for damage assessment using aerial or satellite imagery, they still suffer from poor robustness to domains where manual labeled data is unavailable, directly impacting post-disaster humanitarian assistance in such under-resourced geographies. Our contribution towards improving domain robustness in this scenario is two-fold. Firstly, we leverage the text-guided mask-based image editing capabilities of generative models and build an efficient and easily scalable pipeline to generate thousands of post-disaster images from low-resource domains. Secondly, we propose a simple two-stage training approach to train robust models while using manual supervision from different source domains along with the generated synthetic target domain data. We validate the strength of our proposed framework under cross-geography domain transfer setting from xBD and SKAI images in both single-source and multi-source settings, achieving significant improvements over a source-only baseline in each case.
In practical applications, remotesensing (RS) scene classification faces data shift problems, including novel class and data discrepancy problems. Due to these problems, it is difficult to obtain representative and d...
详细信息
暂无评论