In surveillance video, target tracking is an important part. Based on imageprocessing technology, this paper studies a real-time and effective method to collect and recognize camera motion information. Firstly, the i...
In surveillance video, target tracking is an important part. Based on imageprocessing technology, this paper studies a real-time and effective method to collect and recognize camera motion information. Firstly, the influence of visual dead angle and illumination on recognition is analyzed. Secondly, according to the characteristic of background light intensity, the corresponding algorithm is designed to realize the positioning and tracking control strategy of the target and surrounding environment scenery. Finally, the correctness of the method is verified by MATLAB simulation software, so as to obtain a better and scalable scheme, which is more economical and feasible after the occlusion rate is minimized.
This demo paper gives a real-time learned image codec on FPGA. By using Xilinx VCU128, the proposed system reaches 720P@30fps codec, which is 7.76x faster than prior work.
ISBN:
(纸本)9781665475938
This demo paper gives a real-time learned image codec on FPGA. By using Xilinx VCU128, the proposed system reaches 720P@30fps codec, which is 7.76x faster than prior work.
Plant infection/disease is one of the ongoing challenges for farmers, which imposes a threat on their income and food security. Detecting infection in plants or crops is an onerous task because the analysis of each cr...
详细信息
ISBN:
(纸本)9781665428644
Plant infection/disease is one of the ongoing challenges for farmers, which imposes a threat on their income and food security. Detecting infection in plants or crops is an onerous task because the analysis of each crop in large fields takes too much time, effort, work force and expertise. In this paper, we have proposed AI-based detection of pest infected crop and leaf. The proposed method is helpful in analyzing crops in a time-efficient manner and gives more accurate results. The classification of the crops is done on the basis of their images. A publicly available dataset is used to analyze the proposed methodology. imageprocessing methods are used in order to analyze the crops, further convolutional neural networks are applied to differentiate the healthy crops from the ones that are infected from some disease and also show some visual remarks.
The leading cause of visual impairment after cataract, is glaucoma and the only way to combat it is to detect it early. It is imperative to develop a system that can work effectively without a lot of equipment, qualif...
The leading cause of visual impairment after cataract, is glaucoma and the only way to combat it is to detect it early. It is imperative to develop a system that can work effectively without a lot of equipment, qualified medical personnel, and takes less time in order to address this fundamental issue. A Computer-Aided Diagnosis (CAD) system, which employs different algorithms for medical imageprocessing and analysis, can assist in achieving this. One of the ways to diagnose glaucoma is to calculate Optic Cup to Optic Disc ratio (CDR) and this can be done with the help of CAD algorithms. In medical imageprocessing the primary focus is on image segmentationand its classification in order to obtain a result. In this paper, the exploration the best-known CNN model, U-Net for image segmentation of Optic Disc and Optic Cup from a fundus image and Logistic Regression, a classification model to determine a relationship between these two terms rather than previously used CDR formulas.
Exploiting the infrared area of the spectrum for classification problems is getting increasingly popular, because many materials have characteristic absorption bands in this area. However, sensors in the short wave in...
Exploiting the infrared area of the spectrum for classification problems is getting increasingly popular, because many materials have characteristic absorption bands in this area. However, sensors in the short wave infrared (SWIR) area and even higher wavelengths have a very low spatial resolution in comparison to classical cameras that operate in the visible wavelength area. Thus, in this paper an upsampling method for SWIR images guided by a visible image is presented. For that, the proposed guided upsampling network (GUNet) uses a graph-regularized optimization problem based on learned affinities is presented. The evaluation is based on a novel synthetic near-field visible-SWIR stereo database. Different guided upsampling methods are evaluated, which shows an improvement of nearly 1 dB on this database for the proposed upsampling method in comparison to the second best guided upsampling network. Furthermore, a visual example of an upsampled SWIR image of a real-world scene is depicted for showing real-world applicability.
image captioning is an essential task in artificial intelligence that predicts the description of a given input image. In recent years, both computer vision and natural language processing witnessed huge advancement m...
详细信息
This paper presents an overview and analysis of numerous research projects on image fusion methods, with a particular emphasis on deep learning-based methods. The research analyses the inadequacies of current fusion m...
详细信息
Utilizing just noticeable difference(JND) thresholds to describe distortion visibility enables efficient transmission and storage of point clouds while maintaining high quality. Therefore, describing the relationship ...
详细信息
ISBN:
(数字)9798331529543
ISBN:
(纸本)9798331529550
Utilizing just noticeable difference(JND) thresholds to describe distortion visibility enables efficient transmission and storage of point clouds while maintaining high quality. Therefore, describing the relationship between point cloud JND and attribute quantization parameters (QP) is essential. This relationship is ,however, not clear regarding point cloud compression so far. We investigate the correlation between JND and attribute QP using Geometry-based Point Cloud Compression(G-PCC), which is the latest standard for point cloud compression. Moreover, we provide the corresponding datasets in this work. Using the K-means clustering algorithm, we partition the collected raw data into intervals and identify the peaks closest to the means in each sub-interval as the desired JND points. For human-content point clouds, although the number of JND varies, the QP values corresponding to different levels of JND show a regular distribution. The study focuses on the QP corresponding to the starting JND point(JND S ) and the ending JND point (JND E ).The findings indicate that the critical points for human-content point clouds are relatively consistent, with JND S mainly concentrated at QP = 27 and JND E at QP = 45. Conversely, perception in object-content point clouds is more influenced by their content, and the distribution of JND is not as uniform as in human-content point clouds. The experimental datasets and research results are accessible at the following link: https://***/ZhangChen2022/JND-attribute-QP-on-G-PCC.
In many scenarios, sensors installed in USV scan only detect either the above-water or underwater part of ob-stacles, which leads to errors in obstacle avoidance. In this paper, the obstacle information detected by th...
详细信息
ISBN:
(数字)9798350366556
ISBN:
(纸本)9798350366563
In many scenarios, sensors installed in USV scan only detect either the above-water or underwater part of ob-stacles, which leads to errors in obstacle avoidance. In this paper, the obstacle information detected by the sonar and visual camera is used to perform cross-domain data fusion. In low-speed operation mode, the prior binding information (surface and underwater) such as coordinate and direction will be used to match the collected image information (surface and underwater) until the desired matching level is achieved, directly determining the target orientation. When the coordinate matching level falls below the threshold or when the target ID information is missing, data processing module including cluster analysis, confidence filtering, and information entropy assessment will be employed to allocate weights to heterogeneous information, such as the distance and direction of the obstacle. In the fusion model, the Monte Carlo algorithm is applied to calculate the nonsynchronous error between the obstacles detected by the sonar and the visual sensor. Corresponding angle calculations are performed, and the information such as the distance and direction is converted into latitude and longitude coordinates. Compared to the traditional approach of simply merging surface or underwater obstacle information, we have integrated the information of the same obstacle from both above and below the water surface and enhance the reliability and accuracy of decision-making for unmanned surface vehicles.
In recent years, Transformers have achieved significant success in image fusion. These methods utilize self-attention mechanism across different spatial or channel dimensions and have demonstrated impressive performan...
详细信息
ISBN:
(数字)9798350368741
ISBN:
(纸本)9798350368758
In recent years, Transformers have achieved significant success in image fusion. These methods utilize self-attention mechanism across different spatial or channel dimensions and have demonstrated impressive performance. However, existing methods only optimize along a single dimension and struggle to simultaneously capture the complex dependencies between spatial and channel dimensions. To address this problem, we propose a novel multi-dimensional adaptive interaction transformer network, named as MAITFuse, to enhance the multilevel information expression and detail retention capabilities of images. We design a Multi-Dimensional Feature Extraction (MDFE) module to extract features across spatial and channel dimensions in parallel, and introduce a novel weighted cross-attention fusion method to integrate multi-dimensional information effectively. Experimental results show that, compared to existing fusion methods, our proposed method achieves superior fusion performance across various datasets.
暂无评论