Quality assessment of digital images plays an important role in modeling, implementation and optimization of image and video processing applications. One of the most popular methods in image quality assessment (IQA) i...
详细信息
ISBN:
(数字)9781510662117
ISBN:
(纸本)9781510662100;9781510662117
Quality assessment of digital images plays an important role in modeling, implementation and optimization of image and video processing applications. One of the most popular methods in image quality assessment (IQA) is feature based IQA techniques. These feature based image quality assessment (IQA) techniques, which consist of feature extraction and feature pooling phases, extracts features from the images in order to generate objective scores. Various hand-crafted features have been used in the feature extraction phase of the feature based IQA methods. In this work, instead of implementing a hand-crafted feature extraction scheme, automatic feature extraction is utilized by using a pre-trained deep neural network (DNN) inference structure. Feature pooling, which provides mapping between the proposed features and the subjective scores, is carried out by utilizing a fully-connected layer at the end of the network architecture. Experimental results show that the proposed technique obtains promising results for the IQA problem by making use of the generalization capability of deep learning architectures.
Compared with traditional computational fluid dynamics methods, the lattice Boltzmann method (LBM) has the advantages of simple program structure, adaptability to complex boundaries, and easy parallel computation. How...
详细信息
Compared with traditional computational fluid dynamics methods, the lattice Boltzmann method (LBM) has the advantages of simple program structure, adaptability to complex boundaries, and easy parallel computation. However, since LBM is an explicit algorithm, there are many iterations in the computation process, which leads to an increase in computation time. In this paper, we improve LBM based on deep learning by combining a convolutional neural network (CNN) and a gated recurrent unit neural network (GRU). Based on previous test data, the CNN module extracts spatial features during the computation, while the GRU processes the corresponding temporal features. Compared with the conventional LBM, this method can significantly reduce the computation time and improve the computational efficiency with guaranteed low Reynolds numbers of 1000 and 2000. At the high Reynolds number of 4000, the prediction error of the proposed method is increasing but still has a better performance. In order to verify the effectiveness and accuracy of the proposed algorithm, an eddying model widely used in the computational fluid field is developed. The proposed method not only has impressive results but also deals with non-stationary processes and steady-state problems.
This paper introduces an unrolled Expectation Maximization (EM) algorithm for sparse image reconstruction from radio interferometric measurements in the presence of a compound Gaussian distribution noise. Traditional ...
详细信息
ISBN:
(纸本)9789464593617;9798331519773
This paper introduces an unrolled Expectation Maximization (EM) algorithm for sparse image reconstruction from radio interferometric measurements in the presence of a compound Gaussian distribution noise. Traditional model-based reconstruction methods, rooted in inference and optimization fields, provide an initial foundation with theoretical guarantees, but their performance is highly linked to the model accuracy and the choice of hyperparameter values. The popularity of supervised machine learning rose over the last decade, yet they faced hurdles related to interpretability and theoretical foundations. The emergence of unrolled algorithms addresses these limitations by combining the strengths of both approaches. We specifically focus on unrolling a regularized EM algorithm as a feedforward neural network with a residual connection. Experimental results showcase improvements over the iterative EM version.
Compared with natural images, geospatial images cover larger area and have more complex image contents. There are few algorithms for generating controllable geospatial images, and their results are of low quality. In ...
详细信息
Compared with natural images, geospatial images cover larger area and have more complex image contents. There are few algorithms for generating controllable geospatial images, and their results are of low quality. In response to this problem, this paper proposes Geospatial Style Generative Adversarial Network to generate controllable and high-quality geospatial images. Current conditional generators suffer the mode collapse problem in geospatial field. The problem is addressed via a modified mode seeking regularization term with contrastive learning theory. Besides, the discriminator network architecture is modified to process global feature information and texture information of geospatial images. Feature loss in the generator is introduced to stabilize the training process and improve generated image quality. Comprehensive experiments are conducted on UC Merced Land Use Dataset, NWPU-RESISC45 Dataset, and AID Dataset to evaluate all compared methods. Experiment results show our method outperforms state-of-the-art models. Our method not only generates high-quality and controllable geospatial images, but also enhances the discriminator to learn better representations.
Pneumonia is a common and sometimes fatal lung infection that continues to be a major global health concern. The prediction of pneumonia has become a crucial factor in saving people's lives and improving their qua...
详细信息
Pneumonia is a common and sometimes fatal lung infection that continues to be a major global health concern. The prediction of pneumonia has become a crucial factor in saving people's lives and improving their quality of life. For this purpose, traditional clinical procedures are considered time-consuming. In addition, researchers have used various algorithms to forecast pneumonia due to advances in imageprocessing techniques. However, these algorithms have proven ineffective in terms of feature extraction, which negatively impacts prediction rates. This research aims to predict pneumonia in people worldwide and address the problem of low accuracy. This work introduces a novel method for pneumonia prediction using a deep CNN (Deep Convolutional neural Network) and an InceptionV3 model for feature extraction. Additionally, it introduces an entropy-normalized Neighbourhood Component Analysis (NCA) technique, complemented by Ensemble-Modified Classifiers (EMC) with Naive Bayes, XGBoost, and Random Forest for classification to enhance predictive accuracy. Accurate pneumonia diagnosis is crucial for patient care, but misdiagnoses and delays in diagnosis are not uncommon. This research establishes a robust framework for pneumonia prediction based on deep learning, capable of identifying both normal and atypical pneumonia patterns in medical images. To enhance feature extraction and improve model generalization, the proposed approach combines entropy normalization techniques. This method includes an NCA-based reduction in dimensionality, resulting in more efficient and discriminative feature representations. Furthermore, an ensemble-modified classifier is introduced to refine predictions and improve the model's ability to differentiate between pneumonia and non-pneumonia cases. Experimental results demonstrate that the proposed model surpasses existing methods in terms of accuracy, sensitivity, and specificity. The effectiveness of the proposed system has been confirmed b
While recent years have witnessed a dramatic upsurge of exploiting deep neural networks toward solving image denoising,existing methods mostly rely on simple noise assumptions,such as additive white Gaussian noise(AWG...
详细信息
While recent years have witnessed a dramatic upsurge of exploiting deep neural networks toward solving image denoising,existing methods mostly rely on simple noise assumptions,such as additive white Gaussian noise(AWGN),JPEG compression noise and camera sensor noise,and a general-purpose blind denoising method for real images remains *** this paper,we attempt to solve this problem from the perspective of network architecture design and training data ***,for the network architecture design,we propose a swin-conv block to incorporate the local modeling ability of residual convolutional layer and non-local modeling ability of swin transformer block,and then plug it as the main building block into the widely-used image-to-image translation UNet *** the training data synthesis,we design a practical noise degradation model which takes into consideration different kinds of noise(including Gaussian,Poisson,speckle,JPEG compression,and processed camera sensor noises)and resizing,and also involves a random shuffle strategy and a double degradation *** experiments on AGWN removal and real image denoising demonstrate that the new network architecture design achieves state-of-the-art performance and the new degradation model can help to significantly improve the *** believe our work can provide useful insights into current denoising *** source code is available at https://***/cszn/SCUNet.
Gait recognition is a well-known biometric identification technology and is widely employed in different fields. Due to the advantages of deep learning, such as self-learning capability, high accuracy and excellent ge...
详细信息
Gait recognition is a well-known biometric identification technology and is widely employed in different fields. Due to the advantages of deep learning, such as self-learning capability, high accuracy and excellent generalization ability, various deep network algorithms have been applied in biometric recognition. Numerous studies have been conducted in this area;however, they may not always yield the expected outcomes owing to the issue of data imbalance in clinical and healthcare industries. To overcome this problem, deep multi-convolutional stacked capsule network fostered human gait recognition from enhanced gait energy image (HGR-DMCSCN) is proposed in this manuscript. Initially, the input images are taken from CASIA B and OU-ISIR datasets. Then the input images are given to preprocessing segment to enhance the superiority of the images based upon contrast-limited adaptive histogram equalization filtering (CLAHEF). Then preprocessed image is given to classification process using deep multi-convolutional stacked capsule network (DMCSCN) that is utilized for human gait detection under various conditions, like normal walking, carrying a bag and wearing a cloth. The proposed HGR-DMCSCN approach is executed in python and its performance is examined under performance metrics, such as F-Score, accuracy, RoC and computational time. Finally, the proposed approach attains 28.70%, 11.87% and 14.79% higher accuracy for CASIA B compared with existing methods.
This study presents the RBP-CNN model, a convolutional neural network specifically designed for the precise classification of brain tumors in medical imaging. Conventional methods often encounter difficulties in extra...
详细信息
This study presents the RBP-CNN model, a convolutional neural network specifically designed for the precise classification of brain tumors in medical imaging. Conventional methods often encounter difficulties in extracting image noise and texture features, which has led to the incorporation of regional binary patterns (RBP) and Gray Standard Normalization (GSN) preprocessing techniques in CNN. The research addresses fundamental inquiries regarding the impact of the model on accuracy, false classifications, and efficiency. The novelty of RBPCNN lies in its distinctive approach to extracting texture features, which involves optimizing pixel values through GSN preprocessing and generating regional binary patterns based on integral images. The objective of this research is to bridge a critical gap by providing a more accurate and efficient model for classifying brain tumors. The key findings reveal the exceptional performance of RBP-CNN, achieving a classification accuracy of 96% with a reduced false classification ratio of 7% across a dataset of 3000 samples. Comparative analyses position RBP-CNN as superior to alternative models in terms of accuracy, false classification rates, and efficiency. The structural insights and hyperparameter values of the model, as well as its application to the FigShare dataset, demonstrate its robustness and scalability. RBP-CNN emerges as an innovative and effective solution, advancing the field of medical image categorization. The findings of this study contribute a novel methodology, paving the way for future exploration in hyperspectral image applications and positioning RBP-CNN as a potential state-ofthe-art tool for medical image analysis.
With the advancement of deep learning techniques, the classification of remote sensing data using artificial neural networks has emerged as a prominent research area. Despite this progress, the emulation of brain stru...
详细信息
The "Residual-to-Residual DNN series for high-Dynamic range imaging" (R2D2) approach was recently introduced for Radio-Interferometric (RI) imaging in astronomy. R2D2's reconstruction is formed as a seri...
详细信息
ISBN:
(纸本)9789464593617;9798331519773
The "Residual-to-Residual DNN series for high-Dynamic range imaging" (R2D2) approach was recently introduced for Radio-Interferometric (RI) imaging in astronomy. R2D2's reconstruction is formed as a series of residual images, iteratively estimated as outputs of Deep neural Networks (DNNs) taking the previous iteration's image estimate and associated data residual as inputs. In this work, we investigate the robustness of the R2D2 image estimation process, by studying the uncertainty associated with its series of learned models. Adopting an ensemble averaging approach, multiple series can be trained, arising from different random DNN initializations of the training process at each iteration. The resulting multiple R2D2 instances can also be leveraged to generate "R2D2 samples", from which empirical mean and standard deviation endow the algorithm with a joint estimation and uncertainty quantification functionality. Focusing on RI imaging, and adopting a telescope-specific approach, multiple R2D2 instances were trained to encompass the most general observation setting of the Very Large Array (VLA). Simulations and real-data experiments confirm that: (i) R2D2's image estimation capability is superior to that of the state-of-the-art algorithms;(ii) its ultra-fast reconstruction capability (arising from series with only few DNNs) makes the computation of multiple reconstruction samples and of uncertainty maps practical even at large image dimension;(iii) it is characterized by a very low model uncertainty.
暂无评论