Low-light image enhancement is a computer vision task that aims to improve the visual perceptual quality of images captured in poorly illuminated scenes. At present, deep learning-based low-light enhancement methods c...
详细信息
ISBN:
(纸本)9798350349405;9798350349399
Low-light image enhancement is a computer vision task that aims to improve the visual perceptual quality of images captured in poorly illuminated scenes. At present, deep learning-based low-light enhancement methods can obtain high-quality enhanced images. However, it does not consider the statistical characteristics of different regions, such as edge, structure, and texture. The uncertainty of image regions is not well characterized and utilized. To address this problem, we propose a novel UnCertainty-driven Cycle-Consistent Generative Adversarial Network (UTrCGAN) to improve the performance of low-light enhancement. UTrCGAN first decomposes the unpaired low/normal-light images into reflectance and illumination components based on the Retinex theory. Then a generative adversarial network guided by uncertainty constraint is proposed to enhance the illumination component, in which the quality of the enhanced image is further improved by the guidance of variance estimation. Experimental results on the widely-used LOL dataset show that UTrCGAN outperforms the state-of-the-art methods in terms of visual quality and quantitative metrics.
stochastic resonance has always been a research hotspot in the field of imageprocessing. The ability of the nervous system to detect signals is closely related to nonlinear and collective behavior in neural mediators...
详细信息
The symmetric positive definite (SPD) matrices play an important role in diverse domains, including computer vision and signal processing, due to their unique ability to capture the intrinsic structure of nonlinear da...
详细信息
ISBN:
(纸本)9789464593617;9798331519773
The symmetric positive definite (SPD) matrices play an important role in diverse domains, including computer vision and signal processing, due to their unique ability to capture the intrinsic structure of nonlinear data using Riemannian geometry. Despite their significance, a notable gap exists in the absence of statistical distributions capable of effectively characterizing the statistical properties within the SPD matrices space. This paper addresses this gap by introducing a novel Riemannian Generalized Gaussian distribution (RGGD). The primary aspect of this work includes presenting the precise expression for the probability density function (PDF) of the RGGD model, along with the parameter estimation method based on the maximum likelihood for this distribution. The second aspect of this work entails harnessing the second-order statistical information captured in the feature maps originating from the initial layers of deep convolutional neural networks (DCNNs) using the RGGD stochastic model within an image classification framework. The third aspect of this work includes also the comparison of the three-parameter RGGD model with its two-parameter predecessors, namely the Riemannian Gaussian distribution (RGD) and the Riemannian Laplacian distribution (RLD). Besides the mathematical foundations, the model's efficiency is validated through experiments conducted on the three well-known datasets, showcasing its effectiveness in capturing the underlying statistics of SPD matrices.
The stochastic nature of modern Monte Carlo (MC) rendering methods inevitably produces noise in rendered images for a practical number of samples per pixel. The problem of denoising these images has been widely studie...
详细信息
ISBN:
(纸本)9798400711312
The stochastic nature of modern Monte Carlo (MC) rendering methods inevitably produces noise in rendered images for a practical number of samples per pixel. The problem of denoising these images has been widely studied, with most recent methods relying on data-driven, pretrained neural networks. In contrast, in this paper we propose a statistical approach to the denoising problem, treating each pixel as a random variable and reasoning about its distribution. Considering a pixel of the noisy rendered image, we formulate fast pair-wise statistical tests-based on online estimators-to decide which of the nearby pixels to exclude from the denoising filter. We show that for symmetric pixel weights and normally distributed samples, the classical Welch t-test is optimal in terms of mean squared error. We then show how to extend this result to handle non-normal distributions, using more recent confidence-interval formulations in combination with the BoxCox transformation. Our results show that our statistical denoising approach matches the performance of state-of-the-art neural image denoising without having to resort to any computation-intensive pretraining. Furthermore, our approach easily generalizes to other quantities besides pixel intensity, which we demonstrate by showing additional applications to Russian roulette path termination and multiple importance sampling.
Drone-based geophysical surveying is an emerging measuring platform. High time-cost efficiency and flexibility to survey over inaccessible areas make drones attractive or sometimes the only feasible option to carry ge...
详细信息
Drone-based geophysical surveying is an emerging measuring platform. High time-cost efficiency and flexibility to survey over inaccessible areas make drones attractive or sometimes the only feasible option to carry geophysical measurements. This study presents a new drone-towed electromagnetic induction and magnetic gradient sensor system used for near-surface characterization and areal *** system uses various datasets to enhance processing and interpretation. The system includes;an electromagnetic induction instrument;magnetic sensors;GNSS-IMU system;photogrammetry;Lidar model data;and geoid model data. Robust data processing and stochastic inversion subsurface characterization for archaeological prospecting with drone-towed electromagnetic induction and magnetic gradient sensor systems. Robust statisticalmethods were used to process the *** conducted the fieldwork at one of the ancient Viking settlements in Denmark. The surveyed area was approximately 100x$\times$200 m. We then implemented and applied a one-dimensional laterally constrained non-linear stochastic inversion to image the subsurface electrical conductivity. The inversion results show a consistent conductive layer at 5-8 m depths, likely associated with the groundwater level. This conductive layer is disrupted under a prominent anomaly within a 2-4 m wide area. Our analysis showed that this conductivity disruption could be a flint mine extending 7 m deep. This anomaly also has a strong signature in magnetic gradient data.
Recent advancements in neural compression have surpassed traditional codecs in PSNR and MS-SSIM measurements. However, at low bit-rates, these methods can introduce visually displeasing artifacts, such as blurring, co...
详细信息
ISBN:
(纸本)9798331529543;9798331529550
Recent advancements in neural compression have surpassed traditional codecs in PSNR and MS-SSIM measurements. However, at low bit-rates, these methods can introduce visually displeasing artifacts, such as blurring, color shifting, and texture loss, thereby compromising perceptual quality of images. To address these issues, this study presents an enhanced neural compression method designed for optimal visual fidelity. We have trained our model with a sophisticated semantic ensemble loss, integrating Charbonnier loss, perceptual loss, style loss, and a non-binary adversarial loss, to enhance the perceptual quality of image reconstructions. Additionally, we have implemented a latent refinement process to generate content-aware latent codes. These codes adhere to bit-rate constraints, and prioritize bit allocation to regions of greater importance. Our empirical findings demonstrate that this approach significantly improves the statistical fidelity of neural image compression.
The Deep image Prior (DIP) technique has been successfully employed in Compressive Spectral Imaging (CSI) as a non-data-driven deep model approach. DIP methodology updates the deep network's weights by minimizing ...
详细信息
ISBN:
(纸本)9798350302615
The Deep image Prior (DIP) technique has been successfully employed in Compressive Spectral Imaging (CSI) as a non-data-driven deep model approach. DIP methodology updates the deep network's weights by minimizing a loss function that considers the difference between the measurements and the forward operator of the network's output. However, this method often yields local minima as all the measurements are evaluated at each iteration. This paper proposes a stochastic deep image prior (SDIP) approach, which stochastically trains DIP networks using random subsets of measurements from different CSI sensors in a CSI fusion (CSIF) setting, resulting in the improvement of the convergence through stochastic gradient descent optimization. The proposed SDIP method improves upon the deterministic DIP and requires less computational time since fewer forward operators are required per iteration. The SPID method provides comparable performance against the state-of-the-art CSIF techniques based on supervised data-driven and unsupervised methods, achieving up to 5 dB in the reconstruction.
Vector quantization (VQ) methods have been used in a wide range of applications for speech, image, and video data. While classic VQ methods often use expectation maximization, in this paper, we investigate the use of ...
详细信息
A new approach to adaptive traffic light control at an intersection is being developed by using automatic control methods while constantly monitoring local traffic flow. methods of providing it with the necessary data...
详细信息
The machine learning of lattice operators has three possible bottlenecks. From a statistical standpoint, it is necessary to design a constrained class of operators based on prior information with low bias, and low com...
详细信息
ISBN:
(纸本)9783031577925;9783031577932
The machine learning of lattice operators has three possible bottlenecks. From a statistical standpoint, it is necessary to design a constrained class of operators based on prior information with low bias, and low complexity relative to the sample size. From a computational perspective, there should be an efficient algorithm to minimize an empirical error over the class. From an understanding point of view, the properties of the learned operator need to be derived, so its behavior can be theoretically understood. The statistical bottleneck can be overcome due to the rich literature about the representation of lattice operators, but there is no general learning algorithm for them. In this paper, we discuss a learning paradigm in which, by overparametrizing a class via elements in a lattice, an algorithm for minimizing functions in a lattice is applied to learn. We present the stochastic lattice descent algorithm as a general algorithm to learn on constrained classes of operators as long as a lattice overparametrization of it is fixed, and we discuss previous works which are proves of concept. Moreover, if there are algorithms to compute the basis of an operator from its overparametrization, then its properties can be deduced and the understanding bottleneck is also overcome. This learning paradigm has three properties that modern methods based on neural networks lack: control, transparency and interpretability. Nowadays, there is an increasing demand for methods with these characteristics, and we believe that mathematical morphology is in a unique position to supply them. The lattice overparametrization paradigm could be a missing piece for it to achieve its full potential within modern machine learning.
暂无评论