Different spatial point process models and techniques have been developed in the past decades to facilitate the statistical analysis of spatial point patterns. However, in some cases the spatial point process methodol...
详细信息
Different spatial point process models and techniques have been developed in the past decades to facilitate the statistical analysis of spatial point patterns. However, in some cases the spatial point process methodology is scarce and no flexible models nor suitable statistical methods are available. For example, due to its complexity, the statistical analysis of spatial point patterns of several groups observed at a number of time instances has not been studied in-depth, and there are a few limited models and methods available for such data. In the present work, we provide a mathematical framework for coupling neural network methods with the statistical analysis of point patterns. In particular, we discuss an example of deep neural networks in the statistical analysis of highly multivariate spatial point patterns and provide a new strategy for building spatio-temporal point processes using variational autoencoder generative neural networks. (c) 2022 Elsevier B.V. All rights reserved.
The COVID-19 pandemic has highlighted the importance of non-pharmacological interventions (NPIs) for controlling epidemics of emerging infectious diseases. Despite their importance, NPIs have been monitored mainly thr...
详细信息
The COVID-19 pandemic has highlighted the importance of non-pharmacological interventions (NPIs) for controlling epidemics of emerging infectious diseases. Despite their importance, NPIs have been monitored mainly through the manual efforts of volunteers. This approach hinders measurement of the NPI effectiveness and development of evidence to guide their use to control the global pandemic. We present EpiTopics, a machine learning approach to support automation of NPI prediction and monitoring at both the document level and country level by mining the vast amount of unlabeled news reports on COVID-19. EpiTopics uses a 3-stage, transfer-learning algorithm to classify documents according to NPI categories, relying on topic modeling to support result interpretation. We identified 25 interpretable topics under 4 distinct and coherent COVID-related themes. Importantly, the use of these topics resulted in significant improvements over alternative automated methods in predicting the NPIs in labeled documents and in predicting country-level NPIs for 42 countries.
Recent years have seen a surge in interest in building deep learning-based fully data-driven models for weather prediction. Such deep learning models, if trained on observations can mitigate certain biases in current ...
详细信息
Recent years have seen a surge in interest in building deep learning-based fully data-driven models for weather prediction. Such deep learning models, if trained on observations can mitigate certain biases in current state-of-the-art weather models, some of which stem from inaccurate representation of subgrid-scale processes. However, these data-driven models, being over-parameterized, require a lot of training data which may not be available from reanalysis (observational data) products. Moreover, an accurate, noise-free, initial condition to start forecasting with a data-driven weather model is not available in realistic scenarios. Finally, deterministic data-driven forecasting models suffer from issues with long-term stability and unphysical climate drift, which makes these data-driven models unsuitable for computing climate statistics. Given these challenges, previous studies have tried to pre-train deep learning-based weather forecasting models on a large amount of imperfect long-term climate model simulations and then re-train them on available observational data. In this article, we propose a convolutional variational autoencoder (VAE)-based stochastic data-driven model that is pre-trained on an imperfect climate model simulation from a two-layer quasi-geostrophic flow and re-trained, using transfer learning, on a small number of noisy observations from a perfect simulation. This re-trained model then performs stochastic forecasting with a noisy initial condition sampled from the perfect simulation. We show that our ensemble-based stochastic data-driven model outperforms a baseline deterministic encoder-decoder-based convolutional model in terms of short-term skills, while remaining stable for long-term climate simulations yielding accurate climatology. Impact Statement A stochastic VAE-based data-driven model pre-trained on imperfect climate simulations and re-trained with transfer learning, on a limited number of observations, leads to accurate short-term w
In manufacturing industries, monitoring the complicated devices often necessitates automated methods that can leverage the multivariate time series data produced by the machines. However, analyzing this data can be ch...
详细信息
In manufacturing industries, monitoring the complicated devices often necessitates automated methods that can leverage the multivariate time series data produced by the machines. However, analyzing this data can be challenging due to varying noise levels in the data and possible nonlinear relations between the process variables, requiring appropriate tools to deal with such properties.
This thesis proposes a deep learning-based approach to detect anomalies and interpret their root causes from multivariate time series data, which can be applied in a near real-time setting. The proposed approach extends an existing model from the literature, which employs a variational autoencoder architecture and recurrent neural networks to capture both stochasticity and temporal relations of the data.
The anomaly detection and root cause interpretation performance of the proposed method is compared against five baseline algorithms previously proposed in the literature using real-world data collected from plastic injection molding machines and artificially generated multivariate time series data.
The results of this thesis show that the proposed method performs well on the evaluated multivariate time series datasets, mostly outperforming the baseline methods. Additionally, the approach had the best performance among the selected methods on providing root cause interpretation of the detected anomalies. The experiments conducted in this thesis suggest that deep learning-based algorithms are beneficial for anomaly detection in scenarios where the problem is too complicated for traditional methods, and enough training data is available. However, the amount of real-world injection molding machine data used in the experiments is relatively small, and therefore further experiments should be performed with larger datasets to obtain more generalizable results.
Background: One of the main concerns of public health surveillance is to preserve the physical and mental health of older adults while supporting their independence and privacy. On the other hand, to better assist tho...
详细信息
Background: One of the main concerns of public health surveillance is to preserve the physical and mental health of older adults while supporting their independence and privacy. On the other hand, to better assist those individuals with essential health care services in the event of an emergency, their regular activities should be monitored. Internet of Things (IoT) sensors may be employed to track the sequence of activities of individuals via ambient sensors, providing real-time insights on daily activity patterns and easy access to the data through the connected ecosystem. Previous surveys to identify the regular activity patterns of older adults were deficient in the limited number of participants, short period of activity tracking, and high reliance on predefined normal activity. Objective: The objective of this study was to overcome the aforementioned challenges by performing a pilot study to evaluate the utilization of large-scale data from smart home thermostats that collect the motion status of individuals for every 5-minute interval over a long period of time. Methods: From a large-scale dataset, we selected a group of 30 households who met the inclusion criteria (having at least 8 sensors, being connected to the system for at least 355 days in 2018, and having up to 4 occupants). The indoor activity patterns were captured through motion sensors. We used the unsupervised, time-based, deep neural-network architecture long short-term memory-variational autoencoder to identify the regular activity pattern for each household on 2 time scales: annual and weekday. The results were validated using 2019 records. The area under the curve as well as loss in 2018 were compatible with the 2019 schedule. Daily abnormal behaviors were identified based on deviation from the regular activity model. Results: The utilization of this approach not only enabled us to identify the regular activity pattern for each household but also provided other insights by assessing sleep beh
Missing data values and differing sampling rates, particularly for important parameters such as particle size and stream composition, are a common problem in minerals processing plants. Missing data imputation is used...
详细信息
Missing data values and differing sampling rates, particularly for important parameters such as particle size and stream composition, are a common problem in minerals processing plants. Missing data imputation is used to avoid information loss (due to downsampling or discarding incomplete records). A recent deep -learning technique, variational autoencoders (VAEs), has been used for missing data imputation in image data, and was compared here to imputation by mean replacement and by principal component analysis (PCA) imputation. The techniques were compared using a synthetic, nonlinear dataset, and a simulated milling circuit dataset, which included process disturbances, measurement noise, and feedback control. Each dataset was corrupted with missing values in 20% of records (lightly corrupted) and in 90% of records (heavily corrupted). For both lightly and heavily corrupted datasets, the root mean squared error of prediction for VAE imputation was lower than the traditional methods. Possibilities for the extension of missing data imputation to inferential sensing are discussed. (C) 2018, IFAC (International Federation of Automatic Control) Hosting by Elsevier Ltd. All rights reserved.
Machine learning techniques can help to represent and solve quantum systems. Learning measurement outcome distribution of quantum ansatz is useful for characterization of near-term quantum computing devices. In this w...
详细信息
Machine learning techniques can help to represent and solve quantum systems. Learning measurement outcome distribution of quantum ansatz is useful for characterization of near-term quantum computing devices. In this work, we use the popular unsupervised machine learning model, variational autoencoder (VAE), to reconstruct the measurement outcome distribution of quantum ansatz. The number of parameters in the VAE are compared with the number of measurement outcomes. The numerical results show that VAE can efficiently learn the measurement outcome distribution with few parameters. The influence of entanglement on the task is also revealed.
Here we report a method of finding multiple crystal structures similar to the known crystal structures of materials on database through machine learning. The radial distribution function is used to represent the gener...
详细信息
Here we report a method of finding multiple crystal structures similar to the known crystal structures of materials on database through machine learning. The radial distribution function is used to represent the general characteristics of the known crystal structures, and then the variational autoencoder is employed to generate a set of representative crystal replicas defined in a two-dimensional optimal continuous space. For given chemical compositions and crystal volume, we generate random crystal structures using constraints for crystal symmetry and atomic positions and directly compare their radial distribution functions with those of the known and/or replicated crystals. For selected crystal structures, energy minimization is subsequently performed through firstprinciples electronic structure calculations. This approach enables us to predict a set of new low-energy crystal structures using only the information on the radial distribution functions of the known structures.
Designers increasingly rely on parametric design studies to explore and improve structural concepts based on quantifiable metrics, generally either by generating design variations manually or using optimization method...
详细信息
Designers increasingly rely on parametric design studies to explore and improve structural concepts based on quantifiable metrics, generally either by generating design variations manually or using optimization methods. Unfortunately, both of these approaches have important shortcomings: effectively searching a large design space manually is infeasible, and design optimization overlooks qualitative aspects important in architectural and structural design. There is a need for methods that take advantage of computing intelligence to augment a designer's creativity while guiding-not forcing-their search for better-performing solutions. This research addresses this need by integrating conditional variational autoencoders in a performance-driven design exploration framework. First, a sampling algorithm generates a dataset of meaningful design options from an unwieldy design space. Second, a performance-conditioned variational autoencoder with a low-dimensional latent space is trained using the collected data. This latent space is intuitive to explore by designers even as it offers a diversity of high-performing design options.
Most existing image dehazing methods based learning are less able to perform well to real hazy *** important reason is that they are trained on synthetic hazy images whose distribution is different from real hazy *** ...
详细信息
Most existing image dehazing methods based learning are less able to perform well to real hazy *** important reason is that they are trained on synthetic hazy images whose distribution is different from real hazy *** relieve this issue, this paper proposes a new hazy scene generation model based on domain adaptation, which uses a variational autoencoder to encode the synthetic hazy image pairs and the real hazy images into the latent space to *** synthetic hazy image pairs guide the model to learn the mapping of clear images to hazy images, the real hazy images are used to adapt the synthetic hazy images' latent space to real hazy images through generative adversarial loss, so as to make the generative hazy images' distribution as close to the real hazy images' distribution as *** comparing the results of the model with traditional physical scattering models and Adobe Lightroom CC software, the hazy images generated in this paper is more *** end-to-end domain adaptation model is also very convenient to synthesize hazy images without depth *** traditional method to dehaze the synthetic hazy images generated by this paper, both SSIM and PSNR have been improved, proved that the effectiveness of our *** non-reference haze density evaluation algorithm and other quantitative evaluation also illustrate the advantages of our method in synthetic hazy images.
暂无评论