We introduce a new stochastic model incorporating a mixture of uniform and Laplace distributions. We present basic theoretical properties of this model and discuss related computational issues of parameter estimation ...
详细信息
We introduce a new stochastic model incorporating a mixture of uniform and Laplace distributions. We present basic theoretical properties of this model and discuss related computational issues of parameter estimation via expectation-maximization computa-tional schemes. We check the performance of the estimation algorithm on synthetic data and provide a data example illustrating modeling potential of this novel methodology. A related model involving a mixture of uniform and exponential distributions is studied as well along the same lines.(c) 2023 Elsevier B.V. All rights reserved.
High frequency financial data is characterized by non-normality: asymmetric, leptokurtic and fat-tailed behaviour. The normal distribution is therefore inadequate in capturing these characteristics. To this end, vario...
详细信息
High frequency financial data is characterized by non-normality: asymmetric, leptokurtic and fat-tailed behaviour. The normal distribution is therefore inadequate in capturing these characteristics. To this end, various flexible distributions have been proposed. It is well known that mixture distributions produce flexible models with good statistical and probabilistic properties. In this work, a finite mixture of two special cases of Generalized Inverse Gaussian distribution has been constructed. Using this finite mixture as a mixing distribution to the Normal Variance Mean Mixture we get a Normal Weighted Inverse Gaussian (NWIG) distribution. The second objective, therefore, is to construct and obtain properties of the NWIG distribution. The maximum likelihood parameter estimates of the proposed model are estimated via emalgorithm and three data sets are used for application. The result shows that the proposed model is flexible and fits the data well.
Normal Variance-Mean Mixture (NVMM) provides a general framework for deriving models with desirable properties for modelling financial market variables such as exchange rates, equity prices, and interest rates measur...
详细信息
Normal Variance-Mean Mixture (NVMM) provides a general framework for deriving models with desirable properties for modelling financial market variables such as exchange rates, equity prices, and interest rates measured over short time intervals, i.e. daily or weekly. Such data sets are characterized by non-normality and are usually skewed, fat-tailed and exhibit excess kurtosis. The Generalised Hyperbolic distribution (GHD) introduced by Barndorff-Nielsen (1977) which act as Normal variance-mean mixtures with Generalised Inverse Gaussian (GIG) mixing distribution nest a number of special and limiting case distributions. The Normal Inverse Gaussian (NIG) distribution is obtained when the Inverse Gaussian is the mixing distribution, i.e., the index parameter of the GIG is. The NIG is very popular because of its analytical tractability. In the mixing mechanism, the mix
The normal distribution is inadequate in capturing skewed and heavy-tailed behaviour of data taken over short time intervals. In addition the data can be leptokurtic. For this reason a normal weighted inverse Gaussian...
详细信息
The normal distribution is inadequate in capturing skewed and heavy-tailed behaviour of data taken over short time intervals. In addition the data can be leptokurtic. For this reason a normal weighted inverse Gaussian distribution is proposed as an alternative to the normal inverse Gaussian distribution to handle such data. The mixing distribution used in the normal variance mean mixture is a finite mixture of two special cases of Generalized Inverse Gaussian ( GIG ) distribution. The two special cases and the finite mixture are weighted inverse Gaussian distribution. The motivation for this work is that a finite mixture is more flexible than a single/standard distribution. The em -algorithm has been used for parameter estimation. Pour traiter des données présentant une asymétrie, une queue lourde et un caractère leptokurtique, nous proposons un modèle de mélange en variance où la variable de mélange est elle-même un mélange pondéré de deux loi normales inverses. L'algorithme em est utilisé pour l'estimation. Le modèle a été étudié et simulé avec succès.
In this paper, we develop a subspace system identification method for linear stochastic systems subject to observation noise with outliers. By using the least-trimmed-squares (LTS), we identify the outliers and substi...
详细信息
In this paper, we develop a subspace system identification method for linear stochastic systems subject to observation noise with outliers. By using the least-trimmed-squares (LTS), we identify the outliers and substitute the median of the output data for them, and then we apply the orthogonal decomposition (ORT) based method (Picci and Katayama, 1996 a ; Picci and Katayama, 1996 b ) to get state space models. A numerical example demonstrates the effectiveness of the proposed method. Also it is shown by numerical simulation that the model estimated by the ORT coupled with LTS can be further improved by the em-algorithm (Shumway and Stoffer, 1982; ALMutawa et al. , 2003) with considerable numerical effort.
Differences in spatial units among spatial data often complicate analyses. Spatial unit convergence, called areal interpolation, is often applied to address this problem. Of the many proposed areal interpolation metho...
详细信息
Differences in spatial units among spatial data often complicate analyses. Spatial unit convergence, called areal interpolation, is often applied to address this problem. Of the many proposed areal interpolation methods, few consider spatial autocorrelation, which is the general property of spatial data. In this paper, by employing a spatial process model, a new areal interpolation method that considers spatial autocorrelation is presented. First, we briefly survey previous areal interpolation techniques and demonstrate that the stochastic method is superior to the deterministic method in archiving accurate interpolations. Next, after a discussion on the spatial process model, a new areal interpolation method is suggested. In this method, both spatial autocorrelation and the volume preserving property, a property that should be considered in areal interpolation, are considered using a combination of a linear regression based areal interpolation method, and the spatial process model. Finally, a case study on the areal interpolation of a population is provided to demonstrate that the suggested method succeeds in improving the predictive accuracy. This case study indicates that the consideration of spatial autocorrelation is important for accurate areal interpolation.
Mixture models have become more popular in modelling compared to standard distributions. The mixing distributions play a role in capturing the variability of the random variable in the conditional distribution. Studie...
详细信息
Mixture models have become more popular in modelling compared to standard distributions. The mixing distributions play a role in capturing the variability of the random variable in the conditional distribution. Studies have lately focused on finite mixture models as mixing distributions in the mixing mechanism. In the present work, we consider a Normal Variance Mean mixture model. The mixing distribution is a finite mixture of two special cases of Generalised Inverse Gaussian distribution with indexes -1/2 and -3/2. The parameters of the mixed model are obtained via the Expectation-Maximization (EM) algorithm. The iterative scheme is based on a presentation of the normal equations. An application to some financial data has been done.
Exponentiated Generalized Weibull distribution is a probability distribution which generalizes the Weibull distribution introducing two more shapes parameters to best adjust the non-monotonic shape. The parameters of ...
详细信息
Exponentiated Generalized Weibull distribution is a probability distribution which generalizes the Weibull distribution introducing two more shapes parameters to best adjust the non-monotonic shape. The parameters of the new probability distribution function are estimated by the maximum likelihood method under progressive type II censored data via expectation maximization algorithm.
Incomplete data due to premature withdrawal (dropout) constitute a serious problem in prospective economic evaluations that has received only little attention to date. The aim of this simulation study was to investiga...
详细信息
Incomplete data due to premature withdrawal (dropout) constitute a serious problem in prospective economic evaluations that has received only little attention to date. The aim of this simulation study was to investigate how standard methods for dealing with incomplete data perform when applied to cost data with various distributions and various types of dropout. Selected methods included the product-limit estimator of Lin et al. the expectation maximisation (em-) algorithm, several types of multiple imputation (MI) and various simple methods like complete case analysis and mean imputation. Almost all methods were unbiased in the case of dropout completely at random (DCAR), but only the product-limit estimator, the em-algorithm and the MI approaches provided adequate estimates of the standard error (SE). The best estimates of the mean and SE for dropout at random (DAR) were provided by the bootstrap em-algorithm, MI regression and MI Monte Carlo Markov chain. These methods were able to deal with skewed cost data in combination with DAR and only became biased when costs also included the costs of expensive events. None of the methods were able to deal adequately with informative dropout. In conclusion, the em-algorithm with bootstrap, MI regression and MI MCMC are robust to the multivariate normal assumption and are the preferred methods for the analysis of incomplete cost data when the assumption of DCAR is not justified. Copyright (c) 2005 John Wiley & Sons, Ltd.
The problem of probabilistic topic modeling is as follows. Given a collection of text documents, find the conditional distribution over topics for each document and the conditional distribution over words (or terms) f...
详细信息
The problem of probabilistic topic modeling is as follows. Given a collection of text documents, find the conditional distribution over topics for each document and the conditional distribution over words (or terms) for each topic. Log-likelihood maximization is used to solve this problem. The problem generally has an infinite set of solutions and is ill-posed according to Hadamard. In the framework of Additive Regularization of Topic Models (ARTM), a weighted sum of regularization criteria is added to the main log-likelihood criterion. The numerical method for solving this optimization problem is a kind of an iterative em-algorithm written in a general form for an arbitrary smooth regularizer as well as for a linear combination of smooth regularizers. This paper studies the problem of convergence of the em iterative process. Sufficient conditions are obtained for the convergence to a stationary point of the regularized log-likelihood. The constraints imposed on the regularizer are not too restrictive. We give their interpretations from the point of view of the practical implementation of the algorithm. A modification of the algorithm is proposed that improves the convergence without additional time and memory costs. Experiments on a news text collection have shown that our modification both accelerates the convergence and improves the value of the criterion to be optimized.
暂无评论