In this paper, a novel method as a combination of the expectation-maximization (em) algorithm and Variogram is proposed to decompose the longitudinal measurement errors in the absence of replications and the presence ...
详细信息
In this paper, a novel method as a combination of the expectation-maximization (em) algorithm and Variogram is proposed to decompose the longitudinal measurement errors in the absence of replications and the presence of multiple instrumental measurement errors. In the proposed method, multiple measurements are considered where the units are observed by several distinct instruments (gauges). The approach decouples the observed variance of the measurement model into the process and measurement system variances. In addition, it decomposes the variance of multiple instruments into the process and instrument variances. In the end, the proposed model is validated and tested based on simulated longitudinal data as well as a real case study related to the Framingham Heart study, measuring systolic blood pressure by multiple instruments. In addition, the robustness of the proposed method to the missing values, a common problem in longitudinal data, is demonstrated.
The probabilistic distribution of wind speed is critical information required in evaluating wind resources, designing wind farms, and mitigating the possible risks in wind power expansion. Various distributions have b...
详细信息
The probabilistic distribution of wind speed is critical information required in evaluating wind resources, designing wind farms, and mitigating the possible risks in wind power expansion. Various distributions have been used in the research literature to estimate wind speed distribution. In this paper, we propose a flexible family of mixture distributions, whose elements are convex linear combinations of the skew-t -normal Birnbaum-Saunders distributions and is suitable for modelling heavy-tailed data with a heterogeneous population. These distributions are then used to estimate the wind speed distribution. The performance of the proposed family has been compared with five mixture models of the already used distributions, using wind speed data collected at nine stations across Ontario, Canada. The results indicate that mixture models generally provide a better fit than unimodal distributions, according to the model selection criterion. Based on the obtained results, the proposed family provides highly flexible models at all selected stations. It functions better than the other considered distributions at seven of the stations, whereas it is ranked second in the remaining two stations. The proposed family can be utilized in a widespread manner to describe the wind speed in Canada, as well as other regions with similar features. (c) 2020 Elsevier Ltd. All rights reserved.
The gamma-Poisson and beta-binomial mixture distributions are used for analyzing count-valued data, and the estimation of the hyper-parameters including the shape and/or scale parameters is important in the empirical ...
详细信息
The gamma-Poisson and beta-binomial mixture distributions are used for analyzing count-valued data, and the estimation of the hyper-parameters including the shape and/or scale parameters is important in the empirical Bayes inference. The maximum likelihood method requires the nested loops for solving the non-linear equations at each step of iteration in the emalgorithm. To avoid the extra loops, we derive the closed-form updating procedures at each step of iteration by using the score-adjusted method. The performance is compared by simulation with the maximum likelihood estimators.
Suppose a time sequence of networks is observed. It is known that the probabilistic behaviors of the networks do not change over time, except at a few time points. These time points are usually called change points, w...
详细信息
Suppose a time sequence of networks is observed. It is known that the probabilistic behaviors of the networks do not change over time, except at a few time points. These time points are usually called change points, whose number and locations are unknown. This paper proposes a method for automatically estimating such change points and the community structures of the networks. The proposed method invokes the minimum description length principle to derive a model selection criterion, where the best estimates are defined as its minimizer. It is shown that this selection criterion yields consistent estimates for the change points as well as the community structures. For practical minimization of the selection criterion, a bottom-up search algorithm that combines the em-algorithm with variational approximation is developed. The promising empirical properties of the proposed method are illustrated via a sequence of numerical experiments and applications to some real datasets. To the best of the authors' knowledge, this method is one of the earliest that provides consistent estimates in the context of change point detection for time-evolving networks.
Probabilistic topic modeling of text collections has been recently developed mainly within the framework of graphical models and Bayesian inference. In this paper we introduce an alternative semi-probabilistic approac...
详细信息
Probabilistic topic modeling of text collections has been recently developed mainly within the framework of graphical models and Bayesian inference. In this paper we introduce an alternative semi-probabilistic approach, which we call additive regularization of topic models (ARTM). Instead of building a purely probabilistic generative model of text we regularize an ill-posed problem of stochastic matrix factorization by maximizing a weighted sum of the log-likelihood and additional criteria. This approach enables us to combine probabilistic assumptions with linguistic and problem-specific requirements in a single multi-objective topic model. In the theoretical part of the work we derive the regularized em-algorithm and provide a pool of regularizers, which can be applied together in any combination. We show that many models previously developed within Bayesian framework can be inferred easier within ARTM and in some cases generalized. In the experimental part we show that a combination of sparsing, smoothing, and decorrelation improves several quality measures at once with almost no loss of the likelihood.
In this paper we propose a new four-parameters distribution with increasing, decreasing, bathtub-shaped and unimodal failure rate, called as the exponentiated Weibull-Poisson (EWP) distribution. The new distribution a...
详细信息
In this paper we propose a new four-parameters distribution with increasing, decreasing, bathtub-shaped and unimodal failure rate, called as the exponentiated Weibull-Poisson (EWP) distribution. The new distribution arises on a latent complementary risk problem base and is obtained by compounding exponentiated Weibull (EW) and Poisson distributions. This distribution contains several lifetime sub-models such as: generalized exponential-Poisson (GEP), complementary Weibull-Poisson (CWP), complementary exponential-Poisson (CEP), exponentiated Rayleigh-Poisson (ERP) and Rayleigh-Poisson (RP) distributions. We obtain several properties of the new distribution such as its probability density function, its reliability and failure rate functions, quantiles and moments. The maximum likelihood estimation procedure via an em-algorithm is presented in this paper. Sub-models of the EWP distribution are studied in details. In the end, applications to two real data sets are given to show the flexibility and potentiality of the new distribution. (C) 2013 IMACS. Published by Elsevier B.V. All rights reserved.
Order selection is an important step in the application of finite mixture models. Classical methods such as AIC and BIC discourage complex models with a penalty directly proportional to the number of mixing components...
详细信息
Order selection is an important step in the application of finite mixture models. Classical methods such as AIC and BIC discourage complex models with a penalty directly proportional to the number of mixing components. In contrast, Chen and Khalili propose to link the penalty to two types of overfitting. In particular, they introduce a regularization penalty to merge similar subpopulations in a mixture model, where the shrinkage idea of regularized regression is seamlessly employed. However, the new method requires an effective and efficient algorithm. When the popular expectation-maximization (em)algorithm is used, we need to maximize a nonsmooth and nonconcave objective function in the M-step, which is computationally challenging. In this article, we show that such an objective function can be transformed into a sum of univariate auxiliary functions. We then design an iterative thresholding descent algorithm (ITD) to efficiently solve the associated optimization problem. Unlike many existing numerical approaches, the new algorithm leads to sparse solutions and thereby avoids undesirable ad hoc steps. We establish the convergence of the ITD and further assess its empirical performance using both simulations and real data examples.
The main object of this paper is to consider structural comparative calibration models under the assumption that the unknown quantity being measured is not identically distributed for all units. We consider the situat...
详细信息
The main object of this paper is to consider structural comparative calibration models under the assumption that the unknown quantity being measured is not identically distributed for all units. We consider the situation where the mean of the unknown quantity being measured is different within subgroups of the population. Method of moments and maximum likelihood estimators are considered for estimating the parameters in the model. Large sample inference is facilitated by the derivation of the asymptotic variances. An application to a data set which indeed motivated the consideration of such general model and was obtained by measuring the heights of a group of trees with five different instruments is considered.
A new distribution with increasing, decreasing, bathtub-shaped and unimodal failure rate forms called as the generalized modified Weibull power series (GMWPS) distribution is proposed. The new distribution is construc...
详细信息
A new distribution with increasing, decreasing, bathtub-shaped and unimodal failure rate forms called as the generalized modified Weibull power series (GMWPS) distribution is proposed. The new distribution is constructed based on a latent complementary risk problem and is obtained by compounding generalized modified Weibull (GMW) and power series distributions. The new distribution contains, as special submodels, several important distributions which are discussed in the literature, such as generalized modified Weibull Poisson (GMWP) distribution, generalized modified Weibull Geometric (GMWG) distribution, generalized modified Weibull Logarithmic (GMWL) distribution, generalized modified Weibull Binomial (GMWB) distribution, among others. A comprehensive mathematical treatment of the new distribution is provided. We provide closed-form expressions for the density, cumulative distribution, survival function, failure rate function, the rth raw moment, and also the moments of order statistics. Expressions for the Renyi and Shannon entropies are also given. Moreover, we discuss maximum likelihood estimation and provide formula for the elements of the Fisher information matrix. We consider the em-algorithm for computing the estimates. Simulation studies and two real data set applications are also given for illustration of the flexibility and potentiality of the new distribution. (C) 2015 Elsevier B.V. All rights reserved.
We propose a method for estimating parameters in binomial regression models when the response variable is missing and the missing data mechanism is nonignorable. We assume throughout that the covariates are fully obse...
详细信息
We propose a method for estimating parameters in binomial regression models when the response variable is missing and the missing data mechanism is nonignorable. We assume throughout that the covariates are fully observed. Using a legit model for the missing data mechanism, we show how parameter estimation can be accomplished using the emalgorithm by the method of weights proposed in Ibrahim (1990, Journal of the American Statistical Association 85, 765-769). An example from the Six Cities Study (Ware et al., 1984, American Review of Respiratory Diseases 129, 366-374) is presented to illustrate the method.
暂无评论