Recently the authors have proposed a homogeneous and self-dual algorithm for solving the monotone complementarity problem (MCP) [5]. The algorithm is a single phase interior-point type method;nevertheless, it yields e...
详细信息
Recently the authors have proposed a homogeneous and self-dual algorithm for solving the monotone complementarity problem (MCP) [5]. The algorithm is a single phase interior-point type method;nevertheless, it yields either an approximate optimal solution or detects a possible infeasibility of the problem. In this paper we specialize the algorithm to the solution of general smooth convexoptimization problems, which also possess nonlinear inequality constraints and free variables. We discuss an implementation of the algorithm for large-scale sparse convexoptimization. Moreover, we present computational results for solving quadratically constrained quadratic programming and geometric programming problems, where some of the problems contain more than 100,000 constraints and variables. The results indicate that the proposed algorithm is also practically efficient.
作者:
Wang, HaoyueIbrahim, ShibalMazumder, RahulMIT
Operat Res Ctr Cambridge MA 02142 USA MIT
Dept Elect Engn & Comp Sci Cambridge MA 02142 USA MIT
Operat Res Ctr Sloan Sch Management Cambridge MA 02142 USA MIT
Ctr Stat Cambridge MA 02142 USA
We explore computational aspects of maximum likelihood estimation of the mixture proportions in a nonparametric finite mixture model---a convexoptimization problem with old roots in statistics. Motivated by problems ...
详细信息
We explore computational aspects of maximum likelihood estimation of the mixture proportions in a nonparametric finite mixture model---a convexoptimization problem with old roots in statistics. Motivated by problems in shape constrained inference, we also consider structured variants of this problem with additional convex polyhedral constraints. We propose a new cubic regularized Newton method for this problem and present novel worst-case and local computational guarantees for our algorithm. We extend earlier work by Nesterov and Polyak to the case of a self-concordant objective with polyhedral constraints, such as the ones considered herein. We propose a Frank--Wolfe method to solve the cubic regularized Newton subproblem and derive efficient solutions for the linear optimization oracles that may be of independent interest. In the particular case of Gaussian mixtures without shape constraints, we derive bounds on how well the finite mixture problem approximates the infinite-dimensional Kiefer-Wolfowitz maximum likelihood estimator. Experiments on synthetic and real datasets suggest that our proposed algorithms exhibit improved runtimes and scalability features over prior benchmarks.
convex nonsmooth optimization problems, whose solutions live in very high dimensional spaces, have become ubiquitous. To solve them, the class of first-order algorithms known as proximal splitting algorithms is partic...
详细信息
convex nonsmooth optimization problems, whose solutions live in very high dimensional spaces, have become ubiquitous. To solve them, the class of first-order algorithms known as proximal splitting algorithms is particularly adequate: they consist of simple operations, handling the terms in the objective function separately. In this overview, we demystify a selection of recent proximal splitting algorithms: we present them within a unified framework, which consists in applying splitting methods for monotone inclusions in primal-dual product spaces, with well-chosen metrics. Along the way, we easily derive new variants of the algorithms and revisit existing convergence results, extending the parameter ranges in several cases. In particular, we emphasize that when the smooth term in the objective function is quadratic, e.g., for least-squares problems, convergence is guaranteed with larger values of the relaxation parameter than previously known. Such larger values are usually beneficial for the convergence speed in practice.
We propose a framework for building preconditioners for sequences of linear systems of the form (A + Delta(k))x(k) = b(k), where A is symmetric positive semidefinite and Delta(k) is diagonal positive semidefinite. Suc...
详细信息
We propose a framework for building preconditioners for sequences of linear systems of the form (A + Delta(k))x(k) = b(k), where A is symmetric positive semidefinite and Delta(k) is diagonal positive semidefinite. Such sequences arise in several optimization methods, e.g., in affine-scaling methods for bound-constrained convex quadratic programming and bound-constrained linear least squares, as well as in trust-region and overestimation methods for convex unconstrained optimization problems and nonlinear least squares. For all the matrices of a sequence, the preconditioners are obtained by updating any preconditioner for A available in the LDLT form. The preconditioners in the framework satisfy the natural requirement of being effective on slowly varying sequences;furthermore, under an additional property they are also able to cluster eigenvalues of the preconditioned matrix when some entries of Delta(k) are sufficiently large. We present two low-cost preconditioners sharing the above-mentioned properties and evaluate them on sequences of linear systems generated by the reflective Newton method applied to bound-constrained convex quadratic programming problems and on sequences arising in solving nonlinear least-squares problems with the regularized Euclidean residual method. The results of the numerical experiments show the effectiveness of these preconditioners.
We propose the first general and scalable framework to design certifiable algorithms for robust geometric perception in the presence of outliers. Our first contribution is to show that estimation using common robust c...
详细信息
We propose the first general and scalable framework to design certifiable algorithms for robust geometric perception in the presence of outliers. Our first contribution is to show that estimation using common robust costs, such as truncated least squares (TLS), maximum consensus, Geman-McClure, Tukey's biweight, among others, can be reformulated as polynomial optimization problems (POPs). By focusing on the TLS cost, our second contribution is to exploit sparsity in the POP and propose a sparse semidefinite programming(SDP) relaxation that is much smaller than the standard Lasserre's hierarchy while preserving empirical exactness, i.e., the SDP recovers the optimizer of the nonconvex POP with an optimality certificate. Our third contribution is to solve the SDP relaxations at an unprecedented scale and accuracy by presenting STRIDE, a solver that blends global descent on the convex SDP with fast local search on the nonconvex POP. Our fourth contribution is an evaluation of the proposed framework on six geometric perception problems including single and multiple rotation averaging, point cloud and mesh registration, absolute pose estimation, and category-level object pose and shape estimation. Our experiments demonstrate that (i) our sparse SDP relaxation is empirically exact with up to 60%-90% outliers across applications;(ii) while still being far from real-time, STRIDE is up to 100 times faster than existing SDP solvers on medium-scale problems, and is the only solver that can solve large-scale SDPs with hundreds of thousands of constraints to high accuracy;(iii) STRIDE safeguards existing fast heuristics for robust estimation (e.g., RANSAC or Graduated Non-convexity), i.e., it certifies global optimality if the heuristic estimates are optimal, or detects and allows escaping local optima when the heuristic estimates are suboptimal.
The entropic value-at-risk (EVaR) is a new coherent risk measure, which is an upper bound for both the value-at-risk (VaR) and conditional value-at-risk (CVaR). One of the important properties of the EVaR is that it i...
详细信息
The entropic value-at-risk (EVaR) is a new coherent risk measure, which is an upper bound for both the value-at-risk (VaR) and conditional value-at-risk (CVaR). One of the important properties of the EVaR is that it is strongly monotone over its domain and strictly monotone over a broad sub-domain including all continuous distributions, whereas well-known monotone risk measures such as the VaR and CVaR lack this property. A key feature of a risk measure, besides its financial properties, is its applicability in large-scale sample-based portfolio optimization. If the negative return of an investment portfolio is a differentiable convex function for any sampling observation, the portfolio optimization with the EVaR results in a differentiable convex program whose number of variables and constraints is independent of the sample size, which is not the case for the VaR and CVaR even if the portfolio rate linearly depends on the decision variables. This enables us to design an efficient algorithm using differentiable convexoptimization. Our extensive numerical study indicates the high efficiency of the algorithm in largescales, when compared with the existing convexoptimization software packages. The computational efficiency of the EVaR and CVaR approaches are generally similar, but the EVaR approach outperforms the other as the sample size increases. Moreover, the comparison of the portfolios obtained for a real case by the EVaR and CVaR approaches shows that the EVaR-based portfolios can have better best, mean, and worst return rates as well as Sharpe ratios. (C) 2019 Published by Elsevier B.V.
In this work we develop a numerical method for solving a type of convex graph- structured tensor optimization problem. This type of problem, which can be seen as a generalization of multimarginal optimal transport pro...
详细信息
In this work we develop a numerical method for solving a type of convex graph- structured tensor optimization problem. This type of problem, which can be seen as a generalization of multimarginal optimal transport problems with graph-structured costs, appears in many applications. Examples are unbalanced optimal transport and multispecies potential mean field games, where the latter is a class of nonlinear density control problems. The method we develop is based on coordinate ascent in a Lagrangian dual, and under mild assumptions we prove that the algorithm converges globally. Moreover, under a set of stricter assumptions, the algorithm converges R-linearly. To perform the coordinate ascent steps one has to compute projections of the tensor, and doing so by brute force is in general not computationally feasible. Nevertheless, for certain graph structures it is possible to derive efficient methods for computing these projections, and here we specifically consider the graph structure that occurs in multispecies potential mean field games. We also illustrate the methodology on a numerical example from this problem class.
Pixel-level feature detection from images is an essential but challenging task encountered in domains such as detecting defects in manufacturing systems and detecting tumors in medical imaging. Often, the real image c...
详细信息
Pixel-level feature detection from images is an essential but challenging task encountered in domains such as detecting defects in manufacturing systems and detecting tumors in medical imaging. Often, the real image contains multiple feature types. The types with higher pixel intensities are termed as positive (extreme) features and the ones with lower pixel intensities as negative (extreme) features. For example, when planning a medical treatment, it is important to identify, (a) calcification (a pathological feature which can result in a post-surgical complications) as positive features, and (b) soft tissues (organ morphology, knowledge of which can support pre-surgical planning) as negative features, from a preoperative computed tomography image of the human heart. However, this is not an easy task because (a) conventional segmentation techniques require manual intervention and post-processing, and (b) existing automatic approaches do not distinguish positive features from negative. In this work, we propose a novel, automatic image decomposition-based sparse extreme pixel-level feature detection model to decompose an image into mean and extreme features. To estimate model parameters, a high-dimensional least squares regression with regularization and constraints is utilized. An efficient algorithm based on the alternating direction method of multipliers and the proximal gradient method is developed to solve the large-scaleoptimization problem. The effectiveness of the proposed model is demonstrated using synthetic tests and a real-world case study, where the model exhibits superior performance over existing methods.
暂无评论