In this paper we present a randomized selection algorithm that with high probability 1 - 1/n(rho), for any constant p > 1 requires n + C + o(n) comparisons to determine the Cth order statistic of n keys thus matchi...
详细信息
In this paper we present a randomized selection algorithm that with high probability 1 - 1/n(rho), for any constant p > 1 requires n + C + o(n) comparisons to determine the Cth order statistic of n keys thus matching a corresponding lower bound on the average number of comparisons required. Low order terms in the number of comparisons performed can also be reduced by extending the algorithm of Floyd and Rivest and analyzing its resulting performance more carefully. (C) 2003 Elsevier B.V. All rights reserved.
In Mutual Search, recently introduced by Buhrman et al. (1998), static agents are searching for each other: each agent is assigned one of n locations, and the computations proceed by agents sending queries from their ...
详细信息
In Mutual Search, recently introduced by Buhrman et al. (1998), static agents are searching for each other: each agent is assigned one of n locations, and the computations proceed by agents sending queries from their location to other locations, until one of the queries arrives at the other agent. The cost of a search is the number of queries made. The best known bounds for randomized protocols using private coins are (1) a protocol with worst-case expected cost of [(n + 1)/2], and (2) a lower bound of (n - 1)/8 queries for randomized protocols which make only a bounded number of coin-tosses. In this paper we strictly improve the lower bound, and present a new upper bound for shared random coins. Specifically, we first prove that the worst-case expected cost of any randomized protocol for two-agent mutual search is at least (Ir + 1)/3. This is an improvement both in terms of number of queries and in terms of applicability. We also give a randomized algorithm for mutual search with worst-case expected cost of (n + 1)/3. This algorithm works under the assumption that the agents share a random bit string. This bound shows that no better lower bound can be obtained using our technique. (C) 1999 Published by Elsevier Science B.V. All rights reserved.
Consideration was given to estimation of the eigenvector corresponding to the greatest eigenvalue of a stochastic matrix. There exist numerous applications of this problem arising at ranking the results of search, coo...
详细信息
Consideration was given to estimation of the eigenvector corresponding to the greatest eigenvalue of a stochastic matrix. There exist numerous applications of this problem arising at ranking the results of search, coordination of the multiagent system actions, network control, and data analysis. The standard technique for its solution comes to the power method with an additional regularization of the original matrix. A new randomized algorithm was proposed, and a uniform-over the entire class of the stochastic matrices of a given size-upper boundary of the convergence rate was validated. It is given by C root ln(N)/n, where C is an absolute constant, N is size, and n is the number of iterations. This boundary seems promising because ln(N) is smallish even for a very great size. The algorithm relies on the mirror descent method for the problems of convex stochastic optimization. Applicability of the method to the PageRank problem of ranking the Internet pages was discussed.
The scenario-based optimization approach ("scenario approach") provides an intuitive way of approximating the solution to chance-constrained optimization programs, based on finding the optimal solution under...
详细信息
The scenario-based optimization approach ("scenario approach") provides an intuitive way of approximating the solution to chance-constrained optimization programs, based on finding the optimal solution under a finite number of sampled outcomes of the uncertainty ("scenarios"). A key merit of this approach is that it neither requires explicit knowledge of the uncertainty set, as in robust optimization, nor of its probability distribution, as in stochastic optimization. The scenario approach is also computationally efficient because it only requires the solution to a convex optimization program, even if the original chance-constrained problem is nonconvex. Recent research has obtained a rigorous foundation for the scenario approach, by establishing a direct link between the number of scenarios and bounds on the constraint violation probability. These bounds are tight in the general case of an uncertain optimization problem with a single chance constraint. This paper shows that the bounds can be improved in situations where the chance constraints have a limited "support rank," meaning that they leave a linear subspace unconstrained. Moreover, it shows that also a combination of multiple chance constraints, each with individual probability level, is admissible. As a consequence of these results, the number of scenarios can be reduced from that prescribed by the existing theory for problems with the indicated structural property. This leads to an improvement in the objective value and a reduction in the computational complexity of the scenario approach. The proposed extensions have many practical applications, in particular, high-dimensional problems such as multistage uncertain decision problems or design problems of large-scale systems.
We study the topic of dimensionality reduction for k-means clustering. Dimensionality reduction encompasses the union of two approaches: 1) feature selection and 2) feature extraction. A feature selection-based algori...
详细信息
We study the topic of dimensionality reduction for k-means clustering. Dimensionality reduction encompasses the union of two approaches: 1) feature selection and 2) feature extraction. A feature selection-based algorithm for k-means clustering selects a small subset of the input features and then applies k-means clustering on the selected features. A feature extraction-based algorithm for k-means clustering constructs a small set of new artificial features and then applies k-means clustering on the constructed features. Despite the significance of k-means clustering as well as the wealth of heuristic methods addressing it, provably accurate feature selection methods for k-means clustering are not known. On the other hand, two provably accurate feature extraction methods for k-means clustering are known in the literature;one is based on random projections and the other is based on the singular value decomposition (SVD). This paper makes further progress toward a better understanding of dimensionality reduction for k-means clustering. Namely, we present the first provably accurate feature selection method for k-means clustering and, in addition, we present two feature extraction methods. The first feature extraction method is based on random projections and it improves upon the existing results in terms of time complexity and number of features needed to be extracted. The second feature extraction method is based on fast approximate SVD factorizations and it also improves upon the existing results in terms of time complexity. The proposed algorithms are randomized and provide constant-factor approximation guarantees with respect to the optimal k-means objective value.
The identification of polynomial Nonlinear Autoregressive [Moving Average] models with exogenous variables (NAR[MA]X) is typically carried out with incremental model building techniques that progressively select the t...
详细信息
The identification of polynomial Nonlinear Autoregressive [Moving Average] models with exogenous variables (NAR[MA]X) is typically carried out with incremental model building techniques that progressively select the terms to include in the model. The Model Structure Selection (MSS) turns out to be the hardest task of the identification process due to the difficulty of correctly evaluating the importance of a generic term. As a result, classical MSS methods sometimes yield unsatisfactory models, that are unreliable over long-range prediction horizons. The MSS problem is here recast into a probabilistic framework based on which a randomized algorithm for MSS is derived, denoted RaMSS. The method introduces a tentative probability distribution over models and progressively updates it by extracting useful information on the importance of each term from sampled model structures. The proposed method is validated over models with different characteristics by means of Monte Carlo simulations, which show its advantages over classical and competitor probabilistic MSS methods in terms of both reliability and computational efficiency. (C) 2015 Elsevier Ltd. All rights reserved.
This paper studies the point location problem in Delaunay triangulations without preprocessing and additional storage. The proposed procedure finds the query point by simply "walking through" the triangulati...
详细信息
This paper studies the point location problem in Delaunay triangulations without preprocessing and additional storage. The proposed procedure finds the query point by simply "walking through" the triangulation, after selecting a "good starting point" by random sampling. The analysis generalizes and extends a recent result for d = 2 dimensions by proving this procedure takes expected time close to O(n(1/(d+1))) for point location in Delaunay triangulations of n random points in d = 3 dimensions. Empirical results in both two and three dimensions show that this procedure is efficient in practice. (C) 1999 Elsevier Science B.V. All rights reserved.
We propose a randomized nonmonotone block proximal gradient (RNBPG) method for minimizing the sum of a smooth (possibly nonconvex) function and a block-separable (possibly nonconvex nonsmooth) function. At each iterat...
详细信息
We propose a randomized nonmonotone block proximal gradient (RNBPG) method for minimizing the sum of a smooth (possibly nonconvex) function and a block-separable (possibly nonconvex nonsmooth) function. At each iteration, this method randomly picks a block according to any prescribed probability distribution and solves typically several associated proximal subproblems that usually have a closed-form solution, until a certain progress on objective value is achieved. In contrast to the usual randomized block coordinate descent method [P. Richtarik and M. Takac, Math. Program., 144 (2014), pp. 1-38;A. Patrascu and I. Necoara, J. Global Optim., 61 (2015), pp. 19-46], our method has a nonmonotone flavor and uses variable stepsizes that can partially utilize the local curvature information of the smooth component of objective function. We show that any accumulation point of the solution sequence of the method is a stationary point of the problem almost surely and the method is capable of finding an approximate stationary point with high probability. We also establish a sublinear rate of convergence for the method in terms of the minimal expected squared norm of certain proximal gradients over the iterations. When the problem under consideration is convex, we show that the expected objective values generated by RNBPG converge to the optimal value of the problem. Under some assumptions, we further establish a sublinear and linear rate of convergence on the expected objective values generated by a monotone version of RNBPG. Finally, we conduct some preliminary experiments to test the performance of RNBPG on the l(1)-regularized least-squares problem, a dual support vector machine problem in machine learning, the l(0)-regularized least-squares problem, and a regularized matrix completion model. The computational results demonstrate that our method substantially outperforms the randomized block coordinate descent method with fixed or variable stepsizes.
It is shown that the expected number of leaves evaluated by randomized alpha-beta search for evaluating uniform game trees of degree d and height h is O((B-d)(h)), where B-d = d/2 + Ind + O(1). It was shown by Saks an...
详细信息
It is shown that the expected number of leaves evaluated by randomized alpha-beta search for evaluating uniform game trees of degree d and height h is O((B-d)(h)), where B-d = d/2 + Ind + O(1). It was shown by Saks and Wigderson [Proceedings of 27th Annual Symposium on Foundations of Computer Science (1986), pp. 29-38] that the optimal branching factor of randomized algorithms for evaluating uniform trees of degree d is Bd = (d-1+root d(2)+14d+1)/4=d/2+O(1). As B-d/B*(d)=1+O(lnd/d), randomized alpha-beta search is asymptotically optimal for evaluating uniform game trees as the degree of tree increases.
We introduce a Generalized LU Factorization (GLU) for low-rank matrix approx-imation. We relate this to past approaches and extensively analyze its approximation properties. The established deterministic guarantees ar...
详细信息
We introduce a Generalized LU Factorization (GLU) for low-rank matrix approx-imation. We relate this to past approaches and extensively analyze its approximation properties. The established deterministic guarantees are combined with sketching ensembles satisfying Johnson-Lindenstrauss properties to present complete bounds. Particularly good performance is shown for the subsampled randomized Hadamard transform (SRHT) ensemble. Moreover, the factorization is shown to unify and generalize many past algorithms, sometimes providing strictly better approx-imations. It also helps to explain the effect of sketching on the growth factor during Gaussian elimination.
暂无评论