We develop fixed-point algorithms for the approximation of structured matrices with rank penalties. In particular we use these fixed-point algorithms for making approximations by sums of exponentials, i.e., frequency ...
详细信息
We develop fixed-point algorithms for the approximation of structured matrices with rank penalties. In particular we use these fixed-point algorithms for making approximations by sums of exponentials, i.e., frequency estimation. For the basic formulation of the fixed-point algorithm we show that it converges to the solution of a related minimization problem, namely the one obtained by replacing the original objective function with its convex envelope and keeping the structured matrix constraint unchanged. It often happens that this solution agrees with the solution to the original minimization problem, and we provide a simple criterion for when this is true. We also provide more general fixed-point algorithms that can be used to treat the problems of making weighted approximations by sums of exponentials given equally or unequally spaced sampling. We apply the method to the case of missing data, although the above mentioned convergence results do not hold in this case. However, it turns out that the method often gives perfect reconstruction (up to machine precision) in such cases. We also discuss multidimensional extensions, and illustrate how the proposed algorithms can be used to recover sums of exponentials in several variables, but when samples are available only along a curve. (C) 2017 Elsevier Inc. All rights reserved.
A deep neural network with invertible hidden layers has a nice property of preserving all the information in the feature learning stage. In this paper, we analyse the hidden layers of residual rectifier neural network...
详细信息
A deep neural network with invertible hidden layers has a nice property of preserving all the information in the feature learning stage. In this paper, we analyse the hidden layers of residual rectifier neural networks, and investigate conditions for invertibility under which the hidden layers are invertible. A new fixed-point algorithm is developed to invert the hidden layers of residual networks. The proposed inverse algorithms are capable of inverting some residual networks which cannot be inverted by existing inverting algorithms. Furthermore, a special residual rectifier network is designed and trained on MNIST so that it can achieve comparable performance with the state-of-art performance while its hidden layers are invertible.
This paper concerns underdetermined linear instantaneous and convolutive blind source separation (BSS), i.e., the case when the number P of observed mixed signals is lower than the number N of sources. We propose part...
详细信息
This paper concerns underdetermined linear instantaneous and convolutive blind source separation (BSS), i.e., the case when the number P of observed mixed signals is lower than the number N of sources. We propose partial BSS methods, which separate P supposedly nonstationary sources of interest (while keeping residual components for the other N - P, supposedly stationary, "noise" sources). These methods are based on the general differential BSS concept that we introduced before. In the instantaneous case, the approach proposed in this paper consists of a differential extension of the FastICA method (which does not apply to underdetermined mixtures). In the convolutive case, we extend our recent-time-domain fast fixed-point C-FICA algorithm to underdetermined mixtures. Both proposed approaches thus keep the attractive features of the FastICA and C-FICA methods. Our approaches are based on differential sphering processes, followed by the optimization of the differential nonnormalized kurtosis that we introduce in this paper. Experimental tests show that these differential algorithms are much more robust to noise sources than the standard FastICA and C-FICA algorithms.
This letter presents new blind separation methods for moving average (MA) convolutive mixtures of independent MA processes. They consist of time-domain extensions of the FastICA algorithms developed by Hyvarinen and O...
详细信息
This letter presents new blind separation methods for moving average (MA) convolutive mixtures of independent MA processes. They consist of time-domain extensions of the FastICA algorithms developed by Hyvarinen and Oja for instantaneous mixtures. They perform a convolutive sphering in order to use parameter-free fast fixed-point algorithms associated with kurtotic or negentropic non-Gaussianity criteria for estimating the source innovation processes. We prove the relevance of this approach by mapping the mixtures into linear instantaneous ones. Test results are presented for artificial colored signals and speech signals.
We re-examine a fixed-point algorithm proposed recently by Hyvarinen for independent component analysis, wherein local convergence is proved subject to an ideal signal model using a square invertible mixing matrix. He...
详细信息
We re-examine a fixed-point algorithm proposed recently by Hyvarinen for independent component analysis, wherein local convergence is proved subject to an ideal signal model using a square invertible mixing matrix. Here, we derive step-size bounds which ensure monotonic convergence to a local extremum for any initial condition. Our analysis does not assume an ideal signal model but appeals rather to properties of the contrast function itself, and so applies even with noisy data and/or more sources than sensors. The results help alleviate the guesswork that often surrounds step-size selection when the observed signal does not fit an idealized model.
In this paper, we analyze the convergence of two general classes of optimization algorithms for regularized kernel methods with convex loss function and quadratic norm regularization. The first methodology is a new cl...
详细信息
In this paper, we analyze the convergence of two general classes of optimization algorithms for regularized kernel methods with convex loss function and quadratic norm regularization. The first methodology is a new class of algorithms based on fixed-point iterations that are well-suited for a parallel implementation and can be used with any convex loss function. The second methodology is based on coordinate descent, and generalizes some techniques previously proposed for linear support vector machines. It exploits the structure of additively separable loss functions to compute solutions of line searches in closed form. The two methodologies are both very easy to implement. In this paper, we also show how to remove non-differentiability of the objective functional by exactly reformulating a convex regularization problem as an unconstrained differentiable stabilization problem.
fixed-point algorithms for computing equilibria in economies with production or stationary points in constrained optimization generally use point-to-set mappings and therefore converge slowly. An alternative implement...
详细信息
fixed-point algorithms for computing equilibria in economies with production or stationary points in constrained optimization generally use point-to-set mappings and therefore converge slowly. An alternative implementation uses continuous functions with a higher dimensionality corresponding to the inclusion of activity levels or dual variables. Here we develop algorithms that only increase the dimensionality implicitly. The solution path is piecewise-linear as in other algorithms. However, when viewed in the low-dimensional space, the path within each simplex can be piecewise-linear rather than linear. Asymptotically, these paths are linear and quadratic convergence is attained.
Adjoint algorithms, and in particular those obtained through the adjoint mode of Automatic Differentiation (AD), are probably the most efficient way to obtain the gradient of a numerical simulation. This however needs...
详细信息
ISBN:
(纸本)9788494284472
Adjoint algorithms, and in particular those obtained through the adjoint mode of Automatic Differentiation (AD), are probably the most efficient way to obtain the gradient of a numerical simulation. This however needs to use the flow of data of the original simulation in reverse order, at a cost that increases with the length of the simulation. AD research looks for strategies to reduce this cost, taking advantage of the structure of the given program. One such frequent structure is fixed-point iterations, which occur e.g. in steady-state simulations, but not only. It is common wisdom that the first iterations of a fixed-point search operate on a meaningless state vector, and that reversing the corresponding data-flow may be suboptimal. An adapted adjoint strategy for this iterative process should consider only the last or the few last iterations. At least two authors, B. Christianson and A. Griewank, have studied mathematically fixed-point iterations with the goal of defining an efficient adjoint. In this paper, we describe and contrast these two strategies with the objective of implementing the best suited one into the AD tool that we are developing. We select a representative application to test the chosen strategy, to propose a set of user directives to trigger it, and to discuss the implementation implications in our tool.
We study the problem of decomposing a measured signal as a sum of decaying exponentials. There is a direct connection to sums of these types and positive semi-definite (PSD) Hankel matrices, where the rank of these ma...
详细信息
ISBN:
(纸本)9780992862657
We study the problem of decomposing a measured signal as a sum of decaying exponentials. There is a direct connection to sums of these types and positive semi-definite (PSD) Hankel matrices, where the rank of these matrices equals the number of exponentials. We propose to solve the identification problem by forming an optimization problem with a misfit function combined with a rank penalty function that also ensures the PSD-constraint. This problem is non-convex, but we show that it is possible to compute the minimum of an explicit closely related convexified problem. Moreover, this minimum can be shown to often coincide with the minimum of the original non-convex problem, and we provide a simple criterion that enables to verify if this is the case.
We analyze a fixed-point algorithm for reinforcement learning (RL) of optimal portfolio mean-variance preferences in the setting of multivariate generalized autoregressive conditional-heteroskedasticity (MGARCH) with ...
详细信息
We analyze a fixed-point algorithm for reinforcement learning (RL) of optimal portfolio mean-variance preferences in the setting of multivariate generalized autoregressive conditional-heteroskedasticity (MGARCH) with a small penalty on trading. A numerical solution is obtained using a neural network (NN) architecture within a recursive RL loop. A fixed-point theorem proves that NN approximation error has a big-oh bound that we can reduce by increasing the number of NN parameters. The functional form of the trading penalty has a parameter epsilon > 0 that controls the magnitude of transaction costs. When epsilon is small, we can implement an NN algorithm based on the expansion of the solution in powers of epsilon. This expansion has a base term equal to a myopic solution with an explicit form, and a first-order correction term that we compute in the RL loop. Our expansion-based algorithm is stable, allows for fast computation, and outputs a solution that shows positive testing performance.
暂无评论