Given a set P of h pairwise disjoint convex polygonal obstacles in the plane, defined with n vertices, we preprocess P and compute one routing table at each vertex in a subset of vertices of P. For routing a packet fr...
详细信息
Given a set P of h pairwise disjoint convex polygonal obstacles in the plane, defined with n vertices, we preprocess P and compute one routing table at each vertex in a subset of vertices of P. For routing a packet from any vertex s ? P to any vertex t ? P, our scheme computes a routing path with a multiplicative stretch 1 + E and an additive stretch 2kg, by consulting routing tables at only a subset of vertices along that path. Here, k is the number of obstacles of P the routing path intersects, and P depends on the geometry of obstacles in P. During the preprocessing phase, we construct routing tables of size O(n + h(3) /?(2) polylog (h/? )) in O(n + h(3) /?(2) polylog (h/? )) time, where ? < 1 is an input parameter.
This letter considers resilient decentralized constrained optimization in multi-agent systems where some agents due to cyberattacks become adversaries. We show that the proposed method is resilient despite the persist...
详细信息
This letter considers resilient decentralized constrained optimization in multi-agent systems where some agents due to cyberattacks become adversaries. We show that the proposed method is resilient despite the persistent influence of up to F anonymous adversaries in the complete graphs. Our approach provides a better approximation of the optimal solution than the current literature. If the agents' objectives are 2F redundant, then the algorithm converges to the optimal solution. In addition to current literature, we consider a constrained optimization problem. Finally, we present numerical simulations to corroborate the theoretical analysis.
During the last decade, classification systems (CSs) received significant research attention, with new learning algorithms achieving high accuracy in various applications. However, their resource-intensive nature, in ...
详细信息
During the last decade, classification systems (CSs) received significant research attention, with new learning algorithms achieving high accuracy in various applications. However, their resource-intensive nature, in terms of hardware and computation time, poses new design challenges. CSs exhibit inherent error resilience, due to redundancy of training sets, and self-healing properties, making them suitable for Approximate Computing (AxC). AxC enables efficient computation by using reduced precision or approximate values, leading to energy, time, and silicon area savings. Exploiting AxC involves estimating the introduced error for each approximate variant found during a Design-Space Exploration (DSE). This estimation has to be both rapid and meaningful, considering a substantial number of test samples, which are utterly conflicting demands. In this paper, we investigate on sources of error resiliency of CSs, and we propose a technique to haste the DSE that reduces the computational time for error estimation by systematically reducing the test set. In particular, we cherry-pick samples that are likely to be more sensitive to approximation and perform accuracy-loss estimation just by exploiting such a sample subset. In order to demonstrate its efficacy, we integrate our technique into two different approaches for generating approximate CSs, showing an average speed-up up to 18. Authors
In view of the fact that impulsive differential equations have the discreteness due to the impulse phenomenon, this article proposes a single hidden layer neural networkmethod-based extreme learning machine and a phys...
详细信息
In view of the fact that impulsive differential equations have the discreteness due to the impulse phenomenon, this article proposes a single hidden layer neural networkmethod-based extreme learning machine and a physics-informed neural network method combined with learning rate attenuation strategy to solve linear impulsive differential equations and nonlinear impulsive differential equations, respectively. For the linear impulsive differential equations, first, the interval is segmented according to the impulse points, and a single hidden layer neural network model is constructed, the weight parameters of the hidden layer are randomly set, the optimal output parameters, and solution of the first segment are obtained by the extreme learning machine algorithm, then we calculate the initial value of the second segment according to the jumping equation and the remaining segments are solved in turn in the same way. Although the single hidden layer neural network method proposed can solve linear equations with high accuracy, it is not suitable for solving nonlinear equations. Therefore, we propose the physics-informed neural network combined with a learning rate attenuation strategy to solve the nonlinear impulsive differential equations, then the Adam algorithm and L-BFGS algorithm are combined to find the optimal approximate solution of each segment. Numerical examples show that the single hidden layer neural network method with Legendre polynomials as the activation function and the physics-informed neural network method combined with learning rate attenuation strategy can solve linear and nonlinear impulsive differential equations with higher accuracy. Impact Statement-It is difficult to obtain the analytical solutions of impulsive differential equations because of the existence of impulse points, and the current numerical methods are complicated and demanding. In recent years, artificial neural network methods have been widely used due to its simplicity and efficie
The double linear transformation model Y = AXB + W plays an important role in a variety of science and engineering applications, where X is estimated through known transformation matrices A and B from the noisy measur...
详细信息
The double linear transformation model Y = AXB + W plays an important role in a variety of science and engineering applications, where X is estimated through known transformation matrices A and B from the noisy measurement Y. Decoupling X from Y is a formidable task due to the high complexity brought by the multiplication of the unknown matrix (vector) with the transformation matrix (M-UMTM). Unitary approximate message passing (UAMP) has been verified as a low complexity and strong robustness solution to the M-UMTM problems. However, it has only been used to tackle the problems with a single linear transformation matrix. In this work, we develop a generalized algorithm, namely, generalized double UAMP (GD-UAMP) for the target model, which not only inherits the low complexity of AMP, but also enhances robustness by employing double unitary transformation. As a generalized algorithm, GD-UAMP can be applied to address the generalized Bayesian inference problem, i.e., the arbitrary prior probability of X and likelihood function of Z, where Z = AXB is the noiseless measurement. We verify the feasibility of the proposed algorithm in the channel estimation problem for various wireless communication systems. Numerical results demonstrate that the proposed algorithm can perfectly fit different scenarios and showcase superior performance compared with benchmarks.
Many classical and modern machine learning algorithms require solving optimization tasks under orthogonality constraints. Solving these tasks with feasible methods requires a gradient descent update followed by a retr...
详细信息
Many classical and modern machine learning algorithms require solving optimization tasks under orthogonality constraints. Solving these tasks with feasible methods requires a gradient descent update followed by a retraction operation on the Stiefel manifold, which can be computationally expensive. Recently, an infeasible retraction-free approach, termed the landing algorithm, was proposed as an efficient alternative. Motivated by the common occurrence of orthogonality constraints in tasks such as principle component analysis and training of deep neural networks, this letter studies the landing algorithm and establishes a novel linear convergence rate for smooth non-convex functions using only a local Riemannian P & Lstrok;condition. Numerical experiments demonstrate that the landing algorithm performs on par with the state-of-the-art retraction-based methods with substantially reduced computational overhead.
Minimum area polygonization is a well-known NP-complete problem in computational geometry. It is the problem of finding a simple polygon with minimum area for a given set of points in the plane. We present a parallel ...
详细信息
Biological neural networks are equipped with an inherent capability to continuously adapt through online learning. This aspect remains in stark contrast to learning with error backpropagation through time (BPTT) that ...
详细信息
Biological neural networks are equipped with an inherent capability to continuously adapt through online learning. This aspect remains in stark contrast to learning with error backpropagation through time (BPTT) that involves offline computation of the gradients due to the need to unroll the network through time. Here, we present an alternative online learning algorithm framework for deep recurrent neural networks (RNNs) and spiking neural networks (SNNs), called online spatio-temporal learning (OSTL). It is based on insights from biology and proposes the clear separation of spatial and temporal gradient components. For shallow SNNs, OSTL is gradient equivalent to BPTT enabling for the first time online training of SNNs with BPTT-equivalent gradients. In addition, the proposed formulation unveils a class of SNN architectures trainable online at low time complexity. Moreover, we extend OSTL to a generic form, applicable to a wide range of network architectures, including networks comprising long short-term memory (LSTM) and gated recurrent units (GRUs). We demonstrate the operation of our algorithm framework on various tasks from language modeling to speech recognition and obtain results on par with the BPTT baselines.
We initiate a broad study of classical problems in the streaming model with insertions and deletions in the setting where we allow the approximation factor α to be much larger than 1. Such algorithms can use signific...
详细信息
We initiate a broad study of classical problems in the streaming model with insertions and deletions in the setting where we allow the approximation factor α to be much larger than 1. Such algorithms can use significantly less memory than the usual setting for which α = 1 + ǫ for an ǫ ∈ (0, 1), and are motivated by applications to data-driven algorithm design, among other things. We study large approximations for a number of problems in sketching and streaming, assuming that the underlying n-dimensional vector has all coordinates bounded by M throughout the data stream: 1. For the p norm1, 0 Θ(1), which holds even for randomly ordered streams or for streams in the bounded deletion model. Our lower bound also holds for a large class of statistical M -estimators. We also give a 2-pass algorithm that uses less space than the best existing 1-pass algorithm when the entries of the vector are small. 2. For estimating the p norm, p > 2, we show an upper bound of O(n1−2/p(log n log M)/α2) bits for an α-approximation, and give a matching lower bound, for almost the full range of α ≥ 1 for linear sketches. We use this to design algorithms with large approximation factors for cascaded norms and rectangle p norms. 3. For the 2-heavy hitters problem, we show that the known lower bound of Ω(k log n log M) bits for identifying (1/k)-heavy hitters holds even if we are allowed to output items that are 1/(αk)-heavy, for almost the full range of α, provided the algorithm succeeds with probability 1 − O(1/n). We also obtain a lower bound for linear sketches that is tight even for constant probability algorithms. 4. For estimating the number 0 of distinct elements, we give an n1/t-approximation algorithm using O(t log log M) bits of space, as well as a lower bound of Ω(t) bits, both excluding the storage of random bits, where n is the dimension of the underlying frequency vector and M is an upper bound on the magnitude of its coordinates. We also show a separation between 1 and 2 passes
The combination of model predictive control (MPC) and learning methods has been gaining increasing attention as a tool to control systems that may be difficult to model. Using MPC as a function approximator in reinfor...
详细信息
The combination of model predictive control (MPC) and learning methods has been gaining increasing attention as a tool to control systems that may be difficult to model. Using MPC as a function approximator in reinforcement learning (RL) is one approach to reduce the reliance on accurate models. RL is dependent on exploration to learn, and currently, simple heuristics based on random perturbations are most common. This paper considers variance-based exploration in RL geared towards using MPC as function approximator. We propose to use a non-probabilistic measure of uncertainty of the value function approximator in value-based RL methods. Uncertainty is measured by a variance estimate based on inverse distance weighting (IDW). The IDW framework is computationally cheap to evaluate and therefore well-suited in an online setting, using already sampled state transitions and rewards. The gradient of the variance estimate is then used to perturb the policy parameters in a direction where the variance of the value function estimate is increasing. The proposed method is verified on two simulation examples, considering both linear and nonlinear system dynamics, and compared to standard exploration methods using random perturbations.
暂无评论