The derivation of a supervised training algorithm for a neural network implies the selection of a norm criterion which gives a suitable global measure of the particular distribution of errors. The present paper addres...
详细信息
The derivation of a supervised training algorithm for a neural network implies the selection of a norm criterion which gives a suitable global measure of the particular distribution of errors. The present paper addresses this problem and proposes a correspondence between error distribution at the output of a layered feedforward neural network and L(p) norms. The generalized delta rule is investigated in order to verify how its structure can be modified in order to perform a minimization in the generic L(p) norm. Finally, the particular case of the Chebyshev norm is developed and tested.
Popular state estimation techniques in industry are mostly based on the weighted least squares (WLS) method and its derivatives. These estimators usually detect and identify multiple gross measurement errors by repeat...
详细信息
Popular state estimation techniques in industry are mostly based on the weighted least squares (WLS) method and its derivatives. These estimators usually detect and identify multiple gross measurement errors by repeating a cycle of estimation-detection-elimination. It is rather time consuming for large systems. This paper presents a neural network preestimation filter to identify most forms of gross errors, including conforming bad data, in raw measurements before state estimation rather than afterwards. The proposed neural network model is trained to be a measurement estimator by using the correct measurements of typical system operating states. Once trained, the filter quickly identifies most forms of gross measurement errors simultaneously by comparing the square difference of the raw measurements and their corresponding estimated values with some given thresholds. System observability is maintained by replacing bad data with their reasonably accurate estimates. Using the proposed neural network preestimation filter, the efficiency of present state estimators is greatly improved. Results from several case studies are presented.
Supervised neural-network learning algorithms have proven very successful at solving a variety of learning problems. However, they suffer from a common problem of requiring explicit output labels. This requirement mak...
详细信息
Supervised neural-network learning algorithms have proven very successful at solving a variety of learning problems. However, they suffer from a common problem of requiring explicit output labels. This requirement makes such algorithms implausible as biological models. In this paper, it is shown that pattern classification can be achieved, in a multilayered feedforward neural network, without requiring explicit output labels, by a process of supervised self-coding. The class projection is achieved by optimizing appropriate within-class uniformity, and between-class discernibility criteria, The mapping function and the class labels are developed together, iteratively using the derived self-coding backpropagation algorithm. The ability of the self-coding network to generalize on unseen data is also experimentally evaluated on real data sets, and compares favorably with the traditional labeled supervision with neural networks, However, interesting features emerge out of the proposed self-coding supervision, which are absent in conventional approaches. The further implications of supervised self-coding with neural networks are also discussed.
In the backpropagation algorithm, the error calculated from the output of the neural network should backpropagate the layers to update the weights of each layer, making it difficult to parallelize the training process...
详细信息
In the backpropagation algorithm, the error calculated from the output of the neural network should backpropagate the layers to update the weights of each layer, making it difficult to parallelize the training process and requiring frequent off-chip memory access. Local learning algorithms locally generate error signals which are used for weight updates, removing the need for backpropagation of error signals. However, prior works rely on large, complex auxiliary networks for reliable training, which results in large computational overhead and undermines the advantages of local learning. In this work, we propose a local learning algorithm that significantly reduces computational complexity as well as improves training performance. Our algorithm combines multiple consecutive layers into a block and performs local learning on a block-by-block basis, while dynamically changing block boundaries during training. In experiments, our approach achieves 95.68% and 79.42% test accuracy on the CIFAR-10 and CIFAR-100 datasets, respectively, using a small fully connected layer as auxiliary networks, closely matching the performance of the backpropagation algorithm. Multiply-accumulate (MAC) operations and off-chip memory access also reduce by up to 15% and 81% compared to backpropagation.
In this brief paper, the Real Time Recurrent Learning (RTRL) algorithm for training fully recurrent neural networks in real time, is extended for the case of a recurrent neural network whose inputs, outputs, weights a...
详细信息
In this brief paper, the Real Time Recurrent Learning (RTRL) algorithm for training fully recurrent neural networks in real time, is extended for the case of a recurrent neural network whose inputs, outputs, weights and activation functions are complex. A practical definition of the complex activation function is adopted and the complex form of the conventional RTRL algorithm is derived. The performance of the proposed algorithm is demonstrated with an application in complex communication channel equalization.
In a previous paper, neural networks were able to correct color errors in television receivers. The results showed that 70% of these errors were reduced by various amounts. This paper will further investigate the use ...
详细信息
In a previous paper, neural networks were able to correct color errors in television receivers. The results showed that 70% of these errors were reduced by various amounts. This paper will further investigate the use of neural networks for error reduction. A simpler network will be employed in conjunction with a recently developed fast training algorithm. The new results are significantly better than previously reported. It will be shown that virtually all of the color errors will be significantly reduced. This paper will also present the details of the training algorithm.
The extended Kalman filter (EKF) is well known as a state estimation method for a nonlinear system, and can be used as a parameter estimation method by augmenting the state with unknown parameters. A multilayered neur...
详细信息
The extended Kalman filter (EKF) is well known as a state estimation method for a nonlinear system, and can be used as a parameter estimation method by augmenting the state with unknown parameters. A multilayered neural network is a nonlinear system having a layered structure, and its learning algorithm is regarded as parameter estimation for such a nonlinear system. In this paper, a new real-time learning algorithm for a multilayered neural network is derived from the EKF. Since this EKF-based learning algorithm approximately gives the minimum variance estimate of the linkweights, the convergence performance is improved in comparison with the backwards error propagation algorithm using the steepest descent techniques. Furthermore, tuning parameters which crucially govern the convergence properties are not included, which makes its application easier. Simulation results for the XOR and parity problems are provided.
A new approach to intelligent prediction and control of manufacturing processes is studied by the use of neural networks in this paper. Based on a back-propagation mechanism a learning algorithm for multilayer neural ...
详细信息
A new approach to intelligent prediction and control of manufacturing processes is studied by the use of neural networks in this paper. Based on a back-propagation mechanism a learning algorithm for multilayer neural networks, the ''one-by-one algorithm'', is presented which is particularly suitable for forecasting and process control. The differences between the one-by-one algorithm and the back-propagation algorithm in current use are clarified. Then an intelligent forecasting and control architecture for a leadscrew grinding process using the one-by-one algorithm is discussed.
We give a unified treatment of gradient descent learning algorithms for neural networks using a general framework of dynamical systems, This general approach organizes and simplifies all the known algorithms and resul...
详细信息
We give a unified treatment of gradient descent learning algorithms for neural networks using a general framework of dynamical systems, This general approach organizes and simplifies all the known algorithms and results which have been originally derived for different problems (fixed point/trajectory learning), for different model (discrete/continuous), for different architectures (forward/recurrent), and using different techniques (backpropagation, variational calculus, adjoint methods, etc.). The general approach can also be applied to derive new algorithms. We then briefly examine some of the complexity issues and limitations intrinsic to gradient descent learning. Throughout the paper, we focus on the problem of trajectory learning.
In this paper, we address the problem of robustness in multilayer perceptrons. We present the main theoretical results in the case of linear neural networks with one hidden layer in order to go beyond the empirical st...
详细信息
In this paper, we address the problem of robustness in multilayer perceptrons. We present the main theoretical results in the case of linear neural networks with one hidden layer in order to go beyond the empirical study we previously made. We show that the robustness can greatly be improved and that even without decreasing performance in normal use. Finally, we show how this behavior, clearly demonstrated in the linear case, is an approximation of the behavior of nonlinear networks.
暂无评论