Non-negative matrix factorization (NMF) as a part-based representation method allows only additive combinations of non-negative basis components to represent the original data, so it provides a realistic approximation...
详细信息
ISBN:
(纸本)9781424410651
Non-negative matrix factorization (NMF) as a part-based representation method allows only additive combinations of non-negative basis components to represent the original data, so it provides a realistic approximation to the original data. However, NMF does not work well when directly applied to face recognition due to its global linear decomposition;this intuitively results in a degradation of recognition performance and non-robustness to the variation in illumination, expression and occlusion. In this paper, we propose a robust method, random subspace sub-pattern NMF (RS-SpNMF), especially for face recognition. Unlike the traditional random subspace method (RSM), which completely randomly selects the features from the whole original pattern feature set, the proposed method randomly samples,features from each local region (or a sub-image) partitioned from the original face image and performs NMF decomposition on each sampled feature set. More specially, we first divide a face image into several sub-images in a deterministic way, then construct a component classifier on sampled feature subset from each sub-image set, and finally combine all of component classifiers for the final decision. Experiments on three benchmarks face databases (ORL,Yale and AR) show that the proposed method is effective, especially to the occlusive face image.
Outlier detection has recently become an important problem in many industrial and financial applications. In this paper, a novel unsupervised algorithm for outlier detection with a solid statistical foundation is prop...
详细信息
ISBN:
(数字)9783540734994
ISBN:
(纸本)9783540734987
Outlier detection has recently become an important problem in many industrial and financial applications. In this paper, a novel unsupervised algorithm for outlier detection with a solid statistical foundation is proposed. First we modify a nonparametric density estimate with a variable kernel to yield a robust local density estimation. Outliers are then detected by comparing the local density of each point to the local density of its neighbors. Our experiments performed on several simulated data sets have demonstrated that the proposed approach can outperform two widely used outlier detection algorithms (LOF and LOCI).
In this paper, we introduced new adaptive learning algorithms to extract linear discriminant analysis (LDA) features from multidimensional data in order to reduce the data dimension space. For this purpose, new adapti...
详细信息
ISBN:
(纸本)9783540742586
In this paper, we introduced new adaptive learning algorithms to extract linear discriminant analysis (LDA) features from multidimensional data in order to reduce the data dimension space. For this purpose, new adaptive algorithms for the computation of the square root of the inverse covariance matrix Sigma (-1/2) are introduced. the proof for the convergence of the new adaptive algorithm is given by presenting the related cost function and discussing about its initial conditions. the new adaptive algorithms are used before an adaptive principal component analysis algorithm in order to construct an adaptive multivariate multi-class LDA algorithm. Adaptive nature of the new optimal feature extraction method makes it appropriate for on-line patternrecognition applications. Both adaptive algorithms in the proposed structure are trained simultaneously, using a stream of input data. Experimental results using synthetic and real multi-class multi-dimensional sequence of data, demonstrated the effectiveness of the new adaptive feature extraction algorithm.
Based on all phase theory this paper designed three kinds of true 2-D all phase filter bank (true 2-D APFB), which can be used to decompose and recompose image data in true 2-D directly. If quantification error of fil...
详细信息
ISBN:
(纸本)9781424410651
Based on all phase theory this paper designed three kinds of true 2-D all phase filter bank (true 2-D APFB), which can be used to decompose and recompose image data in true 2-D directly. If quantification error of filters is ignored, the true 2-D APFBs have perfect reconstruction property. To reduce computation, they are implemented in lifting scheme. Simulation has shown that true 2-D APFBs have nicer data compression property. Withthe same compression rate, PSNR of IDCT_AFB7.7 is less 0.7dB at most than Daubechies9/7 wavelet's, for true 2-D APFBs adopt quadtree SPIHT coding method which is suitable for separable 2-D wavelet transform. For true 2-D filter banks, binary tree SPIHT coding should be adopted to gel better performance in compression.
CoreWar is a computer simulation where two programs written in an assembly language called redcode compete in a virtual memory array. these programs are referred to as warriors. Over more than twenty years of developm...
详细信息
ISBN:
(数字)9783540734994
ISBN:
(纸本)9783540734987
CoreWar is a computer simulation where two programs written in an assembly language called redcode compete in a virtual memory array. these programs are referred to as warriors. Over more than twenty years of development a number of different battle strategies have emerged, making it possible to identify different warrior types. Systems for automatic warrior creation appeared more recently, evolvers being the dominant kind. this paper describes an attempt to analyze the output of the CCAI evolver, and explores the possibilities for performing automatic categorization by warrior type using representations based on redcode source, as opposed to instruction execution frequency. Analysis was performed using EM clustering, as well as information gain and gain ratio attribute evaluators, and revealed which mainly brute-force types of warriors were being generated. this, along withthe observed correlation between clustering and the workings of the evolutionary algorithm justifies our approach and calls for more extensive experiments based on annotated warrior benchmark collections.
Advances in wireless and mobile technology flood us with amounts of moving object datathat preclude all means of manual data processing. the volume of data gathered from position sensors of mobile phones, PDAs, or ve...
详细信息
ISBN:
(数字)9783540734994
ISBN:
(纸本)9783540734987
Advances in wireless and mobile technology flood us with amounts of moving object datathat preclude all means of manual data processing. the volume of data gathered from position sensors of mobile phones, PDAs, or vehicles, defies human ability to analyze the stream of input data. On the other hand, vast amounts of gathered data hide interesting and valuable knowledge patterns describing the behavior of moving objects. thus, new algorithms for mining moving object data are required to unearththis knowledge. An important function of the mobile objects management system is the prediction of the unknown location of an object. In this paper we introduce a data mining approach to the problem of predicting the location of a moving object. We mine the database of moving object locations to discover frequent trajectories and movement rules. then, we match the trajectory of a moving object withthe database of movement rules to build a probabilistic model of object location. Experimental evaluation of the proposal reveals prediction accuracy close to 80%. Our original contribution includes the elaboration on the location prediction model, the design of an efficient mining algorithm, introduction of movement rule matching strategies, and a thorough experimental evaluation of the proposed model.
data perturbation with random noise signals has been shown to be useful for data hiding in privacy-preserving data mining. Perturbation methods based on additive randomization allows accurate estimation of the Probabi...
详细信息
ISBN:
(数字)9783540734994
ISBN:
(纸本)9783540734987
data perturbation with random noise signals has been shown to be useful for data hiding in privacy-preserving data mining. Perturbation methods based on additive randomization allows accurate estimation of the Probability Density Function (PDF) via the Expectation-Maximization (EM) algorithm but it has been shown that noise-filtering techniques can be used to reconstruct the original data in many cases, leading to security breaches. In this paper, we propose a generic PDF reconstruction algorithm that can be used on non-additive (and additive) randomization techiques for the purpose of privacy-preserving data mining. this two-step reconstruction algorithm is based on Parzen-Window reconstruction and Quadratic Programming over a convex set - the probability simplex. Our algorithm eliminates the usual need for the iterative EM algorithm and it is generic for most randomization models. the simplicity of our two-step reconstruction algorithm, without iteration, also makes it attractive for use when dealing with large datasets.
Transduction is an inference mechanism "from particular to particular". Its application to classification tasks implies the use of both labeled (training) data and unlabeled (working) data to build a classif...
详细信息
ISBN:
(数字)9783540734994
ISBN:
(纸本)9783540734987
Transduction is an inference mechanism "from particular to particular". Its application to classification tasks implies the use of both labeled (training) data and unlabeled (working) data to build a classifier whose main goal is that of classifying (only) unlabeled data as accurately as possible. Unlike the classical inductive setting, no general rule valid for all possible instances is generated. Transductive learning is most suited for those applications where the examples for which a prediction is needed are already known when training the classifier. Several approaches have been proposed in the literature on building transductive classifiers from data stored in a single table of a relational database. Nonetheless, no attention has been paid to the application of the transduction principle in a (multi-) relational setting, where data are stored in multiple tables of a relational database. In this paper we propose a new transductive classifier, named TRANSC, which is based on a probabilistic approach to making transductive inferences from relational data. this new method works in a transductive setting and employs a principled probabilistic classification in multi-relational data mining to face the challenges posed by some spatial data mining problems. Probabilistic inference allows us to compute the class probability and return, in addition to result of transductive classification, the confidence in the classification. the predictive accuracy of TRANSC has been compared to that of its inductive counterpart in an empirical study involving both a benchmark relational dataset and two spatial datasets. the results obtained are generally in favor of TRANSC, although improvements are small by a narrow margin.
We propose visual tracking of multiple objects (faces of people) in a meeting scenario based on low-level features such as skin-color, target motion, and target size. Based on these features automatic initialization a...
详细信息
ISBN:
(纸本)9783540742586
We propose visual tracking of multiple objects (faces of people) in a meeting scenario based on low-level features such as skin-color, target motion, and target size. Based on these features automatic initialization and termination of objects is performed. Furthermore, on-line learning is used to incrementally update the models of the tracked objects to reflect the appearance changes. For tracking a particle filter is incorporated to propagate sample distributions over time. We discuss the close relationship between our implemented tracker based on particle filters and genetic algorithms. Numerous experiments on meeting data demonstrate the capabilities of our tracking approach.
We propose a method to video segmentation via active learning. Shot segmentation is an essential first step to video segmentation.. the color histogram-based shot boundary detection algorithm is one of the most reliab...
详细信息
ISBN:
(纸本)9780769529295
We propose a method to video segmentation via active learning. Shot segmentation is an essential first step to video segmentation.. the color histogram-based shot boundary detection algorithm is one of the most reliable variants of histogram-based detection algorithms. It is not unreasonable to assume that the color content does not change rapidly within but across shots. thus, we present a metric based on blocked color histogram (BCH) for inter-frame difference. Our metric is the normalized intersection of BCH between contiguous frames. Hard cuts and gradual shot transitions can be detected as valleys in the time series of the differences between color histograms of contiguous frames or of frames a certain distance apart. We try to estimate the valleys on the frame-to-frame difference curve. Each kind of shot transition (Cut or Gradual shot transition) has its own characteristic. pattern corresponding with valleys. therefore shot detection can be viewed as patternrecognition. We employ the support vector machine (SVM) via active learning to classify shot boundaries and non-boundaries Our method is evaluated on the TRECVID benchmarking platform and the experimental results reveal the effectiveness and robustness of the method.
暂无评论