the goal of statistical pattern feature extraction (SPFE) is 'low loss dimension reduction'. As the key link of patternrecognition, dimension reduction has become the research hot spot and difficulty in the f...
详细信息
ISBN:
(纸本)9783540859833
the goal of statistical pattern feature extraction (SPFE) is 'low loss dimension reduction'. As the key link of patternrecognition, dimension reduction has become the research hot spot and difficulty in the fields of patternrecognition, machinelearning, datamining and so on. pattern feature extraction is one of the most challenging research fields and has attracted the attention from many scholars. this paper summarily introduces the basic principle of SPFE, and discusses the latest progress of SPFE from the aspects such as classical statistical theories and their modifications, kernel-based methods, wavelet analysis and its modifications, algorithms integration and so on. At last we discuss the development trend of SPFE.
Over the last couple of months a large number of Distributed Denial of Service (DDoS) attacks have occurred across the world, especially targeting those who provide web services. IP traceback a counter measure against...
详细信息
ISBN:
(纸本)9781424442362
Over the last couple of months a large number of Distributed Denial of Service (DDoS) attacks have occurred across the world, especially targeting those who provide web services. IP traceback a counter measure against DDoS, is the ability to trace IP packets back to the true source/s of the attack. In this paper, an IP traceback scheme using a machinelearning technique called Intelligent Decision Prototype (IDP), is proposed. IDP can be used on both Probabilistic Packet Marking (PPM) and Deterministic Packet Marking (DPM) traceback schemes to identify DDoS attacks. this will greatly reduce the packets that are marked and in effect make the system more efficient and effective at tracing the source of an attack compared with other methods. IDP can be applied to many security systems such as datamining, Forensic Analysis, Intrusion Detection Systems (IDS) and DDoS defense systems.
In this talk we review the recent work done by our group on datamining (DM) technologies deduced from simulating visual principle. through viewing a DM problem as a cognition problems and treading a data set as an im...
详细信息
ISBN:
(数字)9780387876856
ISBN:
(纸本)9780387876849
In this talk we review the recent work done by our group on datamining (DM) technologies deduced from simulating visual principle. through viewing a DM problem as a cognition problems and treading a data set as an image with each light point located at a datum position, we developed a series of high efficient algorithms for clustering, classification and regression via mimicking visual principles. In patternrecognition, human eyes seem to possess a singular aptitude to group objects and find important structure in an efficient way. thus, a DM algorithm simulating visual system may solve some basic problems in DM research. From this point of view, we proposed a new approach for data clustering by modeling the blurring effect of lateral retinal interconnections based on scale space theory. In this approach, as the data image blurs, smaller light blobs merge into large ones until the whole image becomes one light blob at a low enough level of resolution. By identifying each blob with a cluster, the blurring process then generates a family of clustering along the hierarchy. the proposed approach provides unique solutions to many long standing problems, such as the cluster validity and the sensitivity to initialization problems, in clustering. We extended such an approach to classification and regression problems, through combatively employing the Weber's law in physiology and the cell response classification facts. the resultant classification and regression algorithms are proven to be very efficient and solve the problems of model selection and applicability to huge size of data set in DM technologies. We finally applied the similar idea to the difficult parameter setting problem in support vector machine (SVM). Viewing the parameter setting problem as a recognition problem of choosing a visual scale at which the global and local structures of a data set can be preserved, and the difference between the two structures be maximized in the feature space, we derived a di
A modified version for semi-supervised learning algorithm with local and global consistency was proposed in this paper. the new method adds the label information, and adopts the geodesic distance rather than Euclidean...
详细信息
ISBN:
(纸本)9781424421749
A modified version for semi-supervised learning algorithm with local and global consistency was proposed in this paper. the new method adds the label information, and adopts the geodesic distance rather than Euclidean distance as the measure of the difference between two data points when conducting calculation. In addition we add class prior knowledge. It was found that the effect of class prior knowledge was different between under high label rate and low label rate. the experimental results show that the changes attain the satisfying classification performance better than the original algorithms.
Associative classifiers that utilize association rules have been widely studied. It has been shown that associative classifiers often outperform traditional classifiers. Associative classifiers usually find only rules...
详细信息
Associative classifiers that utilize association rules have been widely studied. It has been shown that associative classifiers often outperform traditional classifiers. Associative classifiers usually find only rules with high support values, because reducing the minimum support to be satisfied increases computational cost. However, rules with low support but high confidence may contribute to classification. We have proposed an approach to build a classifier composed of almost all consistent (100% confident) rules. the proposed classifier was extended by introducing item reduction and bagging in order to relax the constraint of consistency, which resulted in slightly increased performance for 26datasets from the UCI machinelearning repository.
In real-world digital libraries, artificial intelligence techniques are essential for tackling the automatic document processing task with sufficient flexibility. the great variability in document kind, content and sh...
详细信息
ISBN:
(纸本)9781424421749
In real-world digital libraries, artificial intelligence techniques are essential for tackling the automatic document processing task with sufficient flexibility. the great variability in document kind, content and shape requires powerful representation formalisms to catch all the domain complexity. the continuous flow of new documents requires adaptable techniques that can progressively adjust the acquired knowledge on documents as long as new evidence becomes available, even extending if needed the set of recognized document types. Boththese issues have not yet been thoroughly studied. this paper presents an incremental first-order logic learning framework for automatically dealing with various kinds of evolution in digital repositories content: evolution in the definition of class definitions, evolution in the set of known classes and evolution by addition of new unknown classes. Experiments show that the approach can be applied to real-world.
In this paper we propose a confidence rated boosting algorithm based on Ada-boost for generic object detection. Confidence rated Ada-boost algorithm has not been applied to generic object detection problem; in that se...
详细信息
ISBN:
(纸本)9781424421749
In this paper we propose a confidence rated boosting algorithm based on Ada-boost for generic object detection. Confidence rated Ada-boost algorithm has not been applied to generic object detection problem; in that sense our work is novel. We represent images as bag of words, where the words are SIFT descriptors extracted over some interest points. We compare our boosting algorithm to another version of boosting algorithm called Gentle-boost. Our approach generalizes well and performs equal or better than Gentle-boost. We show our results on four categories from the Caltech data sets, in terms of ROC curves.
Face recognition is nowadays one of the most challenging biometric modalities for the identification of individuals. In the last two decades several experimental as well as commercial systems have been developed explo...
详细信息
Face recognition is nowadays one of the most challenging biometric modalities for the identification of individuals. In the last two decades several experimental as well as commercial systems have been developed exploiting different physical properties of the face image. Either being based on processing 2D or 3D information all these methods perform a face classification of the individuals based on some relevant features extracted from the raw image data. the data acquisition, preprocessing and the feature extraction/selection are all topics of the greatest importance to design a good performing recognition system. At the same time, the right choice of the features to be used as the basis for the face representation, which must be based on the uniqueness of such features, as well as most advanced issues such as the incorporation of quality information and the cope for ageing effects, are all of paramount importance. the tutorial will consists of two sessions (half day of total duration) devoted to the description of boththe basic and most advanced techniques related to face recognition. the lectures will provide a comprehensive outline of face- based biometrics, its relation to biological systems (the psychophysics of the human visual system), including the existing applications and commercial systems. the lectures will provide an in-depth analysis of the state-of-the-art algorithms for face-image analysis including: face detection and tracking, landmark localization, feature extraction, face representation and classification. the lectures will mainly explore the image processing aspects of the recognition process. As for classification, machinelearning algorithms will be also presented, including kernel methods as related to learning and the approximation theory. the most relevant issues and problems will be raised, providing practical solutions and algorithms responding to them. Particular attention will be given to the most advanced and new techniques for face
In this paper, we propose sparse non-negative patternlearning (SNPL) based on self-taught learning framework. In the algorithm, visual patterns are first learned from unlabeled data by non-negative matrix approximati...
详细信息
In this paper, we propose sparse non-negative patternlearning (SNPL) based on self-taught learning framework. In the algorithm, visual patterns are first learned from unlabeled data by non-negative matrix approximation with sparseness constraints, and then features are extracted by the second part of the algorithm, a conjugate family based non-negative sparse feature extraction method. By combining sparse and non-negative constraints of patterns together, SNPL model gives a better representation for images than state-of-art methods. Beyond that, we give an analytical solution for feature extraction although it is approximate, and thereby we extract the features for self-taught learning framework in a faster and more stable way. We apply the new model to various areas, including pattern coding, feature extraction, and recognition. Experimental results show the advantages of SNPL model.
暂无评论