Traditional machine-learning algorithms are struggling to handle the exceedingly large amount of data being generated by the internet. In real-world applications, there is an urgent need for machine-learning algorithm...
详细信息
Traditional machine-learning algorithms are struggling to handle the exceedingly large amount of data being generated by the internet. In real-world applications, there is an urgent need for machine-learning algorithms to be able to handle large-scale, high-dimensional text data. Cloud computing involves the delivery of computing and storage as a service to a heterogeneous community of recipients, Recently, it has aroused much interest in industry and academia. Most previous works on cloud platforms only focus on the parallel algorithms for structured data. In this paper, we focus on the parallel implementation of web-mining algorithms and develop a parallel web-mining system that includes parallel web crawler; parallel text extract, transform and load (ETL) and modeling; and parallel text mining and application subsystems. The complete system enables variable real-world web-mining applications for mass data.
Cross-media is the outstanding characteristics of the age of big data with large scale and complicated processing task. This article presents 5 issues and briefly summarizes the research progress of cross-media knowle...
详细信息
Cross-media is the outstanding characteristics of the age of big data with large scale and complicated processing task. This article presents 5 issues and briefly summarizes the research progress of cross-media knowledge discovery. Furthermore, we propose a framework for cross-media semantic understanding which contains discriminative modeling, generative modeling and cognitive modeling. In cognitive modeling, a new model entitled CAM is proposed which is suitable for cross-media semantic understanding. Moreover, a Cross-Media intelligent Retrieval System (CMIRS) will be illustrated. In the final, the research directions and problems encountered are presented.
PLSA(Probabilistic Latent Semantic Analysis) is a popular topic modeling technique for exploring document collections. Due to the increasing prevalence of large datasets, there is a need to improve the scalability of ...
详细信息
Concept learning in information systems is actually performed in knowledge granular space on information systems. But no much attention has been paid to study such a knowledge granular space and its structure so far, ...
详细信息
Concept learning in information systems is actually performed in knowledge granular space on information systems. But no much attention has been paid to study such a knowledge granular space and its structure so far, and its structure characteristics are still poorly understood. In this paper, the granular space is firstly topologized and is decomposed into granular worlds. Then it is modeled as a bounded lattice. Finally, by using graph theory, the bounded lattice obtained is expressed as a hass graph, and the mechanism of concept learning in information systems can be visually explained. With related properties of topological space, bounded lattice and graph theory, the "mysterious" granular space can be delved more deeply into. This work can form a basis for designing concept learning algorithm as well as can richen the theory system for granular computing.
This paper presents a novel WQA (Web Question Answering) approach based on the combination CCG (Combinatory Categorial Grammar) and DL (Description Logic) ontology, in order to promote semantic-level accuracy through ...
详细信息
ISBN:
(纸本)9781467312882
This paper presents a novel WQA (Web Question Answering) approach based on the combination CCG (Combinatory Categorial Grammar) and DL (Description Logic) ontology, in order to promote semantic-level accuracy through deep text understanding capabilities. We propose to take DL based semantic modeling, i.e., translating lambda-expression encoding of question meaning into DL based semantic representations. The advantage of such approach is a seamless exploitation of existing semantic resource coded as DL ontology, which is widespread in such area as the Semantic Web and conceptual modeling. The experiments are conducted with a repository of complex Chinese questions which involves the satisfaction of some object property restrictions. The experimental results show that producing the semantic representations with the combination of CCG parsing and DL reasoning is an effective approach for question understanding at semantic level, in terms of both understanding accuracy promotion and semantic resource exploitation.
We present a hierarchical chunk-to-string translation model, which can be seen as a compromise between the hierarchical phrase-based model and the tree-to-string model, to combine the merits of the two models. With th...
ISBN:
(纸本)9781622761715
We present a hierarchical chunk-to-string translation model, which can be seen as a compromise between the hierarchical phrase-based model and the tree-to-string model, to combine the merits of the two models. With the help of shallow parsing, our model learns rules consisting of words and chunks and meanwhile introduce syntax cohesion. Under the weighed synchronous context-free grammar defined by these rules, our model searches for the best translation derivation and yields target translation simultaneously. Our experiments show that our model significantly outperforms the hierarchical phrase-based model and the tree-to-string model on English-Chinese Translation tasks.
We study the visual learning models that could work efficiently with little ground-truth annotation and a mass of noisy unlabeled data for large scale Web image applications, following the subroutine of semi-supervise...
详细信息
We study the visual learning models that could work efficiently with little ground-truth annotation and a mass of noisy unlabeled data for large scale Web image applications, following the subroutine of semi-supervised learning (SSL) that has been deeply investigated in various visual classification tasks. However, most previous SSL approaches are not able to incorporate multiple descriptions for enhancing the model capacity. Furthermore, sample selection on unlabeled data was not advocated in previous studies, which may lead to unpredictable risk brought by real-world noisy data corpse. We propose a learning strategy for solving these two problems. As a core contribution, we propose a scalable semi-supervised multiple kernel learning method (S 3 MKL) to deal with the first problem. The aim is to minimize an overall objective function composed of log-likelihood empirical loss, conditional expectation consensus (CEC) on the unlabeled data and group LASSO regularization on model coefficients. We further adapt CEC into a group-wise formulation so as to better deal with the intrinsic visual property of real-world images. We propose a fast block coordinate gradient descent method with several acceleration techniques for model solution. Compared with previous approaches, our model better makes use of large scale unlabeled images with multiple feature representation with lower time complexity. Moreover, to address the issue of reducing the risk of using unlabeled data, we design a multiple kernel hashing scheme to identify the “informative” and “compact” unlabeled training data subset. Comprehensive experiments are conducted and the results show that the proposed learning framework provides promising power for real-world image applications, such as image categorization and personalized Web image re-ranking with very little user interaction.
Software testing is the key validation technique used by industry up to today, but remain error prone and expensive cost. Automatically generating test cases from formal models of the system under test is a promising ...
详细信息
Software testing is the key validation technique used by industry up to today, but remain error prone and expensive cost. Automatically generating test cases from formal models of the system under test is a promising improvement approach to cut down the testing cost. This paper introduces a technique that automatically generate real-time conformance test cases from timed automata specifications. First, both reactive system and its environment is modeled by restricted automata with the notion of deterministic, input enabled and output urgent. Then demonstration is given to show how to efficiently generate real-time test cases with optimal execution time from diagnostic trace. Finally, we formally specify user's single purpose or coverage criteria to convert the test case generation problem into a reachability problem. This approach is implemented using model checkers as test case generation tools and experiment results on three different coverage criteria specifications show feasibility and effectiveness of our technique.
This research aims to evaluate the internal structure of concrete material configuration using an immersed ultrasonic computed tomography imaging technique. We propose a relative difference method of time of flight da...
详细信息
Using Particle Swarm Optimization to handle complex functions with high-dimension it has the problems of low convergence speed and sensitivity to local convergence. The convergence of particle swarm algorithm is studi...
详细信息
Using Particle Swarm Optimization to handle complex functions with high-dimension it has the problems of low convergence speed and sensitivity to local convergence. The convergence of particle swarm algorithm is studied from the dynamic system theory, and the condition for the convergence of particle swarm algorithm is given. The analysis provided qualitative guidelines for the general algorithm parameter selection. Results of numerical tests show the efficiency of the results.
暂无评论