The Semantic Web is a project and vision of the World Wide Web Consortium to extend the current Web, so that information is given well-defined meaning and structure, enhancing computers and people to work in cooperati...
详细信息
ISBN:
(纸本)9788988678183
The Semantic Web is a project and vision of the World Wide Web Consortium to extend the current Web, so that information is given well-defined meaning and structure, enhancing computers and people to work in cooperation. Semantic technologies are being added to enterprise solutions to accommodate new techniques for discovering relationships across different database, business applications and Web services. In this paper, we present an architectural model for a software tool which combines the use of Semantic Web mechanisms with database metadata and data warehousing mechanisms. If the benefits of the Semantic Web concept are combined with a powerful database server, then the information management will be much improved.
Nowadays, we have to deal with a large quantity of unstructured, heterogeneous data, produced by an increasing number of sources. Clustering heterogeneous data is essential to getting structured information in respons...
详细信息
ISBN:
(纸本)9781424473359
Nowadays, we have to deal with a large quantity of unstructured, heterogeneous data, produced by an increasing number of sources. Clustering heterogeneous data is essential to getting structured information in response to user queries. In this paper, we assess the results of a new clustering technique - clustering by compression - when applied to metadata associated with heterogeneous sets of data. The clustering by compression procedure is based on a parameter-free, universal, similarity distance, the normalized compression distance or NCD, computed from the lengths of compressed data files (singly and in pair-wise concatenation). Experimental results show that using metadata could improve the average clustering performances with about 20% over clustering the same sample data set without using metadata.
The proposed method is designed for a data acquisition system acquiring data from n independent sources. The data sources are supposed to produce fragments that together constitute some logical wholeness. These fragme...
The proposed method is designed for a data acquisition system acquiring data from n independent sources. The data sources are supposed to produce fragments that together constitute some logical wholeness. These fragments are produced with the same frequency and in the same sequence. The discussed algorithm aims to balance the data dynamically between m logically autonomous processing units (consisting of computing nodes) in case of variation in their processing power which could be caused by some faults like failing computing nodes, or broken network connections. As a case study we consider the Data Acquisition System of the Compact Muon Solenoid Experiment at CERN's new Large Hadron Collider. The system acquires data from about 500 sources and combines them into full events. Each data source is expected to deliver event fragments of an average size of 2 kB with 100 kHz frequency. In this paper we present the results of applying proposed load metric and load communication pattern. Moreover, we discuss their impact on the algorithm's overall efficiency and scalability, as well as on fault tolerance of the whole system. We also propose a general concept of an algorithm that allows for choosing the destination processing unit in all source nodes asynchronously and asserts that all fragments of same logical data always go to same unit.
This paper tries to define mechanisms which insure the right functionality for systems based on the Volunteer Grid concept and propose a new framework for Volunteer Grid computing management. Volunteer Grid systems al...
详细信息
ISBN:
(纸本)9781424473359
This paper tries to define mechanisms which insure the right functionality for systems based on the Volunteer Grid concept and propose a new framework for Volunteer Grid computing management. Volunteer Grid systems allow setting up high performance computer networks, easily, rapidly and at low costs, the main characteristic of their nodes being the volunteer participation. Thus, a supercomputer is created, able to perform most complex calculations in a relatively short period of time. This volunteering itself is a weakness of the system, because not only physical faults must be taken into consideration, but also sabotages through which participants could try to increase their rating.
The current trend in processor's design is to add multiple cores to increase the system's overall performance but this is not a solution to increasing the performance of serial applications. Due to its potenti...
详细信息
ISBN:
(纸本)9781424473359
The current trend in processor's design is to add multiple cores to increase the system's overall performance but this is not a solution to increasing the performance of serial applications. Due to its potential to greatly accelerate a wide variety of serial applications, reconfigurable computing has become a subject of a great deal of research. Its key feature is the ability to perform computations in hardware in order to increase performance, while retaining much of the flexibility of a software solution. In this paper, we address the problem of fully automating the process of selecting the code to be used for hardware acceleration. We present a software-hardware partitioning system that transforms Impulse C source code into blocks of C and VHDL code. The resulting C code wiD be run on the CPU, while the VHDL code will be implemented on a reconfigurable hardware, e.g. a FPGA.
In the paper is presented the brief history of research conducted in Poland (at AGH University) in the area of image processing, analysis and recognition. The history is connected with changes of the technology used f...
详细信息
In this paper we address the topic of embedding 2D barcode detection and reading technology into specific online environment applications. We propose a framework for the integration of barcode technology into common u...
详细信息
The paper presents a software application for modeling, simulation and analysis of urban traffic networks behavior in different functioning contexts. A module of control techniques, based on different approaches (as g...
详细信息
Current Web search engines return long lists of ranked documents that users are forced to sift through to find relevant documents. This paper introduces a new approach for clustering Web search results, based on the n...
详细信息
ISBN:
(纸本)9789898425225
Current Web search engines return long lists of ranked documents that users are forced to sift through to find relevant documents. This paper introduces a new approach for clustering Web search results, based on the notion of clustering by compression. Compression algorithms allow defining a similarity measure based on the degree of common information. Classification methods allow clustering similar data without any previous knowledge. The clustering by compression procedure is based on a parameter-free, universal, similarity distance, the normalized compression distance or NCD, computed from the lengths of compressed data files. Our goal is to apply the clustering by compression algorithm in order to cluster the documents returned by a Web search engine in response to a user query.
Skew detection and correction is an important step in automated content conversion systems, on which overall system performance is dependent. Although there are many working solutions at the present time, the search f...
详细信息
暂无评论