This article offers basic principles for use in determining how "relational" a DBMS product is — a challenge which faces many buyers today, because almost every vendor is claiming that his DBMS product is &...
详细信息
This article offers basic principles for use in determining how "relational" a DBMS product is — a challenge which faces many buyers today, because almost every vendor is claiming that his DBMS product is "relational". The material is likely to be of interest to vendors also. Some of them may not realize how far from the mark they are.
In a growing number of information processing applications, data takes the form of continuous data streams rather than traditional stored databases. Monitoring systems that seek to provide monitoring services in cloud...
详细信息
ISBN:
(纸本)9781467352291
In a growing number of information processing applications, data takes the form of continuous data streams rather than traditional stored databases. Monitoring systems that seek to provide monitoring services in cloud environment must be prepared to deal gracefully with huge data collection without compromising system performance. In this paper, we show that by using a concept of urgent data, system can shorten the response time for most `urgent' queries while guarantee lower bandwidth consumption. We argue that monitoring data can be treated differently. Some data capture critical system events, the arrival of these data will significantly influence the monitoring reaction speed, we call them urgent data. High speed urgent data collection would help system to act in real time when facing fatal error. On the other hand, slowing down the collection speed of others may render more bandwidth. Then several urgent data collection strategies that focus on reducing the urgent data volume are also proposed and evaluated to guarantee the efficiency.
Provides an abstract of the keynote presentation and may include a brief professional biography of the presenter. The complete presentation was not made available for publication as part of the conference proceedings.
ISBN:
(纸本)9781538655566;9781538655559
Provides an abstract of the keynote presentation and may include a brief professional biography of the presenter. The complete presentation was not made available for publication as part of the conference proceedings.
Given the rise of ubiquitous computing and communication devices like biosensors, smart watch, and smartphones, real-time online systems can provide users with a wide range of supports including monitoring daily activ...
详细信息
ISBN:
(纸本)9781479984268
Given the rise of ubiquitous computing and communication devices like biosensors, smart watch, and smartphones, real-time online systems can provide users with a wide range of supports including monitoring daily activities and retrieving of personal data. User activity pattern can give an abstraction and summarization about physical behavior for certain group of people. However, one of the biggest challenges in this topic that we are facing today is the big data problem associated with large, complex, and dynamic data. In addition, as the demand for the integration and analysis of dynamic data as well as static historical data from different sources has been growing steadily, smartphones with limited capacity and computing abilities can hardly manage and process such a huge task. To address these above issues, a new framework has to be used to assist in the process, analysis, and integration of big data for a mobile platform. In this paper, I propose a distributed cloud based pervasive framework to help do complicated computing for a mobile platform. The framework has the ability to collect, process, analyze, and integrate different types of data from different sources by using state-of-the-art technologies. The purpose of this framework is to provide an intelligent and efficient approach to analyze and combine new incoming data with historical data to build and refine a solid user activity pattern.
Partitioned Global Address Space (PGAS) parallel programming models can provide an efficient mechanism for managing shared data stored across multiple nodes in a distributed memory system. However, these models are tr...
详细信息
Partitioned Global Address Space (PGAS) parallel programming models can provide an efficient mechanism for managing shared data stored across multiple nodes in a distributed memory system. However, these models are traditionally directly addressed and, for applications with loosely-structured or sparse data, determining the location of a given data element within a PGAS can incur significant overheads. Applications incur additional overhead from the network latency of lookups from remote location resolution structures. Further, for large data, caching such structures locally incurs space and coherence overheads that can limit scaling. We observe that the matching structures used by implementations of the Message Passing Interface (MPI) establish a separation between incoming data writes and the location where data will be stored. In this work, we investigate extending such structures to add a layer of indirection between incoming data reads and the location from which data will be read, effectively extending PGAS models with logical addressing.
Edge caching can provide a limited delay guarantee for delay-sensitive industrial applications. However, in a highly distributed edge environment, cached data is prone to be damaged, and its integrity must be ensured....
详细信息
ISBN:
(数字)9798331509712
ISBN:
(纸本)9798331509729
Edge caching can provide a limited delay guarantee for delay-sensitive industrial applications. However, in a highly distributed edge environment, cached data is prone to be damaged, and its integrity must be ensured. In this paper, we propose an integrity audit scheme based on collaborative caching in the Industrial Internet of Things (IIoT) environment. This scheme shifts the data storage function from the cloud core network to the edge network, allowing cooperation to fulfill user requests. Firstly, an end-edge-cloud collaborative caching mechanism is designed, partitioning cache space and achieving hierarchical placement through edge collaboration. Secondly, a cache value function combining content recommendation and cost is proposed; cache hit rate and value serve as constraints for replacement and update to dynamically adjust cache data in the edge network. Finally, the analysis proves the correctness and reliability of the scheme. Experimental results show that this scheme can quickly and effectively audit cached data, reduce waiting delay of industrial tasks, and address resource waste from invalid audits.
Timely access to quality data and linkage of data beyond disciplinary boundaries is essential for the marine research community. Therefore the national “Marine Network for Integrated Data Access” is establishing the...
详细信息
Timely access to quality data and linkage of data beyond disciplinary boundaries is essential for the marine research community. Therefore the national “Marine Network for Integrated Data Access” is establishing the “Data Portal German Marine Research” to facilitate seamless access to marine data and services and to promote the exchange and dissemination of marine data interlinked with corresponding scientific publications. In that course the data portal was conceptualized and developed to provide a “one-stop-shop” approach to marine research data from various data providers in terms of coherent discovery, view, download and dissemination of scientific data and publications. The data portal is based on a central harvesting and interfacing approach by connecting distributed data sources. To achieve interoperability the data portal makes use of internationally endorsed standards (e.g., ISO, OGC, OAI-PMH). In this paper we provide information on details of content, functionality, services, architecture, interfaces and standards of the data portal and the network of contributing data providers.
In recent years, Hadoop framework is popularly known for providing cost-effective solutions to process large-scale data intensive applications in a distributed manner. Storage imbalance during replica placement in Had...
详细信息
In recent years, Hadoop framework is popularly known for providing cost-effective solutions to process large-scale data intensive applications in a distributed manner. Storage imbalance during replica placement in Hadoop is harmful. Replica placement in HDFS plays a major role in data availability and balanced utilization of clusters. In this paper we propose a solution for load-aware replica placement in Hadoop such that a cluster is divided into small size partitions where at-least one partition will be nearer to its users. Partitions are created using minimum spanning tree and results are compared with the default replica placement policy of Hadoop. Experimental results of the proposed solution confirm that, load-aware replica placement gives uniform rack level utilization and reduces read access time over the default replica placement policy of Hadoop in a multiuser environment.
Modern multimedia applications use an increasing number of sensors including cameras, microphones and microphone arrays. These applications must acquire and process data from sensors in real-time, which is usually bey...
详细信息
Modern multimedia applications use an increasing number of sensors including cameras, microphones and microphone arrays. These applications must acquire and process data from sensors in real-time, which is usually beyond the capabilities of single machines. We present our distributed sensor data transport middleware, the NIST data flow system II, which offers network transparent services for data acquisition and processing across a local network of computers. An application is thus represented as a data flow graph, with streaming media flowing between the different computational components. This is highlighted by presenting a multimedia application tracking persons using a sensor fusion of audio and video streams.
Marvel is a distributed computing environment that allows the creation of scalable management services using intelligent agents and the World-Wide Web. Marvel is based on an information model that generates computed v...
详细信息
Marvel is a distributed computing environment that allows the creation of scalable management services using intelligent agents and the World-Wide Web. Marvel is based on an information model that generates computed views of management information and a distributed computing model that makes this information available to a variety of client applications. Marvel does not replace existing element management agents but rather builds on top of them a hierarchy of servers that aggregate the underlying information in a synchronous or asynchronous fashion and present it in the form of Java-enriched Web pages. It uses a distributed database to reduce the cost associated with centralized network management systems and mobile agent technology to: (a) support thin clients by uploading the necessary code to access Marvel services; and (b) extend its functionality dynamically by downloading code that incorporates new objects and services. A prototype implementation in Java is presented together with results from its first application on a broadband home access network using cable modems.
暂无评论