Real-time database systems must maintain consistency while minimizing the number of transactions that miss the deadline. To satisfy both the consistency and real-time constraints, there is the need to integrate synchr...
详细信息
Concurrency manipulates the control of concurrent transaction execution. distributed database management system enforce concurrency manipulate to make sure serializability and isolation of transaction. Lots of researc...
详细信息
ISBN:
(纸本)9781467366809;9781467366793
Concurrency manipulates the control of concurrent transaction execution. distributed database management system enforce concurrency manipulate to make sure serializability and isolation of transaction. Lots of research has been done on this area and a number of algorithms have been purposed. In this article, we are comparing few algorithms for preserving the ACID property (atomicity, consistency, isolation, and durability) of transactions in DDBMS.
In time-domain astronomy, STLF (Short-Timescale and Large Field-of-view) sky survey is the latest way of sky observation. Compared to traditional sky survey who can only find astronomical phenomena, STLF sky survey ca...
详细信息
ISBN:
(数字)9783030280611
ISBN:
(纸本)9783030280611;9783030280604
In time-domain astronomy, STLF (Short-Timescale and Large Field-of-view) sky survey is the latest way of sky observation. Compared to traditional sky survey who can only find astronomical phenomena, STLF sky survey can even reveal how short astronomical phenomena evolve. The difference does not only lead the new survey data but also the new analysis style. It requires that database behind STLF sky survey should support continuous analysis on data streaming, real-time analysis on short-term data and complex analysis on long-term historical data. In addition, both insertion and query latencies have strict requirements to ensure that scientific phenomena can be discovered. However, the existing databases cannot support our scenario. In this paper, we propose AstroServ, a distributed system for analysis and management of large-scale and full life-cycle astronomical data. AstroServ's core components include three data service layers and a query engine. Each data service layer serves for a specific time period of data and query engine can provide the uniform analysis interface on different data. In addition, we also provide many applications including interactive analysis interface and data mining tool to help scientists efficiently use data. The experimental results show that AstroServ can meet the strict performance requirements and the good recognition accuracy.
Recently with the advent of the information technology and the wide spread use of the Internet, new user oriented production and logistic systems, such as the Supply Chain Management System, have been required in orde...
详细信息
ISBN:
(纸本)0769517692
Recently with the advent of the information technology and the wide spread use of the Internet, new user oriented production and logistic systems, such as the Supply Chain Management System, have been required in order to cope with the drastic and continuous changes on the markets and users' preferences. To achieve this purpose, the companies need to mutually integrate their database systems with heterogeneous requirements. Autonomous Decentralized database System (ADDS) is proposed as a system architecture in order to realize assurance when distributed database systems with heterogeneous requirements have to coexist in a common environment without violating each other characteristics. In this system architecture, a loose consistency management technology is proposed in order to let each database to autonomously update its own information while preserving the consistency of the total system. The autonomy of each site can be achieved by defining a data attribute, Allowable Volume (AV), at each DB within which each site can update. Moreover, a background coordination technology, performed by an autonomous mobile agent (MA), permits that the sites can coordinate and cooperate each other. The MA moves around the sites in order to continuously adjust and allocate AV by negotiating with them under evolving situations. Furthermore, the effectiveness of this assurance systems is experimentally shown in the supply-chain management system model by simulation.
The query cost usually as an important criterion for a distributed database. The genetic algorithm is an adaptive probabilistic search algorithm, but the crossover and mutation probability usually keep a probability i...
详细信息
ISBN:
(纸本)9783038351153
The query cost usually as an important criterion for a distributed database. The genetic algorithm is an adaptive probabilistic search algorithm, but the crossover and mutation probability usually keep a probability in traditional genetic algorithm. If the crossover probability keep a large value, the possibility of damage for genetic algorithm model is greater;In turn, if the crossover probability keep a small value, the search process will transform a slow processing or even stagnating. If the mutation probability keep a small value, a new individual can be reproduced difficultly;In turn, if the mutation probability keep a large value, the genetic algorithm will as a Pure random search algorithm. To solve this problem, proposed a improved genetic algorithm that multiple possibility of crossover and mutation based on k-means clustering algorithm. The experiment results indicate that the algorithm is effective.
In this paper, on the basis of the research on the existing distributed query optimization, against the higher repetition rate, a new query processing program is presented, which completes the data dictionary, the ove...
详细信息
ISBN:
(纸本)9781424445189
In this paper, on the basis of the research on the existing distributed query optimization, against the higher repetition rate, a new query processing program is presented, which completes the data dictionary, the overall model design, and then improves the query flow. Experiments shows that the strategies can improve the efficiency of the query, reduce the amount of middle data, and effectively reduce the total cost of network communication.
Currently the massive power quality monitoring data are stored in the centralized database of the monitoring master station, the problems such as large storage space;low query retrieval speed;low reliability and poor ...
详细信息
ISBN:
(纸本)9783037857434
Currently the massive power quality monitoring data are stored in the centralized database of the monitoring master station, the problems such as large storage space;low query retrieval speed;low reliability and poor scalability will be caused. This paper proposes a data management scheme for massive power quality monitoring data based-on the distributed-database-system. The monitoring data of different power quality indexes are stored in the distributed servers of the existing monitoring sub-stations;the server of monitoring master station is used for storing data characteristics value and data indexes, it is also used for unified management of distributed database system. The scheme takes full advantage of each server's storage space and network bandwidth, and saves the storage space and improves the access efficiency.
In the past few years, the number of sensors is rapidly increasing on the edge of the network. IoT(Internet of things) devices play not only data producers but also data consumers. It is valuable to deploy a distribut...
详细信息
ISBN:
(纸本)9783030858995;9783030858988
In the past few years, the number of sensors is rapidly increasing on the edge of the network. IoT(Internet of things) devices play not only data producers but also data consumers. It is valuable to deploy a distributed database on the edge of the network. However, flash memory is the mainstream storage medium on the edge, which is different from cloud environment. Flash memory can wear out through repeated writes while large amount of data are written on the edge per day. Thus, in this paper, SardineDB is presented, which is a decentralized distributed database optimized for edge. The engine of SardineDB is Sardine-Core, which is a flash-optimized key-value separation storage based on LevelDB. SardineCore has low GC (garbage collection) burden, which can be used to low the write amplification and improve the write performance on the edge. From evaluation results, the write performance and random read performance of SardineDB have great advantages compared with existing distributed databases on the edge. As a result, SardineDB is very suitable for edge because it has high write performance, low GC burden and low write amplification.
With the continuous development of modern information technology, the use of scientific and technological literature data is increasing, among which semi-structured data accounts for a large proportion, and it has the...
详细信息
ISBN:
(纸本)9781450362788
With the continuous development of modern information technology, the use of scientific and technological literature data is increasing, among which semi-structured data accounts for a large proportion, and it has the characteristics of large quantity and extensive growth. Therefore, distributed database has been widely applied. At the same time, the optimization algorithm for distributed database becomes the research hot spot. However, when using optimization algorithm to solve problems, researchers have not fully solved the limitations of optimization algorithm. In order to improve the retrieval speed of distributed database, this paper combined with the global optimization ability of ant colony algorithm and the local optimization ability of simulated annealing algorithm, the DDQO algorithm is proposed. At the same time, this paper conducted a control experimental on the DDQO algorithm. Experimental results show that the DDQO algorithm is superior to ant colony algorithm and simulated annealing algorithm, it increase the query speed of distributed database by 24%.
This paper introduces SMA, a sampling algorithm for mining association rules in distributed database. SMA adopts two-way sampling technique in every website and amalgamates those local samples into a sampling dataset....
详细信息
ISBN:
(纸本)9780769536040
This paper introduces SMA, a sampling algorithm for mining association rules in distributed database. SMA adopts two-way sampling technique in every website and amalgamates those local samples into a sampling dataset. In an empirical study, SMA was able to accelerate the speed and improve the accuracy of mining association rules in distributed database.
暂无评论