For a transactional database system, the efficiency of logging is usually crucial to its performance. The emergence of new hardware, such as NVM and SSD, eliminated the traditional I/O bottleneck of logging and releas...
详细信息
ISBN:
(纸本)9783319914589;9783319914572
For a transactional database system, the efficiency of logging is usually crucial to its performance. The emergence of new hardware, such as NVM and SSD, eliminated the traditional I/O bottleneck of logging and released the potential of multi-core CPUs. As a result, the parallelism of logging becomes important. We propose a parallel logging subsystem called TwinBuf and implemented it in PostgreSQL. This solution can make better use of multi-core CPUs, and is generally applicable to all kinds of storage devices, such as hard disk, SSD and NVM. TwinBuf adopts per-thread logging slots to parallelize logging, and a twin-log-buffer mechanism to make sure that logging can be performed in a non-stop manner. It performs group commit to minimize the persistence overheads. Experimental evaluation was conducted to demonstrate its advantages.
Evaluating the comprehensive performance of a warehouse-scale computer (WSC) has been a long-standing challenge. Traditional load-testing benchmarks become ineffective because they cannot accurately reproduce the beha...
详细信息
ISBN:
(纸本)9781450349116
Evaluating the comprehensive performance of a warehouse-scale computer (WSC) has been a long-standing challenge. Traditional load-testing benchmarks become ineffective because they cannot accurately reproduce the behavior of thousands of distinct jobs co-located on a WSC. We therefore evaluate WSCs using actual job behaviors in live production environments. From our experience of developing multiple generations of WSCs, we identify two major challenges of this approach: 1) the lack of a holistic metric that incorporates thousands of jobs and summarizes the performance, and 2) the high costs and risks of conducting an evaluation in a live environment. To address these challenges, we propose WSMeter, a cost-effective methodology to accurately evaluate a WSC's performance using a live production environment. We first define a new metric which accurately represents a WSC's overall performance, taking a wide variety of unevenly distributed jobs into account. We then propose a model to statistically embrace the performance variance inherent in WSCs, to conduct an evaluation with minimal costs and risks. We present three real-world use cases to prove the effectiveness of WSMeter. In the first two cases, WSMeter accurately discerns 7% and 1% performance improvements from WSC upgrades using only 0.9% and 6.6% of the machines in the WSCs, respectively. We emphasize that naive statistical comparisons incur much higher evaluation costs (> 4x) and sometimes even fail to distinguish subtle differences. The third case shows that a cloud customer hosting two services on our WSC quantifies the performance benefits of software optimization (+ 9.3%) with minimal overheads (2.3% of the service capacity).
Author name disambiguation is a very important and complex research topic. During the retrieval and research of literature the quality of the investigation results has been reduced because of the high probability of d...
详细信息
ISBN:
(纸本)9781538626672
Author name disambiguation is a very important and complex research topic. During the retrieval and research of literature the quality of the investigation results has been reduced because of the high probability of different authors sharing the same name, which lengthens the whole cycle of the scientific research. Therefore, it is necessary to find a reasonable and efficient method to distinguish the different authors who share the same name. In this paper, an author name disambiguation algorithm based on the fusion of multiple features (NDFMF) is proposed. First we proposed a single feature similarity detection algorithm (SFSD). SFSD is used to compute the degree of similarity between two features of a paper and to assess the threshold value. Then, SFSD is used to realize the preliminary SFSD-based disambiguation algorithm (SFSDD). Furthermore, different features are evaluated according to the disambiguation results of author names and the evaluation metrics, including precision, recall and F-measure with SFSDD. The evaluation parameter of weight (W) is introduced to express each feature's influence in disambiguation. NDFMF can disambiguate author names more efficiently based on the fusion of multiple features. Experiments were implemented to test the performance of NDFMF. Experimental results show that NDFMF was effective in the disambiguation precision, recall and F-measure.
Improving the company's performance to compete globally, in rapid growth market economy condition, is practically required. One of the important things is right decision in issues of supplier selection. The improp...
详细信息
PiDrone is a quadrotor platform created to accompany an introductory robotics course. Students build an autonomous flying robot from scratch and learn to program it through assignments and projects. Existing education...
PiDrone is a quadrotor platform created to accompany an introductory robotics course. Students build an autonomous flying robot from scratch and learn to program it through assignments and projects. Existing educational robots do not have significant autonomous capabilities, such as high-level planning and mapping. We present a hardware and software framework for an autonomous aerial robot, in which all software for autonomy can run onboard the drone, implemented in Python. We present an Unscented Kalman Filter (UKF) for accurate state estimation. Next, we present an implementation of Monte Carlo (MC) Localization and FastSLAM for Simultaneous Localization and Mapping (SLAM). The performance of UKF, localization, and SLAM is tested and compared to ground truth, provided by a motion-capture system. Our evaluation demonstrates that our autonomous educational framework runs quickly and accurately on a Raspberry Pi in Python, making it ideal for use in educational settings.
This study proposes how to implement the computer based program able to process the image, estimate and classify images the landscape images based on adaption of RGB Color Space technique into the procedure of image p...
详细信息
ISBN:
(纸本)9781538693858
This study proposes how to implement the computer based program able to process the image, estimate and classify images the landscape images based on adaption of RGB Color Space technique into the procedure of image processing. The database consisted of total 705 satellite images categorized into main four groups of landscapes: Building, Agricultural, Forest and Water. The procedure of study included image segmentation, histogram approximation for RUB color intensity, feature estimation and classification. The results obtained from processing have shown that the histograms of RGB among four categorized image groups are different in color intensity represented in term of dominant peaks and their intensity location. As visually observed, it can be found that most images of water resource have their histogram with a very high peak of color intensity obviously different from forest, building and agriculture. This kind of similar trend on histogram intensity can be found for all three color spaces of Red, Green and Blue. Therefore, only the Red histogram was selected for further procedure. The color intensity value and location were then chosen to serve as input feature set to classification. Based on the comparative performance resulted from classification, the Maximum likelihood (ML) classifier revealed its performance to be the highest accuracy at approximately 87.85% and the lowest accuracy of 51.69%, and another classifier the Least Square (LS) showed its highest accuracy at 87.85% and the lowest at 60.54%. Both types of classifiers have a very similar performance. The results from this study could suggest the applicable identification of various landscapes in images based simply on the intensity of color in histogram of image as the common effective feature set for identification of landscape images.
Logic synthesis and physical design (LSPD) tools automate complex design tasks previously performed by human designers. One time-consuming task that remains manual is configuring the LSPD flow parameters, which signif...
详细信息
ISBN:
(纸本)9781728124261
Logic synthesis and physical design (LSPD) tools automate complex design tasks previously performed by human designers. One time-consuming task that remains manual is configuring the LSPD flow parameters, which significantly impacts design results. To reduce the parameter-tuning effort, we propose an LSPD parameter recommender system that involves learning a collaborative prediction model through tensor decomposition and regression. Using a model trained with archived data from multiple state-of-the-art 14nm processors, we reduce the exploration cost while achieving comparable design quality. Furthermore, we demonstrate the transfer-learning properties of our approach by showing that this model can be successfully applied for 7nm designs.
One of the common and pressing challenges in solving real-world problems in various domains, such as in smart cities, involves solving large sparse systems of linear equations. Jacobi iterative method is used to solve...
详细信息
To build a storage management platform for music big data, we need to collect massive heterogeneous music resources from the Internet and store them on big data platforms. Therefore, it is a key problem to build a sto...
详细信息
Optical multilayer optimization continuously reorganizes layer 0-1-2 network elements to handle both existing and dynamic traffic requirements in the most efficient manner. This delays the need to add new resources fo...
详细信息
暂无评论