Github facilitates the pull-request mechanism as an outstanding social coding paradigm by integrating with social media. The review process of pull-requests is a typical crowd sourcing job which needs to solicit opini...
详细信息
Github facilitates the pull-request mechanism as an outstanding social coding paradigm by integrating with social media. The review process of pull-requests is a typical crowd sourcing job which needs to solicit opinions of the community. Recommending appropriate reviewers can reduce the time between the submission of a pull-request and the actual review of it. In this paper, we firstly extend the traditional Machine Learning (ML) based approach of bug triaging to reviewer recommendation. Furthermore, we analyze social relations between contributors and reviewers, and propose a novel approach to recommend highly relevant reviewers by mining comment networks (CN) of given projects. Finally, we demonstrate the effectiveness of these two approaches with quantitative evaluations. The results show that CN-based approach achieves a significant improvement over the ML-based approach, and on average it reaches a precision of 78% and 67% for top-1 and top-2 recommendation respectively, and a recall of 77% for top-10 recommendation.
parallel programs face a new security problem - concurrency vulnerability, which is caused by a special thread scheduling instead of inputs. In this paper, we propose to automatically fix concurrency vulnerabilities b...
详细信息
parallel programs face a new security problem - concurrency vulnerability, which is caused by a special thread scheduling instead of inputs. In this paper, we propose to automatically fix concurrency vulnerabilities by reducing thread scheduling space. Our method is based on two observations. First, most concurrency vulnerabilities are caused by atomicity violation errors. Second, reducing thread scheduling space does not harm the correctness of the original program. We designed a prototype runtime system shield using deterministic multithreading techniques. Shield is designed to transparently run parallel programs and schedule threads in large instruction blocks to prevent atomicity violation at best effort. In case some concurrency vulnerabilities cannot be fixed by shield's scheduling reducing scheme, we also provide a remedy strategy by integrating shield with record&replay function, so that it can help programmers to analyze attacker's behavior for manually fixing.
Non-negative matrix factorization (NMF) reconstructs the original samples in a lower dimensional space and has been widely used in pattern recognition and data mining because it usually yields sparse representation. S...
详细信息
ISBN:
(纸本)9781479938414
Non-negative matrix factorization (NMF) reconstructs the original samples in a lower dimensional space and has been widely used in pattern recognition and data mining because it usually yields sparse representation. Since NMF leads to unsatisfactory reconstruction for the datasets that contain translations of large magnitude, it is required to develop translation NMF (TNMF) to first remove the translation and then conduct a decomposition. However, existing multiplicative update rule based algorithm for TNMF is not efficient enough. In this paper, we reformulate TNMF and show that it can be efficiently solved by using the state-of-the-art solvers such as NeNMF. Experimental results on face image datasets confirm both efficiency and effectiveness of the reformulated TNMF.
Internet of things is an emerging technology that aims to connect various smart objects in our daily life. It facilitates the information exchange and control among objects. In the Internet of things, it is important ...
详细信息
ISBN:
(纸本)9781479966226
Internet of things is an emerging technology that aims to connect various smart objects in our daily life. It facilitates the information exchange and control among objects. In the Internet of things, it is important to discover various relations among objects for analyzing and mining useful knowledge. Existing works on relation discovery mainly focus on centralized processing. It is not suitable for Internet of things due to unavailable of server, one-point failure, computation bottleneck, and security and business concerns. In this paper, we propose a distributed approach to discover the relations among objects in Internet of things. We first build the distritbuted system model which may include multiple relation discovery tasks. Based on that, we design an approach utilizing distributed spanning tree to extract the relations. Rather than sending all the information to the server, the objects only need to send the information to a local leader object in our algorithm. We also discuss more about how to improve the performance of the proposed approach and how to relax the system constraints. Extensive simulation have been done and the results show that the proposed approach outperforms existing approaches in terms of the data amount of transmission.
Morphological operation constitutes one of a powerful and versatile image and video applications applied to a wide range of domains, from object recognition, to feature extraction and to moving objects detection in co...
详细信息
Morphological operation constitutes one of a powerful and versatile image and video applications applied to a wide range of domains, from object recognition, to feature extraction and to moving objects detection in computer vision where real-time and high-performance are required. However, the throughput of morphological operation is constrained by the convolutional characteristic. In this paper, we analysis the parallelism of morphological operation and parallel implementations on the graphics processing unit (GPU), and field programming gate array (FPGA) are presented. For GPU platform, we propose the optimized schemes based on global memory, texture memory and shared memory, achieving the throughput of 942.63 Mbps with 3×3 structuring element. For FPGA platform, we present an optimized method based on the traditional delay-line architecture. For 3×3 structuring element, it achieves a throughput of 462.64 Mbps.
Nonparametric density estimation is a fundamental problem of statistics and data mining. Even though kernel density estimation is the most widely used method, its performance highly depends on the choice of the kernel...
详细信息
Image defogging (IDF) removes influences of fogs from an image to improve its quality. Since defogged images can significantly boost the performance of subsequent processing, IDF has attracted many attentions from the...
详细信息
Image defogging (IDF) removes influences of fogs from an image to improve its quality. Since defogged images can significantly boost the performance of subsequent processing, IDF has attracted many attentions from the computer vision community. However, existing IDF algorithms are built on the assumption that light is scattered once by a grain. Since such assumption is violated if images are contaminated by dense haze or heavy fog, traditional IDF algorithms often fail in this situation. In this paper, we propose a hybrid image defogging (HIDF) algorithm to overcome this deficiency. In particular, HIDF applies the single scattering physics model (SSPM) to pixels dominated by single scattering of light, and applies the multiple scattering physics model (MSPM) to remaining pixels. To distinguish two types of pixels, HIDF utilizes the optical thickness of corresponding pixels. If optical thickness is smaller than a threshold that determines whether the single scattering or the multiple scattering dominates, HIDF applies the SSPM, and HIDF applies the MSPM otherwise. Experimental results on several popular foggy images demonstrate that HIDF competes with the state-of-the-art algorithms, and show the promise of HIDF for defogging heavily foggy images.
Hi-GAL is a large-scale survey of the Galactic plane, performed with Herschel in five infrared continuum bands between 70 and 500 µm. We present a band-merged catalogue of spatially matched sources and their prop...
详细信息
The concerns of data-intensiveness and energy awareness are actively reshaping the design of high-performance computing (HPC) systems nowadays. The Graph500 is a widely adopted benchmark for evaluating the performance...
详细信息
ISBN:
(纸本)9781479947638
The concerns of data-intensiveness and energy awareness are actively reshaping the design of high-performance computing (HPC) systems nowadays. The Graph500 is a widely adopted benchmark for evaluating the performance of computing systems for data-intensive workloads. In this paper, we introduce a data-parallel implementation of Graph500 on the Intel Single-chip Cloud Computer (SCC). The SCC features a non-coherent many-core architecture and multi-domain on-chip DVFS support for dynamic power management. With our custom-made shared virtual memory programming library, memory sharing among threads is done efficiently via the shared physical memory (SPM) while the library has taken care of the coherence. We conduct an in-depth study on the power and performance characteristics of the Graph500 workloads running on this system with varying system scales and power states. Our experimental results are insightful for the design of energy-efficient many-core systems for data-intensive applications.
Virtualization is the foundation for cloud computing, and the virtualization can not be achieved without software defined, elastic, flexible and scalable virtual layers. Unfortunately, if multiple virtual storage devi...
详细信息
Virtualization is the foundation for cloud computing, and the virtualization can not be achieved without software defined, elastic, flexible and scalable virtual layers. Unfortunately, if multiple virtual storage devices are chained together, the system may be subject to severe performance degradation. While the read-ahead (RA) mechanism in storage devices plays a very important role to improve I/O performance, RA may not be effective as expected for multiple virtualization layers, since it is originally designed for one layer only. When I/O requests are passed through a long I/O path, they may trigger a chain reaction and lead to unnecessary data transmission and thus bandwidth waste. In this paper, we study the dynamic behavior of RA through multiple I/O layers and demonstrate that if controlled well, RA can greatly accelerate I/O speed. We present RAFlow, a RA control mechanism, to effectively improve I/O performance by strategically expanding RA window at each layer. Our real-world experiments show that it can achieve 20% to 50% performance improvement in I/O paths with up to 8 virtualized storage devices.
暂无评论