Erasure codes are promising for improving the reliability of the storage system due to its space efficiency compared to the replication methods. Traditional erasure codes split data into equalsized data blocks and enc...
详细信息
Erasure codes are promising for improving the reliability of the storage system due to its space efficiency compared to the replication methods. Traditional erasure codes split data into equalsized data blocks and encode strips in different data blocks. This brings heavy repairing traffic when clients read parts of the data, since most strips read for repairing are not in the expected blocks. This paper proposes a novel discrete data dividing method to completely avoid this problem. The key idea is to encode strips from the same data block. We could see that for repairing failed blocks, the strips to be read are either in the same data block with corrupted strips or from the encoded strips. Therefore, no data is wasted. We design and implement this data layout into a HDFS-like storage system. Experiments over a small-scale testbed shows that the proposed discrete data divided method avoids downloading data blocks that are not needed for clients during the repairing operations.
Context situation, which means a snapshot of the status of the real world, is formed by integrating a large amount of contexts collected from various resources. How to get the context situation and use the situation t...
详细信息
Context situation, which means a snapshot of the status of the real world, is formed by integrating a large amount of contexts collected from various resources. How to get the context situation and use the situation to provide better services is a challenging issue. In this paper, we focused on this challenge on the basis of the mobile cloud computing architecture. An abstract model is proposed in this paper to uniformly collect the context and send them to cloud. A rule-based large-scale context aggregation algorithm is also proposed which utilizes the MapReduce computing paradigm. Finally, a large-scale context management framework based on the abstract model and the context aggregation algorithm is proposed, and a real-time traffic demo is implemented to verify the validity of the framework.
Audio matching automatically retrieves all excerpts that have the same content as the query audio clip from given audio recordings. The extracted feature is critical for audio matching and the Chroma Energy Normalized...
详细信息
Audio matching automatically retrieves all excerpts that have the same content as the query audio clip from given audio recordings. The extracted feature is critical for audio matching and the Chroma Energy Normalized Statistics(CENS) feature is the state-of-the-arts. However, CENS might behave unsatisfactorily on some audio because it is a handcraft feature. In this paper, we propose to utilize the features learned by Convolutional Deep Belief Network(CDBN) to enhance the performance of audio matching. Benefit from the strong generalization ability of CDBN, our method works better than CENS based methods on most audio datasets. Since the features learned by CDBN are binary-valued, we can develop a more efficient audio matching algorithm by taking the advantage of this property. Experimental results on both TIMIT dataset and a simulated music dataset confirm effectiveness of the proposed CDBN based method comparing with the traditional CENS feature based algorithm.
With the exponential growth of mobile traffic data, mobile traffic classification is in a great need. It is an essential step to improve the performance of network services such as QoS and security monitoring. However...
详细信息
ISBN:
(数字)9781728186955
ISBN:
(纸本)9781728186962
With the exponential growth of mobile traffic data, mobile traffic classification is in a great need. It is an essential step to improve the performance of network services such as QoS and security monitoring. However, the widespread use of encrypted protocols, especially the TLS protocol, has posed great challenges to traditional traffic classification techniques. As the rule-based deep packet inspection approaches are ineffective for encrypted traffic classification, various machine learning methods have been studied and used. Recently, deep learning solutions which enable automatic feature extraction are also proposed to classify encrypted traffic. In this paper we propose App-Net, an end-to-end hybrid neural network, to learn effective features from raw TLS flows for mobile app identification. App-Net is designed by combining RNN and CNN in a parallel way. So that it can learn a joint flow-app embedding to characterize both flow sequence patterns and unique app signatures. We evaluate App-Net on a real-world dataset that covering 80 apps. The results show that our method can achieve an excellent performance and outperform the state-of-the-art methods.
Power-gating is a representative circuit level technique to mitigate leakage power. While in low-power Network-on-Chip (NoC) design, the former fine-grained power-gating methods will decrease network performance due t...
详细信息
Power-gating is a representative circuit level technique to mitigate leakage power. While in low-power Network-on-Chip (NoC) design, the former fine-grained power-gating methods will decrease network performance due to serial wake-up latency and head-of-line blocking. Therefore, we propose a flexible Virtual Channel (VC) management scheme for fine-grained power-gating to achieve high throughput and low-power. The proposed power-gating method with the early wake-up is evaluated by using some synthetic workloads. When compared with an optimized early wake-up power-gating technique, it can improve performance effectively in medium and high network loads, and increases the network throughput by 15.7%~44.1% for different synthetic loads, while keeps network power consumption as low as the optimized method. For the PARSEC application traces of token based protocol, it can significantly decrease packet latency by 20.3% on average, however only increases less than 3.6% peak power when compared with the optimized method.
Meteorology Grid Computing aims to provide scientist with seamless, reliable, secure and inexpensive access to meteorological resources. In this paper, we presented a semantic-based meteorology grid service registry, ...
Meteorology Grid Computing aims to provide scientist with seamless, reliable, secure and inexpensive access to meteorological resources. In this paper, we presented a semantic-based meteorology grid service registry, discovery and composition framework by combining grid technologies and the advantages of semantic web techniques. The main objective of the framework is to support automating the discovery, selection, and workflow composition of semantically described heterogeneous meteorological grid services, which offers the possibility of facilitating geographically distributed meteorological scientists to resolve complex scientific problems cooperately. With this framework, the key technologies such as semantic registry, semantic matchmaking, QoS ranking and composition model, will be discussed.
Maximum likelihood estimation has been widely adopted along with the encoder-decoder framework for video captioning. However, it ignores the structure of sentences and restrains the diversity and distinction of genera...
详细信息
ISBN:
(纸本)9781665487900
Maximum likelihood estimation has been widely adopted along with the encoder-decoder framework for video captioning. However, it ignores the structure of sentences and restrains the diversity and distinction of generated captions. To address this issue, we propose a hard contrastive learning (HCL) method for video captioning. Specifically, built on the encoder-decoder framework, we introduce mismatched pairs to learn a reference distribution of video descriptions. The target model on the matched pairs is learned on top the reference model, which improves the distinctiveness of generated captions. In addition, we further boost the distinctiveness of the captions by developing a hard mining technique to select the hardest mismatched pairs within the contrastive learning framework. Finally, the relationships among multiple relevant captions for each video is consider to encourage the diversity of generated captions. The proposed method generates high quality captions which effectively capture the specialties in individual videos. Extensive experiments on two benchmark datasets, i.e., MSVD and MSR-VTT, show that our approach outperforms state-of-the-art methods.
Monte Carlo (MC) simulation plays an important part in dose calculation for radiotherapy treatment planning. Since the accuracy of MC simulation relies on the number of simulated particles histories, it's very tim...
详细信息
Monte Carlo (MC) simulation plays an important part in dose calculation for radiotherapy treatment planning. Since the accuracy of MC simulation relies on the number of simulated particles histories, it's very time-consuming. The Intel Many Integrated Core (MIC) architecture, which consists of more than 50 cores and supports many parallel programming models, provides an efficient alternative for accelerating MC dose calculation. This paper implements the OpenMP-based MC Dose Planning Method (DPM) for radiotherapy treatment problems on the Intel MIC architecture. The implementation has been verified on the target MIC coprocessor including 57 cores. The results demonstrate that the OpenMP-based DPM implementation exhibits very accurate results and achieves the maximum speedup of 10.53 times in comparison to the original DPM one on a Xeon E5-2670 CPU. Additionally, speedup and efficiency of the implementation running on the different number of cores in MIC are also reported.
Convolutional layers are ubiquitous in a variety of deep neural networks. Due to the lower computation complexity and the smaller number of parameters, convolutions with small filter sizes are often used, such as one ...
详细信息
ISBN:
(数字)9781728169811
ISBN:
(纸本)9781728169828
Convolutional layers are ubiquitous in a variety of deep neural networks. Due to the lower computation complexity and the smaller number of parameters, convolutions with small filter sizes are often used, such as one by one convolution. Nevertheless, these small convolution operations are still time-consuming. A common approach to implementing convolutions is to transform them into matrix multiplications, known as GEMM-based convolutions. The approach maybe incurs additional memory overhead and calls matrix multiplication routines, which are not optimized for matrices generated by convolutions. In this paper, we present a new parallel one by one direct convolution implementation on ARMv8 multi-core CPUs, which doesn't incur any additional memory space requirement. Our implementation is verified on two ARMv8 CPUs, Phytium FT-1500A and FT-2000plus. In terms of performance and scalability, our implementation is better than GEMM-based implementations in all the tests on Phytium FT-1500A. On Phytium FT-2000plus, our approach gives much better performance and scalability than GEMM-based approaches in most cases.
暂无评论