Cloud computing and big data technologies are converging to offer a cost-effective delivery model for cloud-based big data analytics. Though impacts of size and scaling of big data on cloud have been extensively studi...
详细信息
ISBN:
(纸本)9781509066438
Cloud computing and big data technologies are converging to offer a cost-effective delivery model for cloud-based big data analytics. Though impacts of size and scaling of big data on cloud have been extensively studied, the effects of complexity of underlying analytic methods on cloud performance have received less attention. This paper will develop and evaluate a computationally intensive statistical methodology to perform inference in the presence of both non-Gaussian data and missing data. Two well-established statistical approaches, bootstrap and multiple imputations (MI), will be combined to form the methodology. Bootstrap is a computer-based nonparametric resampling procedure that involves randomly selecting data many thousands of times to construct an empirical distribution, which is then used to construct confidence intervals for significance tests. This statistical technique enables scientists who conduct studies on data with known non-normality to obtain higher quality significance tests than is possible with a traditional asymptotic, normal-theory based significance test. However, the bootstrapping procedure only works when no data are missing or the data are missing completely at random (MCAR). Missing data can lead to biased estimates when the MCAR assumption is violated. It is unclear how to best implement a bootstrapping procedure in the presence of missing data. The proposed methods will provide guidelines and procedures that will enable researchers to use the technique in all areas of health, behavior and developmental science in which a study has missing data and cannot rely on parametric inference. Either bootstrapping or MI can be computationally expensive, and combining these two can lead to further computation costs in the cloud. Using carefully constructed simulation examples, we demonstrate that it is feasible to implement the proposed methodology in a high performance Knights Landing platform. However, the computation costs are substantial
In this paper, we apply the copulas and fuzzy sets to approximate the dependencies in causes and lives, where under each cause of decrement the decrement times of the lives are assumed to be weak dependence. We propos...
详细信息
ISBN:
(纸本)9783319429724;9783319429717
In this paper, we apply the copulas and fuzzy sets to approximate the dependencies in causes and lives, where under each cause of decrement the decrement times of the lives are assumed to be weak dependence. We propose utilizing a mixture of both randomness and fuzziness to describe the concept of weak dependence. An application is considered for a general symmetric status of multiple life under dependent causes of decrement.
Quantum computer simulation provides researchers with tools for verification of quantum algorithms. GPU (Graphics Processing Units) cluster is an advisable platform for this task. However, the high cost of communicati...
详细信息
ISBN:
(纸本)9783319685052;9783319685045
Quantum computer simulation provides researchers with tools for verification of quantum algorithms. GPU (Graphics Processing Units) cluster is an advisable platform for this task. However, the high cost of communication between GPUs makes the simulation inefficiency. To overcome this drawback, we propose the following two methods. (1) A method for GPU cluster quantum simulation to improve the data locality is introduced, and two schemes for data exchanging are proposed. (2) A novel data distribution method for quantum computer simulation on GPU cluster is proposed. Experimental results show that the simulation of 33-qubit Quantum Fourier Transform algorithm using 4 nodes outperforms the serial program of the CPU cluster with a speedup of 129 times.
The proceedings contain 15 papers. The special focus in this conference is on Future Network Systems and Security. The topics include: Efficient Certificate Verification for Vehicle-to-Grid Communications;The GENI Tes...
ISBN:
(纸本)9783319655475
The proceedings contain 15 papers. The special focus in this conference is on Future Network Systems and Security. The topics include: Efficient Certificate Verification for Vehicle-to-Grid Communications;The GENI Test Automation Framework for New Protocol Development;Computational Security and the Economics of Password Hacking;DNS DDoS Mitigation, via DNS Timer Design Changes;RFID-Based Non-repudiation Protocols for Supply Chains;Towards Realizing a Distributed Event and Intrusion Detection System;Adjusting Matryoshka Protocol to Address the Scalability Issue in IoT Environment;Is High Performance computing (HPC) Ready to Handle Big data?;Transportation at the Confluence of Engineering and Big data;Competing in a Rapidly Changing World;Persuasive Educational Platform Design for Underprivileged Children;Business Analytics Generated data Brokerage;Android Application Collusion Demystified and Acceptance of Technology-Driven Interventions for Improving Medication Adherence.
Coordinated Multi-Point (CoMP) is considered as one of the most important techniques in 3GPP LTE/LTE-A. In CoMP, several base stations can be grouped together to form the cooperating set, where the cooperating set is ...
详细信息
ISBN:
(纸本)9781538608401
Coordinated Multi-Point (CoMP) is considered as one of the most important techniques in 3GPP LTE/LTE-A. In CoMP, several base stations can be grouped together to form the cooperating set, where the cooperating set is used to improve the system throughput as well as the throughput of cell edge users. Most of the studies discussed how to relieve the overload condition or maximize the system throughput by exploiting CoMP. In these studies, static cooperating set is adopted, i.e., the size of the cooperating set is fixed. Different from previous studies, this paper provides a complete discussion by considering both off-peak hours and overload conditions. In off-peak hours, we try to save the transmit power of base stations and utilize the idle radio resource to achieve energy-efficient green communications while guarantee users' service quality. On the contrary, when big activities occur with gathered crowds or during the peak hours, we propose to dynamically form the cooperating set according to the actual traffic demand condition and adapt the transmit power of base stations, so that the system can effectively offload the traffic demand to the member cells with free radio resource in the cooperating set. Our proposed approach is a decentralized-based method, which avoids excess computing being executed in a base station. Simulation results show that our proposed method achieves the best system throughput, radio resource utilization, and energy efficiency in all schemes.
Using genetic data to infer relatedness has been crucial for genetics studies for decades. In a previously published paper together with the KING software, we demonstrated that the kinship coefficient, a measure of re...
详细信息
The article analyses the model of behavior of the Russian participants of volunteer computing (VC) using platform BOINC. In contrast to the literature data based on sociological surveys our study based on clustering t...
详细信息
Intel(R) software Guard eXtensions (SGX) is a hardware-based technology for ensuring security of sensitive data from disclosure or modification that enables user-level applications to allocate protected areas of memor...
详细信息
The book is a compilation of high-quality scientific papers presented at the 3rdinternationalconference on Computer & Communication Technologies (IC3T 2016). The individual papers address cutting-edge technologi...
ISBN:
(数字)9789811032264
ISBN:
(纸本)9789811032257
The book is a compilation of high-quality scientific papers presented at the 3rdinternationalconference on Computer & Communication Technologies (IC3T 2016). The individual papers address cutting-edge technologies and applications of softcomputing, artificial intelligence and communication. In addition, a variety of further topics are discussed, which include data mining, machine intelligence, fuzzy computing, sensor networks, signal and image processing, human-computer interaction, web intelligence, etc. As such, it offers readers a valuable and unique resource.
With the advent of big data, data is being generated, collected, transformed, processed and analyzed at an unprecedented scale. Since data is created at a fast velocity and with a large variety, the quality of big dat...
详细信息
ISBN:
(纸本)9781509063185
With the advent of big data, data is being generated, collected, transformed, processed and analyzed at an unprecedented scale. Since data is created at a fast velocity and with a large variety, the quality of big data is far from perfect. Recent studies have shown that poor quality can bring serious erroneous data costs on the result of big data analysis. data validation is an important process to recognize and improve data quality. In this paper, a case study that is relevant to big data quality is designed to study original big data quality, data quality dimension, data validation process and tools.
暂无评论