This article delves into heterogeneous computing systems, which employ multi-core processors (CPUs) and graphics processing units (GPUs) concurrently, facilitating efficient handling of resource-intensive tasks demand...
详细信息
Android malware detection has become research hotspot in mobile security. When security service providers obtain feature information from target samples, they may involve user privacy information such as identity and ...
详细信息
Driven by the need to support spatial data applications, most relational databases offer spatial SQL query features. However, traditional relational databases are not scalable, and their query processing follows a pul...
详细信息
We provide a comprehensive and updated assessment of Docker versus Docker in Docker (DinD), evaluating its impact on CPU, memory, disk, and network. Using different workloads, we evaluate DinD's performance across...
详细信息
ISBN:
(纸本)9798350363074;9798350363081
We provide a comprehensive and updated assessment of Docker versus Docker in Docker (DinD), evaluating its impact on CPU, memory, disk, and network. Using different workloads, we evaluate DinD's performance across distinct hardware platforms and GNU/Linux distributions on cloud Infrastructure as a Service (IaaS) platforms like Google Compute Engine (GCE) and traditional server-based environments. We developed an automated tools suite to achieve our goal. We execute four well-known benchmarks on Docker and its nested-container variant. Our findings indicate that nested-containers require up to 7 seconds for startup, while the Docker standard containers require less than 0.5 seconds for Debian and Alpine operating systems. Our results suggest that Docker containers based on Debian consistently outperform their Alpine counterparts, showing lower CPU latency. A key distinction among these Docker images lies in the varying number of installed libraries (e.g., stretching from 13 to 119) across different Linux distributions for the same system (e.g., MySQL). Furthermore, the number of events and CPU latency indicates that the influence of DinD over Docker proves that it is insignificant for both operating systems. In terms of memory, running containers of Debian-based images consume 20% more size of memory than those based on Alpine. No significant differences are between nested-containers and Dockers for disk and network IO. It is worth emphasizing that some of the disparities, such as a bigger memory footprint, appear to be a direct result of the software stack in use, including different kernel versions, libraries, and other essential packages.
Recently, parallel and distributedcomputing has become indispensable in various application domains, such as cloud computing and extensive data analysis, to meet the growing demands for efficiency and reliability. In...
详细信息
The proceedings contain papers. The special focus in this conference is on . The topics include: Impact of artificial intelligence on high-penetration renewable physical infrastructure;collaborative attacks and defens...
ISBN:
(纸本)9781032738659
The proceedings contain papers. The special focus in this conference is on . The topics include: Impact of artificial intelligence on high-penetration renewable physical infrastructure;collaborative attacks and defense;building a safe and secure metaverse;distributed-computing based versatile healthcare services framework for diagnostic markers;supervised context-aware Latent Dirichlet allocation-based drug recommendation model;combination kernel support vector machine based digital twin model for prediction of dyslexia in distributed environment;Time series analysis of vegetation change using remote sensing, GIS and FB prophet;advances in smart farming for precision agriculture: Green-IoT and machine learning as a solution;Towards a serverless computing and edge-intelligence architecture for the Personal-Internet-of-Things (PIoT);IOT and developed deep learning based road accident detection system and societal knowledge management;Healthcare CAD model for hierarchical processing of surgical navigation system;applicability of eye tracking technology in virtual keyboard for human-computer interactions;a digital twin framework for smart contract-based DeFi applications in the metaverse: Towards interoperability, service scaleup & resilience;a novel approach to glass identification using ensemble learning for forensics;deep learning-based cognitive digital twin system for wrist pulse diagnostic and classification;Multi-class instance segmentation for the detection of cervical cancer cells using modified mask RCNN;dilated convolution model for lightweight neural network;artificial intelligence based Monte Carlo model for epidemic forecasting for societal aspect;efficient data visualization through novel recurrent neural network-based dimension reduction;An optimized CNN-based model for pneumonia detection;EEG-based emotion recognition: Leveraging CNNs for precision;A seven layer DNN approach for social-media multilevel image-based text classification.
This paper presents techniques for theoretically and practically efficient and scalable Schrodinger-style quantum circuit simulation. Our approach partitions a quantum circuit into a hierarchy of subcircuits and simul...
详细信息
ISBN:
(数字)9798350352917
ISBN:
(纸本)9798350352924;9798350352917
This paper presents techniques for theoretically and practically efficient and scalable Schrodinger-style quantum circuit simulation. Our approach partitions a quantum circuit into a hierarchy of subcircuits and simulates the subcircuits on multi-node GPUs, exploiting available data parallelism while minimizing communication costs. To minimize communication costs, we formulate an Integer Linear Program that rewards simulation of "nearby" gates on "nearby" GPUs. To maximize throughput, we use a dynamic programming algorithm to compute the subcircuit simulated by each kernel at a GPU. We realize these techniques in Atlas, a distributed, multi-GPU quantum circuit simulator. Our evaluation on a variety of quantum circuits shows that Atlas outperforms state-of-the-art GPU-based simulators by more than 2x on average and is able to run larger circuits via offloading to DRAM, outperforming other large-circuit simulators by two orders of magnitude.
作者:
Liu, KangkangChen, NingjiangGuangxi Univ
Coll Comp & Elect Informat Nanning Peoples R China Guangxi Univ
Educ Dept Guangxi Zhuang Autonomous Reg Key Lab Parallel Distributed & Intelligent Comp Nanning Peoples R China
With the increasing performance of deep convolutional neural networks, they have been widely used in many computer vision tasks. However, a huge convolutional neural network model requires a lot of memory and computin...
详细信息
ISBN:
(纸本)9798350349184;9798350349191
With the increasing performance of deep convolutional neural networks, they have been widely used in many computer vision tasks. However, a huge convolutional neural network model requires a lot of memory and computing resources, which makes it difficult to meet the requirements of low latency and reliability of edge computing when the model is deployed locally on resource-limited devices in edge environments. Quantization is a kind of model compression technology, which can effectively reduce model size, calculation cost and inference delay, but the quantization noise will cause the accuracy of the quantization model to decrease. Aiming at the problem of precision loss caused by model quantization, this paper proposes a post-training quantization method based on scale optimization. By reducing the influence of redundant parameters in the model on the quantization parameters in the process of model quantization, the scale factor optimization is realized to reduce the quantization error and thus improve the accuracy of the quantized model, reduce the inference delay and improve the reliability of edge applications. The experimental results show that under different quantization strategies and different quantization bit widths, the proposed method can improve the accuracy of the quantized model, and the absolute accuracy of the optimal quantization model is improved by 1.36%. The improvement effect is obvious, which is conducive to the application of deep neural network in edge environment.
With the advances in the large-scale computing platforms, jobs that run on such systems are becoming complex and they present increased variability. In these systems, the workload usually has complex structure includi...
详细信息
ISBN:
(纸本)9798350319439
With the advances in the large-scale computing platforms, jobs that run on such systems are becoming complex and they present increased variability. In these systems, the workload usually has complex structure including jobs with different characteristics. Along with regular jobs, there are jobs represented as linear workflows (LWs) or even jobs which are bags of LWs. Consequently, efficient scheduling algorithms should be employed to ensure satisfactory performance as well as fair processing of jobs. This research aims to study appropriate scheduling algorithms of mixed workloads consisting of simple jobs including one task only, along with bags-of-LWs (BoLWs). Different techniques for scheduling are studied in a set of distributed resources. Extensive simulation experiments are carried out to investigate their effectiveness for various system utilization levels and variability in job/task service demands. The results demonstrate that the efficacy of the scheduling schemes depends on the variability level in job/task service demands, the system load, and the level of fairness that is considered as satisfactory.
As the size of modern datasets exceeds the disk and memory capacities of a single computer, machine learning practitioners have resorted to parallel and distributedcomputing. Given that optimization is one of the pil...
详细信息
暂无评论