Due to the excessive cost of data collection as well as annotation in dedicated domains, artificial intelligence model training is difficult to achieve optimality with insufficient data. To optimize this issue, a text...
详细信息
With the continuous development and widespread application of big data and cloud computing technologies, there is an urgent need to balance the security and convenience of data during its use and storage. To enable se...
详细信息
In a cloud-native architecture, the operational data of various system components experiences a significant increase. From the distributed complex system, obtaining the operation status data and realizing real-time mo...
详细信息
In data center (DC) environments, machine learning algorithms play an important role in resource management to increase efficiency by means of proper predictive monitoring workload trends and adjusting jobs accordingl...
详细信息
In data center (DC) environments, machine learning algorithms play an important role in resource management to increase efficiency by means of proper predictive monitoring workload trends and adjusting jobs accordingly. In this paper, we propose a system to predict the CPU usage of virtual machines (VMs) of a DC. Our proposal performs clustering of VMs based on their historical information (i.e., time series) by evaluating several traditional ML algorithms using common statistical features of VM time series, which facilitates grouping VMs with similar behaviors and establishing clusters based on these features. Then, training of representative models is performed to finally choose the one with the lowest mean error per cluster. The simulation results show that by performing clustering and training the model with representative time series, it is indeed possible to obtain a low mean error while reducing the local training time per individual VM. (c) 2020 The Authors. Published by Elsevier B.V.
This paper delves into the transformative potential of data science for earthquake prediction techniques. Through a thorough literature review, it explores methodologies spanning machine learning, deep learning, time ...
详细信息
In this digital era, every digital movement or change can be captured. One of the forms can be the telecommunication data and call detail records (CDR) when a person makes a call to each other. This data, after stream...
详细信息
The popularization and application of the Internet of Things (IoT) technology has brought massive time series data, which puts forward higher requirements for data compression technology. At present, most existing com...
详细信息
This study centers on the application of vertical federated learning technology in the context of Internet banking loans, with a particular focus on innovations in data privacy protection, risk control model algorithm...
详细信息
Fuzzy c means is a conventional clustering algorithm that uses complete data sets for the clustering process, making it difficult to deal with incomplete data which is a critical problem in medical research that canno...
详细信息
The data-intensive applications of today's big data era often produce a large memory footprint. As a result, a significant volume of data needs to travel from memory to the CPU under the traditional Von-Neumann co...
详细信息
ISBN:
(纸本)9798350383225
The data-intensive applications of today's big data era often produce a large memory footprint. As a result, a significant volume of data needs to travel from memory to the CPU under the traditional Von-Neumann computing paradigm. Near-memory processing (NMP) or processing-in-memory (PIM) is a potential alternate computation framework where a computation unit is placed near the memory (or inside the memory) and a portion of an application is executed on it (termed computation offloading) aiming to reduce the amount of data movement and its consequences. Although a few computation offloading strategies have been proposed in recent times, the existing approaches do not consider the data locality offered by the last level cache and the overall execution time of the application while designing their policies. In this paper, we propose a data locality-aware computation offloading strategy for a hybrid computing system comprising the host processor and NMP-enabled 3D memory. After the application code is instrumented using the LLVM compiler framework, the strategy offloads a portion of an application to NMP if its estimated overall execution time is less. An extensive simulation performed on a set of standard simulators for a bunch of large graph-based application benchmarks reports the effectiveness of the proposed strategy by achieving a maximum speedup of 40% and 11.8% as compared to the host-only configuration and the state-of-art policy, respectively. The proposed strategy also reduces the off-chip data transfer and energy consumption by a significant margin as compared to the host-only configuration (avg 27%) and the state-of-art policy (avg 28%). Further, the proposed policy reduces the LLC miss rate by 57% as compared to the state-of-art policy.
暂无评论