Crowdsourced repositories have become an increasingly important source of information for users and businesses in multiple domains. Everyday examples of tourism crowdsourcing platforms focusing on accommodation, food ...
详细信息
Crowdsourced repositories have become an increasingly important source of information for users and businesses in multiple domains. Everyday examples of tourism crowdsourcing platforms focusing on accommodation, food or travelling in general, influence consumer behaviour in modern societies. these repositories, due to their intrinsic openness, can strongly benefit from independent data quality modelling mechanisms. In this context, building trust & reputation models of contributors and storing crowdsourced data using distributed ledger technology allows not only to ascertain the quality of crowdsourced contributions, but also ensures the integrity of the built models. this paper presents a survey on distributed trust & reputation modelling using blockchain technology and, for the specific case of tourism crowdsourcing platforms, discusses the open research problems and identifies future lines of research. 2019the Authors. Published by Elsevier B.V.
Withthe fast growing trend in deep learning driven AI services over the past decade, deep learning, especially the resource-intensive and time-consuming training jobs, have become one of the main workload in today...
详细信息
ISBN:
(纸本)9789897583650
Withthe fast growing trend in deep learning driven AI services over the past decade, deep learning, especially the resource-intensive and time-consuming training jobs, have become one of the main workload in today's production clusters. However, due to the complex workload characteristics of deep learning, and the dynamic natural of shared resource environment, managing the resource allocation and execution lifecycle of distributed training jobs in cluster can be challenging. this work aims to address these issues by developing and implementing a scheduling and scaling controller to dynamically manage distributed training jobs on a Kubernetes (K8S) cluster, which is a broadly used platform for managing containerized workloads and services. the objectives of our proposed approach is to enhance K8S withthree capabilities: (1) Task dependency aware gang scheduling to avoid idle resources. (2) Locality aware task placement to minimize communication overhead. (3) Load aware job scaling to improve cost efficiency. Our approach is evaluated by real testbed and simulator using a set of TensorFlow jobs. Comparing to the default K8S scheduler, our approach successfully improved resource utilization by 20% similar to 30% and reduced job elapsed time by over 65%.
the proceedings contain 77 papers. the topics discussed include: ELS: an hard real-time scheduler for homogeneous multi-core platforms;MC-RPL: a new routing approach based on multi-criteria RPL for the Internet of thi...
ISBN:
(纸本)9781728150758
the proceedings contain 77 papers. the topics discussed include: ELS: an hard real-time scheduler for homogeneous multi-core platforms;MC-RPL: a new routing approach based on multi-criteria RPL for the Internet of things;Persian sentiment lexicon expansion using unsupervised learning methods;a case study for presenting bank recommender systems based on bon card transaction data;performance evaluation of classification data mining algorithms on coronary artery disease dataset;DMap: a distributed blockchain-based framework for online mapping in smart city;and a novel parallel jobs scheduling algorithm in the cloud computing.
Increasing the number of computational cores is a primary way of achieving high performance of contemporary supercomputers. However, developing parallelapplications capable to harness the enormous amount of cores is ...
详细信息
Withthe development of Big Data, data storage has been exposed to more challenges. Data compression which can save both storage and network bandwidth, is a very important technology to deal withthe challenges. In th...
详细信息
Approximate computing is a technique to tradeoff accuracy and hardware cost. It increases energy efficiency that leverages application-level tolerance to few errors in many applications including image processing, mul...
详细信息
ISBN:
(数字)9781728166872
ISBN:
(纸本)9781728166889
Approximate computing is a technique to tradeoff accuracy and hardware cost. It increases energy efficiency that leverages application-level tolerance to few errors in many applications including image processing, multimedia, machine learning and wireless communication. Truncated adders, as the most conventional approximate architectures, compute the addition of most significant bits, and produce small errors with high probabilities. In prior art, the adders have been analyzed considering uniformly distributed input data. However, in digital signal processing, the data has a distribution which can be considered as Gaussian distribution characterized by a mean value and standard deviation. this paper studies the effects of input data distribution on small-error approximate adders. We will show that the effects of Gaussian distribution can be modeled for the approximate adder architectures.
Big Data Cyber Security Analytics (BDCA) systems use big data technologies (e.g., Hadoop and Spark) for collecting, storing, and analyzing a large volume of security event data to detect cyber-attacks. the state-of-th...
详细信息
the performances of modern distributed stream processing systems are critically affected by the distribution of the load across workers. Skewed data streams in real world are very common and pose a great challenge to ...
详细信息
Frequent Itemset Mining (FIM) from large-scale databases has emerged as an important problem in the data mining and knowledge discovery research community. However, FIM suffers from three important limitations withth...
详细信息
this paper addresses the problem of continuously finding highly correlated pairs of time series over the most recent time window. the solution builds upon the ParCorr parallel method for online correlation discovery a...
详细信息
ISBN:
(纸本)9789897583773
this paper addresses the problem of continuously finding highly correlated pairs of time series over the most recent time window. the solution builds upon the ParCorr parallel method for online correlation discovery and is designed to run continuously on top of the UPM-CEP data streaming engine through efficient streaming operators. the implementation takes advantage of the flexible API of the streaming engine that provides low level primitives for developing custom operators. thus, each operator is implemented to process incoming tuples on-the-fly and hence emit resulting tuples as early as possible. this guarantees a real pipelined flow of data that allows for outputting early results, as the experimental evaluation shows.
暂无评论