Many industries nowadays use management and decision making based on artificial neural networks. However, the major drawback of neural networks lies in their time and computational complexity. the problem with computa...
详细信息
Many industries nowadays use management and decision making based on artificial neural networks. However, the major drawback of neural networks lies in their time and computational complexity. the problem with computational complexity could be eliminated using sharing of the computing needs on multiple computing nodes. this article focuses on the architectural design of a distributed system, which aims to solve large neural networks. the article describes the technology GPGPU and the next part of the article deals with an overview of methods for speeding up the calculation and distribution of artificial neural network. the main section describes the design of a model architecture description of the algorithm that allows correct data distribution on computational nodes.
the problem of efficiently finding top-k frequent items has attracted much attention in recent years. Storage constraints in the processing node and intrinsic evolving feature of the data streams are two main challeng...
详细信息
ISBN:
(纸本)9781479967162
the problem of efficiently finding top-k frequent items has attracted much attention in recent years. Storage constraints in the processing node and intrinsic evolving feature of the data streams are two main challenges. In this paper, we propose a method to tackle these two challenges based on space-saving and gossip-based algorithms respectively. Our method is implemented on SAMOA, a scalable advanced massive online analysis machine learning framework. the experimental results show its effectiveness and scalability.
In this paper, we propose new variants of unsupervised and competitive learning algorithms designed to deal with temporal sequences. these algorithms combine features from Spiking Neural Networks (SNNs) and the advant...
详细信息
In this paper, we propose new variants of unsupervised and competitive learning algorithms designed to deal with temporal sequences. these algorithms combine features from Spiking Neural Networks (SNNs) and the advantages of the hierarchical self organizing map (HSOM). the first variant named Hierarchical Dynamic recurrent spiking self-organizing map (HD-RSSOM) is characterized by the integration of a temporal controller component to regulate the firing activity of the spiking neurons. the second variant is a hierarchical model which represents a multi-layer extension of HD-RSSOM model. the case study of the proposed HSOM variants is phonemes and words recognition in continuous speech. the applied HSOM variants serve as tools for developing intelligent systems and pursuing artificial intelligence applications.
Hadoop is a convenient framework in e-Science enabling scalable distributed data analysis. In molecular biology, next-generation sequencing produces vast amounts of data and requires flexible frameworks for constructi...
详细信息
Hadoop is a convenient framework in e-Science enabling scalable distributed data analysis. In molecular biology, next-generation sequencing produces vast amounts of data and requires flexible frameworks for constructing analysis pipelines. We extend the popular HTSeq package into the Hadoop realm by introducing massively parallel versions of short read quality assessment as well as functionality to count genes mapped by the short reads. We use the Hadoop-streaming library which allows the components to run in both Hadoop and regular Linux systems and evaluate their performance in two different execution environments: A single node on a computational cluster and a Hadoop cluster in a private cloud. We compare the implementations with Apache Pig showing improved runtime performance of our developed methods. We also inject the components in the graphical platform Cloudgene to simplify user interaction.
For many research endeavours, e-Infrastructures need to provide predictable, on-demand access to large-scale computational resources with high data availability. these need to scale withthe research communities requi...
详细信息
For many research endeavours, e-Infrastructures need to provide predictable, on-demand access to large-scale computational resources with high data availability. these need to scale withthe research communities requirements and use. One example of such an e-Infrastructure is the Australian Urban Research Infrastructure Network (AURIN -- ***) project, which supports Australia-wide research in and across the urban and built environment. this paper describes the architecture of the AURIN infrastructure and its support for access to distributed (federated) and highly heterogeneous data sets from a wide range of providers. We present how this architecture solution leverages the intersection of high throughput computing (HTC), infrastructure as a service (IaaS) Cloud services and big data technologies including use of NoSQL resources. the driving concept in this architecture and the focus of this paper is the ability for scaling up or down depending on resource demands at any given time. this is done automatically and on demand avoiding either under-or over-utilization of resources. this resource-optimization-driven infrastructure has been designed to ensure that peak loads can be predicted and successfully coped with, as well as avoid wasting resources during non-peak times. this overall management strategy has resulted in an e-Infrastructure that provides a flexible, evolving research environment that scales with research needs, rather than providing a rigid (static) end product.
At present,the power system is building up on top of a series of auxiliary systems for examples communication systems,monitoring systems,marketing systems and so *** the systems work based on the shared power system d...
详细信息
ISBN:
(纸本)9781479951499
At present,the power system is building up on top of a series of auxiliary systems for examples communication systems,monitoring systems,marketing systems and so *** the systems work based on the shared power system data which are defined using Common Information Model(CIM).Due to diversiform reasons,errors may exist in the *** the verification technologies are *** far the researchers mainly focus on the accuracy of the ***,along withthe volume of data increasing,efficiency has become an *** paper proposes a CIM verifying algorithm based on MapReduce in terms of efficiency *** experimental results show that it can enhance the performance of verification.
Research has shown that the amount of data now available often overwhelms key functions of an information system. this situation necessitates the design of information architectures that scale to meet the challenges. ...
详细信息
Research has shown that the amount of data now available often overwhelms key functions of an information system. this situation necessitates the design of information architectures that scale to meet the challenges. the Planetary Data System, a NASA funded project, has developed an information architecture for the planetary science community that addresses this and other big science data issues noted in a National Research Council report regarding architectures for big data management and analysis and end-to-end data lifecycle management across diverse disciplines. the report identified enabling technology trends including distributed systems, service-oriented architectures, ontologies, models and information representation, scalable database systems, federated data security mechanisms, and technologies for moving big data. this paper will present the PDS4 information architecture, its successful implementation in a multi-discipline big-data environment.
Location-based applications such as Facebook Places, Foursquare, or Loopt typically use location services to manage mobile object positions. However, exposing precise user positions raises user privacy concerns, espec...
详细信息
Location-based applications such as Facebook Places, Foursquare, or Loopt typically use location services to manage mobile object positions. However, exposing precise user positions raises user privacy concerns, especially if location service providers are not fully trusted. To enable the secure management of private user positions in non-trusted systems, we present two novel position sharing approaches based on the concept of multi-secret sharing. We improve existing geometric position sharing approaches by Durr et al. [2] and Skvortsov et al. [3] by considering continuous position updates and by increasing the robustness against various attacks. Furthermore, we present the first position sharing approach for symbolic location models. (C) 2013 Elsevier B.V. All rights reserved.
Adiabatic quantum computation has been proposed as quantum parallel processing with adiabatic evolution by using a superposition state to solve combinatorial optimization problem, then it has been applied to many prob...
详细信息
ISBN:
(纸本)9781479924189
Adiabatic quantum computation has been proposed as quantum parallel processing with adiabatic evolution by using a superposition state to solve combinatorial optimization problem, then it has been applied to many problems like satisfiability problem. Among them, Deutsch and Deutsch-Jozsa problems have been tried to be solved by using adiabatic quantum computation. In this paper, we modify the adiabatic quantum computation and propose to solve Simon problem more efficiently by a method with higher observation probability.
Large-scale interactive applications and realtime data-processing are facing problems with traditional disk-based storage solutions. Because of the often irregular access patterns they must keep almost all data in RAM...
详细信息
ISBN:
(纸本)9781479924189
Large-scale interactive applications and realtime data-processing are facing problems with traditional disk-based storage solutions. Because of the often irregular access patterns they must keep almost all data in RAM caches, which need to be manually synchronized with secondary storage and need a lot of time to be re-loaded in case of power outages. In this paper we propose a novel key-value storage keeping all data always in RAM by aggregating resources of potentially many nodes in a data center. We aim at supporting management of billions of small data objects (16-64 byte) like for example needed for storing graphs. A scalable low-overhead meta-data management is realized using a novel range-based ID approach combined with a super-overlay network. Furthermore, we provide persistence by a novel SSD-aware logging approach allowing to recover failed nodes very fast.
暂无评论