Recently, businesses have started using MapReduce as a popular computation framework for processing large amount of data, such as spam detection, and different data mining tasks, in both public and private clouds. Two...
详细信息
this paper presents BIGS the Big Image Data Analysis Toolkit, a software framework for large scale image processing and analysis over heterogeneous computing resources, such as those available in clouds, grids, comput...
详细信息
this paper presents BIGS the Big Image Data Analysis Toolkit, a software framework for large scale image processing and analysis over heterogeneous computing resources, such as those available in clouds, grids, computer clusters or throughout scattered computer resources (desktops, labs) in an opportunistic manner. through BIGS, eScience for image processing and analysis is conceived to exploit coarse grained parallelism based on data partitioning and parameter sweeps, avoiding the need of inter-process communication and, therefore, enabling loosely coupled computing nodes (BIGS workers). It adopts an uncommitted resource allocation model where (1) experimenters define their image processing pipelines in a simple configuration file, (2) a schedule of jobs is generated and (3) workers, as they become available, take over pending jobs as long as their dependency on other jobs is fulfilled. BIGS workers act autonomously, querying the job schedule to determine which one to take over. this removes the need for a central scheduling node, requiring only access by all workers to a shared information source. Furthermore, BIGS workers are encapsulated within different technologies to enable their agile deployment over the available computing resources. Currently they can be launched through the Amazon EC2 service over their cloud resources, through Java Web Start from any desktop computer and through regular scripting or SSH commands. this suits well different kinds of research environments, both when accessing dedicated computing clusters or clouds with committed computing capacity or when using opportunistic computing resources whose access is seldom or cannot be provisioned in advance. We also adopt a NoSQL storage model to ensure the scalability of the shared information sources required by all workers, including within BIGS support for HBase and Amazon's DynamoDB service. Overall, BIGS now enables researchers to run large scale image processing pipelines in an easy, affo
In this paper, we assess the biological accuracy and relevance of XenoCluster results compared to a reported set of known xenologs in yeast (Hall et al. Eukaryot. Cell 4(6):1102-1115, 2005). We were able to assign a h...
详细信息
In this paper, we assess the biological accuracy and relevance of XenoCluster results compared to a reported set of known xenologs in yeast (Hall et al. Eukaryot. Cell 4(6):1102-1115, 2005). We were able to assign a high-priority rank to all reported yeast xenologs for which we had sufficient genomic sequence data. Additionally, we now report on a fine-grained extension to our previous implementation (Walters et al. 8thinternationalparallelcomputingtechnologiesconference (PaCT), LNCS, vol. 3606, pp. 355-366, Springer, Berlin, 2005), in which we achieved 80% efficiency on 8 CPU cores. In the coarse-grained component, high degrees parallelism (up to 65x speedup on 4,096 processors) was reported. While an obvious candidate for coarse-grained implementation, this paper now describes a multiple granularity parallelism solution that includes exploitation of multicore shared memory nodes to address fine-grained aspects in the tree-clustering phase of our previous deployment of XenoCluster (Walters et al. 8thinternationalparallelcomputingtechnologiesconference (PaCT), LNCS, vol. 3606, pp. 355-366, Springer, Berlin, 2005).
this paper discusses a parallel immune algorithm (IA) for detection of lung cancer in chest X-ray images based on object shared space. the template matching method is combined to the algorithm and JavaSpaces is used a...
详细信息
Improvements in hardware for parallel shared-memory computing usually involve increments in the number of computing cores and in the amount of memory available for a given application. However, many shared-memory appl...
详细信息
Using effective scheduling strategies to improve turnaround time, slowdown, and utilization is an important consideration in large supercomputing environments. Since such machines have traditionally used non-preemptio...
详细信息
the proceedings contain 142 papers. the topics discussed include: distributed caching strategies in peer-to-peer systems;a power-aware based storage architecture for high performance computing;a distributed reconfigur...
ISBN:
(纸本)9780769545387
the proceedings contain 142 papers. the topics discussed include: distributed caching strategies in peer-to-peer systems;a power-aware based storage architecture for high performance computing;a distributed reconfigurable active SSD platform for data intensive applications;a parallel processing scheme for large-size sliding-window applications;algebraic models for the cube connected cycles and shuffle exchange graphs;a high-performance and energy-efficient virtually tagged stack cache architecture for multi-core environments;providing quality of science in volunteer computing;an effective dynamic scheduling runtime and tuning system for heterogeneous multi and many-core desktop platforms;on the social aspects of personalized ranking for web services;stable adaptive work-stealing for concurrent multi-core runtime systems;rule pattern parallelization of packet filters on multi-core environments;and memory-intensive applications on a many-core processor.
the proceedings contain 7 papers. the topics discussed include: memetic algorithm for web service selection;model-driven performance engineering for wireless sensor networks with feature modeling and event calculus;de...
ISBN:
(纸本)9781450307338
the proceedings contain 7 papers. the topics discussed include: memetic algorithm for web service selection;model-driven performance engineering for wireless sensor networks with feature modeling and event calculus;description and composition of bio-inspired design patterns: the gradient case;methods for self-organizing distributed software;protein structure prediction using particle swarm optimization and a distributedparallel approach;discrete optimization problem solving withthree variants of hybrid binary particle swarm optimization;and self-organized invasive parallel optimization.
Cloud computing grows rapidly as today's advanced information technology. However, by allowing outsourcing computation on the Cloud, users risk of disclosing privacy and obtaining forged results. these potential t...
详细信息
ISBN:
(纸本)9780769546001
Cloud computing grows rapidly as today's advanced information technology. However, by allowing outsourcing computation on the Cloud, users risk of disclosing privacy and obtaining forged results. these potential threats block sensitive applications to join the Cloud. In this paper, we characterize sensitive applications on the Cloud (SAND) problem and define two critical security requirements: confidentiality and verifiability. the former refers to the protection of sensitive programs/data from disclosing to other users or even the Cloud administrators. the latter concerns with user's capability to verify whether computing results are faithfully calculated. To address SAND, we propose a new Cloud model, Trusted Block as a Service (TBaaS), to provide a confidential and verifiable environment for each sensitive application. TBaaS limits Cloud provider's access of sensitive applications while granting user the ability to verify whether the computation is faithfully carried out. Moreover, it offers high flexibility and low performance overhead.
A new secret-sharing-based e-auction scheme is proposed. distributed bid opening is employed to protect bid privacy. It can achieve all the desired properties for sealed-bid auctions at a reasonable cost. Moreover, at...
详细信息
暂无评论