The importance of optimization and NP problems solving cannot be over emphasized. The usefulness and popularity of evolutionary computing methods are also well established. There are various types of evolutionary meth...
详细信息
ISBN:
(纸本)9781467387767
The importance of optimization and NP problems solving cannot be over emphasized. The usefulness and popularity of evolutionary computing methods are also well established. There are various types of evolutionary methods that arc mostly sequential, and some others have parallel implementation. We propose a method to parallelize Imperialist Competitive Algorithm (Multi-Population). The algorithm has been implemented with MPI on two platforms and have tested our algorithms on a shared- memory and message passing architecture. An outstanding performance is obtained, which indicates that the method is efficient concern to speed and accuracy. In the second step, the proposed algorithm is compared with a set of existing well known parallel algorithms and is indicated that it obtains more accurate solutions in a lower time.
Graph partitioning is used to solve the problem of distributing computations to a number of processors, in order to improve the performance of time consuming applications in parallel environments. A common approach to...
详细信息
ISBN:
(纸本)9781467387767
Graph partitioning is used to solve the problem of distributing computations to a number of processors, in order to improve the performance of time consuming applications in parallel environments. A common approach to solve this problem is based on a multilevel framework, where the graph is firstly coarsened to a smaller instance and then it is partitioned in a number of parts using recursive bisection (RB) based methods. However, in applications where initial fixed vertices are used to model additional constraints of the problem, RB based methods often fail to produce partitions of good quality. In this paper, we propose a new direct k-way greedy graph growing algorithm, called KGGGP, that overcomes this issue and succeeds to produce partition with better quality than RB while respecting the constraint of fixed vertices. In the experimental section, we present results which compare KGGGP against state-of-the-art methods for graphs available from the popular DIMACS'10 collection.
Background: Automated image analysis on virtual slides is evolving rapidly and will play an important role in the future of digital pathology. Due to the image size, the computational cost of processing whole slide im...
详细信息
Background: Automated image analysis on virtual slides is evolving rapidly and will play an important role in the future of digital pathology. Due to the image size, the computational cost of processing whole slide images (WSIs) in full resolution is immense. Moreover, image analysis requires well focused images in high magnification. methods: We present a system that merges virtual microscopy techniques, open source image analysis software, and distributedparallelprocessing. We have integrated the parallelprocessing framework JPPF, so batch processing can be performed distributed and in parallel. All resulting meta data and image data are collected and merged. As an example the system is applied to the specific task of image sharpness assessment. imageJ is an open source image editing and processing framework developed at the NIH having a large user community that contributes imageprocessing algorithms wrapped as plug-ins in a wide field of life science applications. We developed an imageJ plug-in that supports both basic interactive virtual microscope and batch processing functionality. For the application of sharpness inspection we employ an approach with non-overlapping tiles. Compute nodes retrieve image tiles of moderate size from the streaming server and compute the focus measure. Each tile is divided into small sub images to calculate an edge based sharpness criterion which is used for classification. The results are aggregated in a sharpness map. Results: Based on the system we calculate a sharpness measure and classify virtual slides into one of the following categories - excellent, okay, review and defective. Generating a scaled sharpness map enables the user to evaluate sharpness of WSIs and shows overall quality at a glance thus reducing tedious assessment work. Conclusions: Using sharpness assessment as an example, the introduced system can be used to process, analyze and parallelize analysis of whole slide images based on open source software.
In this paper, the design and implementation of a recently developed clustering algorithm NNCA [1], Nearest Neighhour Clustering Algorithm, is proposed in conjunction with a Fast K Nearest Neighbour (FKNN) strategy fo...
详细信息
ISBN:
(纸本)9780889866379
In this paper, the design and implementation of a recently developed clustering algorithm NNCA [1], Nearest Neighhour Clustering Algorithm, is proposed in conjunction with a Fast K Nearest Neighbour (FKNN) strategy for further reduction in processing time. The parallel algorithm (PNNCA) has the ability to cluster pixels of retinal images into those belonging to blood vessels and others not belonging to blood vessels in a reasonable time.
With the increasing importance of multiple multiplatform remote sensing missions, digital image registration has been applied into many fields, and specially plays a very important role in remotely sensed data process...
详细信息
ISBN:
(纸本)3540292357
With the increasing importance of multiple multiplatform remote sensing missions, digital image registration has been applied into many fields, and specially plays a very important role in remotely sensed data processing. Firstly a brief introduction of existing parallelmethods of wavelet-based global registration is given. And then the communication optimization for GP method is described. The optimized algorithm is named Group-Optimized-parallel (GOP for short). To find out the reason of occasionally lower efficiency of GOP than other methods, a more careful analysis is presented in theory and proved in experiments. Moreover, we give a quantitative criterion, called Remainder Items, to choose the best solution in different input conditions.
image registration is the process of geometrically aligning images taken from different sensors, viewpoints or instances in time. It plays a key role in the detection of defects or anomalies for automated visual inspe...
详细信息
image registration is the process of geometrically aligning images taken from different sensors, viewpoints or instances in time. It plays a key role in the detection of defects or anomalies for automated visual inspection. A multiagent distributed blackboard system has been developed for intensity-based image registration. The images are divided into segments and allocated to agents on separate processors, allowing parallel computation of a similarity metric that measures the degree of likeness between reference and sensed images after the application of a transform. The need for a dedicated control module is removed by coordination of agents via the blackboard. Tests show that additional agents increase speed, provided the communication capacity of the blackboard is not saturated. The success of the approach in achieving registration, despite significant misalignment of the original images, is demonstrated in the detection of manufacturing defects on screen-printed plastic bottles and printed circuit boards. (c) 2007 Elsevier B.V. All rights reserved.
This paper describes the implementation of a preconditioned CG (Conjugate Gradient) method on GPUs and evaluates the performance compared with CPUs. Our CG method utilizes SP (Splitting-Up) preconditioner, which is su...
详细信息
ISBN:
(纸本)9781467387767
This paper describes the implementation of a preconditioned CG (Conjugate Gradient) method on GPUs and evaluates the performance compared with CPUs. Our CG method utilizes SP (Splitting-Up) preconditioner, which is suitable for parallelprocessing because other dimensions except for one dimension are independent. In order to enhance the memory bandwidth to the global memory of GPUs, our implementation utilizes a pseudo matrix transposition before and after a tridiagonal matrix solver, which results in coalesced memory accesses. In addition, the number of pseudo matrix transpositions can be reduced to only one by using a rotation configuration technique. By these techniques, the speedups of our approach can be enhanced by up to 102.2%.
Purpose - Content-based image retrieval (CBIR) technologies offer many advantages over purely text-based image search. However, one of the drawbacks associated with CBIR is the increased computational cost arising fro...
详细信息
Purpose - Content-based image retrieval (CBIR) technologies offer many advantages over purely text-based image search. However, one of the drawbacks associated with CBIR is the increased computational cost arising from tasks such as imageprocessing, feature extraction, image classification, and object detection and recognition. Consequently CBIR systems have suffered from a lack of scalability, which has greatly hampered their adoption for real-world public and commercial image search. At the same time, paradigms for large-scale heterogeneous distributed computing such as grid computing, cloud computing, and utility-based computing are gaining traction as a way of providing more scalable and efficient solutions to large-scale computing tasks. Design/methodology/approach - This paper presents an approach in which a large distributedprocessing grid has been used to apply a range of CBIR methods to a substantial number of images. By massively distributing the required computational task across thousands of grid nodes, very high through-put has been achieved at relatively low overheads. Findings - This has allowed one to analyse and index about 25 million high resolution images thus far, while using just two servers for storage and job submission. The CBIR system was developed by Imense Ltd and is based on automated analysis and recognition of image content using a semantic ontology. It features a range of image-processing and analysis modules, including image segmentation, region classification, scene analysis, object detection, and face recognition methods. Originality/value - In the case of content-based image analysis, the primary performance criterion is the overall through-put achieved by the system in terms of the number of images that can be processed over a given time frame, irrespective of the time taken to process any given image. As such, grid processing has great potential for massively parallel content-based image retrieval and other tasks with similar p
In this paper, we investigate how patterns could be used in order to generate Event-B refinements automatically through DSL(s) for temporal, timed or distribution patterns. Our ulimate goal is to generate code for a c...
详细信息
ISBN:
(纸本)9781509060580
In this paper, we investigate how patterns could be used in order to generate Event-B refinements automatically through DSL(s) for temporal, timed or distribution patterns. Our ulimate goal is to generate code for a concurrent, or distributed framework, e.g., BIP.
We present a parallel MPEG-2 video encoder on the Intel Paragon parallel computer. Given a video sequence or a set of sequences, the aim of the encoder is to achieve the maximum possible encoding rate. To achieve this...
详细信息
ISBN:
(纸本)0819429074
We present a parallel MPEG-2 video encoder on the Intel Paragon parallel computer. Given a video sequence or a set of sequences, the aim of the encoder is to achieve the maximum possible encoding rate. To achieve this aim, the parallel encoder works by a combined scheduling of processors, I/O nodes, and disks, enabling the system to work in a highly balanced fashion through matching of the encoding and I/O rates. An efficient data layout scheme for video frames is also proposed in order for I/O to sustain the desired data transfer rates. Using a small percentage of processors as the I/O nodes, the utilization of the system is also high. More importantly, our encoder is scalable and with an increase in the number of processors will result in a proportional increase in the encoding rate. Given any machine configuration (that is, the number of compute processors, I/O processors, and disks), our propose strategy can logically partition the system and match the I/O and encoding rates to reach the ideal encoding rate. The experimental results indicate about two-fold gain in performance compared to the previous studies. Our approach is useful for compressing a large video sequence or batches of sequences.
暂无评论