Expressions that involve matrices and vectors, known as linear algebra expressions, are commonly evaluated through a sequence of invocations to highly optimised kernels provided in libraries such as BLAS and LAPACK. A...
详细信息
ISBN:
(纸本)9781450397339
Expressions that involve matrices and vectors, known as linear algebra expressions, are commonly evaluated through a sequence of invocations to highly optimised kernels provided in libraries such as BLAS and LAPACK. A sequence of kernels represents an algorithm, and in general, because of associativity, algebraic identities, and multiple kernels, one expression can be evaluated via many different algorithms. These algorithms are all mathematically equivalent (i.e., in exact arithmetic, they all compute the same result), but often differ noticeably in terms of execution time. When faced with a decision, high-level languages, libraries, and tools such as Julia, Armadillo, and Linnea choose by selecting the algorithm that minimises the FLOP count. In this paper, we test the validity of the FLOP count as a discriminant for dense linear algebra algorithms, analysing "anomalies": problem instances for which the fastest algorithm does not perform the least number of FLOPs. To do so, we focused on relatively simple expressions and analysed when and why anomalies occurred. We found that anomalies exist and tend to cluster into large contiguous regions. For one expression anomalies were rare, whereas for the other they were abundant. We conclude that FLOPs is not a sufficiently dependable discriminant even when building algorithms with highly optimised kernels. Plus, most of the anomalies remained as such even after filtering out the inter-kernel cache effects. We conjecture that combining FLOP counts with kernel performance models will significantly improve our ability to choose optimal algorithms.
Medical image scans and associated electronic medical records (EMR) could be stored locally or transmitted for use in autodiagnosis and remote healthcare in teleradiology. Hence, they require security against unauthor...
详细信息
ISBN:
(纸本)9781728119908
Medical image scans and associated electronic medical records (EMR) could be stored locally or transmitted for use in autodiagnosis and remote healthcare in teleradiology. Hence, they require security against unauthorised access and modification. Among other means of providing this security, information hiding (IH) techniques have gained relevance especially for open networks that are prone to active attacks. However, the evaluation of the suitability of these IH algorithms in terms of preserving medical image diagnostic features is currently limited to signal processing parameters. This paper re-interprets existing evaluation parameters and provides a new framework that allows dynamic selection of medical image IH (watermarking and steganography) security algorithms. Specifically, criteria that capture medical statistics used in the diagnosis and monitoring of patients were incorporated. These criteria and framework were validated on the Pneumonia Chest Xray dataset (used in a Kaggle Competition) using three selected IH algorithms that offer privacy and image tamper detection.
Genetic improvement uses automated search to find improved versions of existing software. Software can either be evolved with general-purpose intentions or with a focus on a specific application (e.g., to improve it...
详细信息
ISBN:
(纸本)9781450367486
Genetic improvement uses automated search to find improved versions of existing software. Software can either be evolved with general-purpose intentions or with a focus on a specific application (e.g., to improve it's efficiency for a particular class of problems). Unfortunately, software specialisation to each problem application is generally performed independently, fragmenting and slowing down an already very time-consuming search process. We propose to incorporate specialisation as an online mechanism of the general search process, in an attempt to automatically devise application classes, by benefiting from past execution history.
Software verifiers have different strengths and weaknesses, depending on properties of the verification task. It is well-known that combinations of verifiers via portfolio and selection approaches can help to combine ...
详细信息
ISBN:
(纸本)9783030994297;9783030994280
Software verifiers have different strengths and weaknesses, depending on properties of the verification task. It is well-known that combinations of verifiers via portfolio and selection approaches can help to combine the strengths. In this paper, we investigate (a) how to easily compose such combinations from existing, 'off-the-shelf' verification tools without changing them and (b) how much performance improvement easy combinations can yield, regarding the effectiveness (number of solved problems) and efficiency (consumed resources). First, we contribute a method to systematically and conveniently construct verifier combinations from existing tools, using the composition framework coVERITEAM. We consider sequential portfolios, parallel portfolios, and algorithm selections. Second, we perform a large experiment on 8 883 verification tasks to show that combinations can improve the verification results without additional computational resources. All combinations are constructed from off-the-shelf verifiers, that is, we use them as published. The result of our work suggests that users of verification tools can achieve a significant improvement at a negligible cost (only configure our composition scripts).
GRAVES-CPA is a verification tool which uses algorithm selection to decide an ordering of underlying verifiers to most effectively verify a given program. GRAVES-CPA represents programs using an amalgam of traditional...
详细信息
ISBN:
(纸本)9783030995270;9783030995263
GRAVES-CPA is a verification tool which uses algorithm selection to decide an ordering of underlying verifiers to most effectively verify a given program. GRAVES-CPA represents programs using an amalgam of traditional program graph representations and uses stateof-the-art graph neural network techniques to dynamically decide how to run a set of verification techniques. The GRAVES technique is implementation agnostic, but it's competition submission, GRAVES-CPA, is built using several CPAchecker configurations as its underlying verifiers.
Using a knowledge discovery approach, we seek insights into the relationships between problem structure and the effectiveness of scheduling heuristics. A large collection of 75,000 instances of the single machine earl...
详细信息
ISBN:
(纸本)9783642111686
Using a knowledge discovery approach, we seek insights into the relationships between problem structure and the effectiveness of scheduling heuristics. A large collection of 75,000 instances of the single machine early/tardy scheduling problem is generated, characterized by six features, and used to explore the performance of two common scheduling heuristics. The best heuristic is selected using rules from a decision tree with accuracy exceeding 97%. A self-organizing map is used to visualize the feature space and generate insights into heuristic performance. This paper argues for such a knowledge discovery approach to be applied to other optimization problems, to contribute to automation of algorithm selection as well as insightful algorithm design.
This article proposes the MASSCAH method realization for Apache Spark clustering algorithms selection and configuration. Optimization of one of the clustering quality measures is used to configure the algorithm. In th...
详细信息
ISBN:
(纸本)9781450385862
This article proposes the MASSCAH method realization for Apache Spark clustering algorithms selection and configuration. Optimization of one of the clustering quality measures is used to configure the algorithm. In the course of this study, additional clustering quality measures were implemented that are not included in the Apache Spark framework, since at the moment only the silhouette criterion is available in the framework.
Automated Machine Learning (autoML) is a novel topic that aims to tackle the parameter configuration issue using automatic monitoring models and comprises different machine learning tasks, such as feature selection, m...
详细信息
This article proposes the MASSCAH method realization for Apache Spark clustering algorithms selection and configuration. Optimization of one of the clustering quality measures is used to configure the algorithm. In th...
详细信息
This article proposes the MASSCAH method realization for Apache Spark clustering algorithms selection and configuration. Optimization of one of the clustering quality measures is used to configure the algorithm. In the course of this study, additional clustering quality measures were implemented that are not included in the Apache Spark framework, since at the moment only the silhouette criterion is available in the framework.
A data driven approach is an emerging paradigm for the handling of analytic prob- lems. In this paradigm the mantra is to let the data speak freely. However, when using machine learning algorithms, the data does not n...
详细信息
A data driven approach is an emerging paradigm for the handling of analytic prob- lems. In this paradigm the mantra is to let the data speak freely. However, when using machine learning algorithms, the data does not naturally reveal the best or even a good approach for algorithm choice. One method to let the algorithm reveal itself is through the use of Meta Learning, which uses the features of a dataset to determine a useful model to represent the entire dataset. This research proposes an improve- ment on the meta-model recommendation system by adding classification problems to the candidate problem space with appropriate evaluation metrics for these additional problems. This research predicts the relative performance of six machine learning algorithms using support vector regression with a radial basis function as the meta learner. Six sets of data of various complexity are explored using this recommendation system and at its best, the system recommends the best algorithm 67% of the time and a "good" algorithm from 67% to 100% of the time depending on how "good" is defined.
暂无评论