Petri nets are a class of models of computation used to compactly represent discrete event systems. Among many application domains, they have now become the most prominent formalism to express process models in Proces...
详细信息
ISBN:
(纸本)9783031742088;9783031742095
Petri nets are a class of models of computation used to compactly represent discrete event systems. Among many application domains, they have now become the most prominent formalism to express process models in Process Mining, thanks to their formal semantics that enables automated analysis techniques. In this context, model repair is the task of aligning a process model with actual executions of the process. Current solutions to model repair do not allow for embedding domain knowledge, providing guarantees of rigor, and enforcing structural requirements at the same time. In this paper, we fill this gap by proposing an approach based on the inductive logic programming system ILASP. We then implement our approach and perform an experimental evaluation, showing both its expressiveness and feasibility.
The goal of inductive logic programming is to learn a logic program that models the examples provided as input. The search space of the possible programs is constrained by a language bias, which defines the atoms and ...
详细信息
This paper presents a novel approach combining inductive logic programming with reinforcement learning to improve training performance and explainability. We exploit inductive learning of answer set programs from...
详细信息
The employment of Machine Learning (ML) techniques in embedded systems has seen constant growth in recent years, especially for black-box ML techniques (such as Artificial Neural Networks (ANNs)). However, despite the...
详细信息
The employment of Machine Learning (ML) techniques in embedded systems has seen constant growth in recent years, especially for black-box ML techniques (such as Artificial Neural Networks (ANNs)). However, despite the successful employment of ML techniques in embedded environments, their performance potential is constrained by the limited computing resources of their embedded computers. Several hardware-based approaches were developed (e.g., using FPGAs and ASICs) to address the constraints of limited computing resources. The scope of this work focuses on improving the performance for inductive logic programming (ILP) on embedded environments. ILP is a powerful logic-based ML technique that uses logicprogramming to construct human-interpretable ML models, where those logic-based ML models are capable of describing complex and multi-relational concepts. In this work, we present a hardware-based approach that accelerates the hypothesis evaluation task for ILPs in embedded environments that use Description logic (DL) languages as their logic-based representation. In particular, we target the ALCQ((D)) language. According to experimental results (through an FPGA implementation), our presented approach has achieved speedups up to 48.7-fold for a disjunction of 32 concepts on 100 M individuals, where the baseline performance is the sequential CPU performance of the Raspberry Pi 4. For role and concrete role restrictions, the FPGA implementation achieved speedups up to 2.4-fold (for MIN cardinality role restriction on 1M role assertions);all FPGA implemented role and concrete role restrictions have achieved similar speedups. In the worst-case scenario, the FPGA implementation achieved either a similar or slightly better performance than the baseline (for all DL operations);the worst-case scenario resulted from using small datasets such as: using conjunction and disjunction on < 100 individuals, and using role and concrete (float/string) role restrictions on < 100,000 asse
Knowledge acquisition is a difficult, error-prone, and time-consuming task. The task of automatically improving an existing knowledge base using learning methods is addressed by the class of systems performing theory ...
详细信息
Knowledge acquisition is a difficult, error-prone, and time-consuming task. The task of automatically improving an existing knowledge base using learning methods is addressed by the class of systems performing theory refinement. This paper presents a system, FORTE (First-Order Revision of Theories from Examples), which refines first-order Horn-clause theories by integrating a variety of different revision techniques into a coherent whole. FORTE uses these techniques within a hill-climbing framework, guided by a global heuristic. It identifies possible errors in the theory and calls on a library of operators to develop possible revisions. The best revision is implemented, and the process repeats until no further revisions are possible. Operators are drawn from a variety of sources, including propositional theory refinement, first-order induction, and inverse resolution. FORTE is demonstrated in several domains, including logicprogramming and qualitative modelling.
Knowledge acquisition with machine learning techniques is a fundamental re-quirement for knowledge discovery from databases and data mining systems. Two techniquesin particular - inductive learning and theory revision...
详细信息
Knowledge acquisition with machine learning techniques is a fundamental re-quirement for knowledge discovery from databases and data mining systems. Two techniquesin particular - inductive learning and theory revision - have been used toward this end. Amethod that combines both approaches to effectively acquire theories (regularity) from a setof training examples is presented. inductive learning is used to acquire new regularity fromthe training examples; and theory revision is used to improve an initial theory. In addition, atheory preference criterion that is a combination of the MDL-based heuristic and the Laplaceestimate has been successfully employed in the selection of the promising theory. The resultingalgorithm developed by integrating inductive learning and theory revision and using the criterionhas the ability to deal with complex problems, obtaining useful theories in terms of its predictiveaccuracy.
A knowledge-based system uses its database (also known as its "theory") to produce answers to the queries it receives. Unfortunately, these answers may be incorrect if the underlying theory is faulty. Standa...
详细信息
A knowledge-based system uses its database (also known as its "theory") to produce answers to the queries it receives. Unfortunately, these answers may be incorrect if the underlying theory is faulty. Standard "theory revision" systems use a given set of "labeled queries" (each a query paired with its correct answer) to transform the given theory, by adding and/or deleting either rules and/or antecedents, into a related theory that is as accurate as possible. After formally defining the theory revision task, this paper provides both sample and computational complexity bounds for this process. It first specifies the number of labeled queries necessary to identify a revised theory whose error is close to minimal with high probability. It then considers the computational complexity of finding this best theory, and proves that, unless P = NP, no polynomial-time algorithm can identify this optimal revision, even given the exact distribution of queries, except in certain simple situations. It also shows that, except in such simple situations, no polynomial-time algorithm can produce a theory whose error is even close to (i.e., within a particular polynomial factor of) optimal. The first (sample complexity) results suggest reasons why theory revision can be more effective than learning from scratch, while the second (computational complexity) results explain many aspects of the standard theory revision systems, including the practice of hill-climbing to a locally-optimal theory, based on a given set of labeled queries. (C) 1999 Elsevier Science B.V. All rights reserved.
With the increasing prevalence of Machine Learning in everyday life, a growing number of people will be provided with Machine-Learned assessments on a regular basis. We believe that human users interacting with system...
详细信息
With the increasing prevalence of Machine Learning in everyday life, a growing number of people will be provided with Machine-Learned assessments on a regular basis. We believe that human users interacting with systems based on Machine-Learned classifiers will demand and profit from the systems' decisions being explained in an approachable and comprehensive way. We developed a general process framework for logic-rule-based classifiers facilitating mutual exchange between system and user. The framework constitutes a guideline for how a system can apply inductive logic programming in order to provide comprehensive explanations for classification choices and empowering users to evaluate and correct the system's decisions. It also includes users' corrections being integrated into the system's core logic rules via retraining in order to increase the overall performance of the human-computer system. The framework suggests various forms of explanations-like natural language argumentations, near misses emphasizing unique characteristics, or image annotations-to be integrated into the system.
We describe an inductive logic programming (ILP) approach called learning from failures. In this approach, an ILP system (the learner) decomposes the learning problem into three separate stages: generate, test, and co...
详细信息
We describe an inductive logic programming (ILP) approach called learning from failures. In this approach, an ILP system (the learner) decomposes the learning problem into three separate stages: generate, test, and constrain. In the generate stage, the learner generates a hypothesis (a logic program) that satisfies a set of hypothesis constraints (constraints on the syntactic form of hypotheses). In the test stage, the learner tests the hypothesis against training examples. A hypothesis fails when it does not entail all the positive examples or entails a negative example. If a hypothesis fails, then, in the constrain stage, the learner learns constraints from the failed hypothesis to prune the hypothesis space, i.e. to constrain subsequent hypothesis generation. For instance, if a hypothesis is too general (entails a negative example), the constraints prune generalisations of the hypothesis. If a hypothesis is too specific (does not entail all the positive examples), the constraints prune specialisations of the hypothesis. This loop repeats until either (i) the learner finds a hypothesis that entails all the positive and none of the negative examples, or (ii) there are no more hypotheses to test. We introduce Popper, an ILP system that implements this approach by combining answer set programming and Prolog. Popper supports infinite problem domains, reasoning about lists and numbers, learning textually minimal programs, and learning recursive programs. Our experimental results on three domains (toy game problems, robot strategies, and list transformations) show that (i) constraints drastically improve learning performance, and (ii) Popper can outperform existing ILP systems, both in terms of predictive accuracies and learning times.
Summarizing content contributed by individuals can be challenging, because people make different lexical choices even when describing the same events. However, there remains a significant need to summarize such conten...
详细信息
Summarizing content contributed by individuals can be challenging, because people make different lexical choices even when describing the same events. However, there remains a significant need to summarize such content. Examples include the student responses to post-class reflective questions, product reviews, and news articles published by different news agencies related to the same events. High lexical diversity of these documents hinders the system's ability to effectively identify salient content and reduce summary redundancy. In this paper, we overcome this issue by introducing an integer linear programming-based summarization framework. It incorporates a low-rank approximation to the sentence-word cooccurrence matrix to intrinsically group semantically similar lexical items. We conduct extensive experiments on datasets of student responses, product reviews, and news documents. Our approach compares favorably to a number of extractive baselines as well as a neural abstractive summarization system. The paper finally sheds light on when and why the proposed framework is effective at summarizing content with high lexical variety.
暂无评论