probabilisticlogicprogramming combines uncertainty and logic-based languages. Liftable probabilisticlogic Programs have been recently proposed to perform inference in a lifted way. LIFTCOVER is an algorithm used to...
详细信息
ISBN:
(数字)9783031492990
ISBN:
(纸本)9783031492983;9783031492990
probabilisticlogicprogramming combines uncertainty and logic-based languages. Liftable probabilisticlogic Programs have been recently proposed to perform inference in a lifted way. LIFTCOVER is an algorithm used to perform parameter and structure learning of liftable probabilisticlogic programs. In particular, it performs parameter learning via Expectation Maximization and LBFGS. In this paper, we present an updated version of LIFTCOVER, called LIFTCOVER+, in which regularization was added to improve the quality of the solutions and LBFGS was replaced by gradient descent. We tested LIFTCOVER+ on the same 12 datasets on which LIFTCOVER was tested and compared the performances in terms of AUC-ROC, AUC-PR, and execution times. Results show that in most cases Expectation Maximization with regularization improves the quality of the solutions.
probabilistic inductive logic programming (PILP) is a statistical relational learning technique which extends inductivelogicprogramming by considering probabilistic data. The ability to use probabilities to represen...
详细信息
probabilistic inductive logic programming (PILP) is a statistical relational learning technique which extends inductivelogicprogramming by considering probabilistic data. The ability to use probabilities to represent uncertainty comes at the cost of an exponential evaluation time when composing theories to model the given problem. For this reason, PILP systems rely on various pruning strategies in order to reduce the search space. However, to the best of the authors' knowledge, there has been no systematic analysis of the different pruning strategies, how they impact the search space and how they interact with one another. This work presents a unified representation for PILP pruning strategies which enables end-users to understand how these strategies work both individually and combined and to make an informed decision on which pruning strategies to select so as to best achieve their goals. The performance of pruning strategies is evaluated both time and quality-wise in two state-of-the-art PILP systems with datasets from three different domains. Besides analysing the performance of the pruning strategies, we also illustrate the utility of PILP in one of the application domains, which is a real-world application.
The main goal of this work is to propose a tool-chain capable of analyzing a data collection of temporally qualified (genetic) mutation profiles, i.e., a collection of DNA-sequences (genes) that present variations wit...
详细信息
probabilisticlogicprogramming (PLP) provides a powerful tool for reasoning with uncertain relational models. However, learning probabilisticlogic programs is expensive due to the high cost of inference. Among the p...
详细信息
probabilisticlogicprogramming (PLP) provides a powerful tool for reasoning with uncertain relational models. However, learning probabilisticlogic programs is expensive due to the high cost of inference. Among the proposals to overcome this problem, one of the most promising is lifted inference. In this paper we consider PLP models that are amenable to lifted inference and present an algorithm for performing parameter and structure learning of these models from positive and negative examples. We discuss parameter learning with EM and LBFGS and structure learning with LIFTCOVER, an algorithm similar to SLIPCOVER. The results of the comparison of LIFTCOVER with SLIPCOVER on 12 datasets show that it can achieve solutions of similar or better quality in a fraction of the time.
PRISM is a probabilisticlogicprogramming formalism which allows defining a probability distribution over possible worlds. This paper investigates learning a class of generative PRISM programs known as failure-free. ...
详细信息
PRISM is a probabilisticlogicprogramming formalism which allows defining a probability distribution over possible worlds. This paper investigates learning a class of generative PRISM programs known as failure-free. The aim is to learn recursive PRISM programs which can be used to model stochastic processes. These programs generalise dynamic Bayesian networks by defining a halting distribution over the generative process. Dynamic Bayesian networks model infinite stochastic processes. Sampling from infinite process can only be done by specifying the length of sequences that the process generates. In this case, only observations of a fixed length of sequences can be obtained. On the other hand, the recursive PRISM programs considered in this paper are self-terminating upon some halting conditions. Thus, they generate observations of different lengths of sequences. The direction taken by this paper is to combine ideas from inductivelogicprogramming and learning Bayesian networks to learn PRISM programs. It builds upon the inductivelogicprogramming approach of learning from entailment. (C) 2015 Elsevier Inc. All rights reserved.
Learning probabilisticlogicprogramming languages is receiving an increasing attention, and systems are available for learning the parameters (PRISM, LeProbLog, LFI-ProbLog and EMBLEM) or both structure and parameter...
详细信息
Learning probabilisticlogicprogramming languages is receiving an increasing attention, and systems are available for learning the parameters (PRISM, LeProbLog, LFI-ProbLog and EMBLEM) or both structure and parameters (SEM-CP-logic and SLIPCASE) of these languages. In this paper we present the algorithm SLIPCOVER for "Structure LearnIng of probabilisticlogic programs by searChing OVER the clause space." It performs a beam search in the space of probabilistic clauses and a greedy search in the space of theories using the log likelihood of the data as the guiding heuristics. To estimate the log likelihood, SLIPCOVER performs Expectation Maximization with EMBLEM. The algorithm has been tested on five real world datasets and compared with SLIPCASE, SEM-CP-logic, Aleph and two algorithms for learning Markov logic Networks (Learning using Structural Motifs (LSM) and ALEPH++ExactL1). SLIPCOVER achieves higher areas under the precision-recall and receiver operating characteristic curves in most cases.
probabilistic inductive logic programming (PILP) is a relatively unexplored area of Statistical Relational Learning which extends classic inductivelogicprogramming (ILP). Within this scope, we introduce SkILL, a Sto...
详细信息
ISBN:
(纸本)9781509002870
probabilistic inductive logic programming (PILP) is a relatively unexplored area of Statistical Relational Learning which extends classic inductivelogicprogramming (ILP). Within this scope, we introduce SkILL, a Stochastic inductivelogic Learner, which takes probabilistic annotated data and produces First Order logic (FOL) theories. Data in several domains such as medicine and bioinformatics have an inherent degree of uncertainty, and because SkILL can handle this type of data, the models produced for these areas are closer to reality. SkILL can then use probabilistic data to extract non-trivial knowledge from databases, and also address efficiency issues by introducing an efficient search strategy for finding hypotheses in PILP environments. SkILL's capabilities are demonstrated using a real world medical dataset in the breast cancer domain.
We propose a framework for reasoning about dynamic Web data, based on probabilistic Answer Set programming (ASP). Our approach, which is prototypically implemented, allows for the annotation of first-order formulas as...
详细信息
ISBN:
(纸本)9783319111131;9783319111124
We propose a framework for reasoning about dynamic Web data, based on probabilistic Answer Set programming (ASP). Our approach, which is prototypically implemented, allows for the annotation of first-order formulas as well as ASP rules and facts with probabilities, and for learning of such weights from examples (parameter estimation). Knowledge as well as examples can be provided incrementally in the form of RDF data streams. Optionally, stream data can be configured to decay over time. With its hybrid combination of various contemporary AI techniques, our framework aims at prevalent challenges in relation to data streams and Linked Data, such as inconsistencies, noisy data, and probabilistic processing rules.
Recently much work in Machine Learning has concentrated on using expressive representation languages that combine aspects of logic and probability. A whole field has emerged, called Statistical Relational Learning, ri...
详细信息
Recently much work in Machine Learning has concentrated on using expressive representation languages that combine aspects of logic and probability. A whole field has emerged, called Statistical Relational Learning, rich of successful applications in a variety of domains. In this paper we present a Machine Learning technique targeted to probabilisticlogic Programs, a family of formalisms where uncertainty is represented using logicprogramming tools. Among various proposals for probabilisticlogicprogramming, the one based on the distribution semantics is gaining popularity and is the basis for languages such as ICL, PRISM, ProbLog and logic Programs with Annotated Disjunctions. This paper proposes a technique for learning parameters of these languages. Since their equivalent Bayesian networks contain hidden variables, an Expectation Maximization (EM) algorithm is adopted. In order to speed the computation up, expectations are computed directly on the Binary Decision Diagrams that are built for inference. The resulting system, called EMBLEM for "EM over Bdds for probabilisticlogic programs Efficient Mining", has been applied to a number of datasets and showed good performances both in terms of speed and memory usage. In particular its speed allows the execution of a high number of restarts, resulting in good quality of the solutions.
In this paper we propose to apply the Information Bottleneck (IB) approach to the sub-class of Statistical Relational Learning (SRL) languages that are reducible to Bayesian networks. When the resulting networks invol...
详细信息
In this paper we propose to apply the Information Bottleneck (IB) approach to the sub-class of Statistical Relational Learning (SRL) languages that are reducible to Bayesian networks. When the resulting networks involve hidden variables, learning these languages requires the use of techniques for learning from incomplete data such as the Expectation Maximization (EM) algorithm. Recently, the IB approach was shown to be able to avoid some of the local maxima in which EM can get trapped when learning with hidden variables. Here we present the algorithm Relational Information Bottleneck (RIB) that learns the parameters of SRL languages reducible to Bayesian Networks. In particular, we present the specialization of RIB to a language belonging to the family of languages based on the distribution semantics, logic Programs with Annotated Disjunction (LPADs). This language is prototypical for such a family and its equivalent Bayesian networks contain hidden variables. RIB is evaluated on the IMDB, Cora and artificial datasets and compared with LeProbLog, EM, Alchemy and PRISM. The experimental results show that RIB has good performances especially when some logical atoms are unobserved. Moreover, it is particularly suitable when learning from interpretations that share the same Herbrand base.
暂无评论