\documentclass[a4paper]{article} \usepackage{hyperref} \hypersetup{ colorlinks=true, linkcolor=blue, } \usepackage{commath} \usepackage{amssymb} % % Local commands % \newcommand{\todo}[1]{{\color{orange}TODO #1}} \newcommand{\naf}{\ensuremath{\sim\!}} \newcommand{\larr}{\ensuremath{\leftarrow}} \newcommand{\at}[1]{\ensuremath{\!\del{#1}}} \newcommand{\co}[1]{\ensuremath{\overline{#1}}} \newcommand{\fml}[1]{\ensuremath{{\cal #1}}} \newcommand{\deft}[1]{\textbf{#1}} \newcommand{\pset}[1]{\ensuremath{\mathbb{P}\at{#1}}} \newcommand{\ent}{\ensuremath{\lhd}} \newcommand{\cset}[2]{\ensuremath{\set{#1,~#2}}} \newcommand{\langof}[1]{\ensuremath{\fml{L}\at{#1}}} \newcommand{\uset}[1]{\ensuremath{\left|{#1}\right>}} \newcommand{\lset}[1]{\ensuremath{\left<{#1}\right|}} \newcommand{\pr}[1]{\ensuremath{\mathrm{p}\at{#1}}} \newcommand{\given}{\ensuremath{~\middle|~}} \title{Zugzwang\\\textit{Logic and Artificial Intelligence}} \author{Francisco Coelho\\ \texttt{fc@uevora.pt}} \begin{document} \maketitle \begin{abstract} A major limitation of logical representations is the implicit assumption that the Background Knowledge (BK) is perfect. This assumption is problematic if data is noisy, which is often the case. Here we aim to explore how ASP specifications with probabilistic facts can lead to characterizations of probability functions on the specification's domain. \end{abstract} \section{Introduction and Motivation } Answer Set Programming (ASP) is a logic programming paradigm based on the Stable Model semantics of Normal Logic Programs (NP) that can be implemented using the latest advances in SAT solving technology. ASP is a truly declarative language that supports language constructs such as disjunction in the head of a clause, choice rules, and hard and weak constraints. The Distribution Semantics (DS) is a key approach to extend logical representations with probabilistic reasoning. Probabilistic Facts (PF) are the most basic stochastic DS primitive and they take the form of logical facts, $a$, labelled with a probability, such as $p::a$; Each probabilistic fact represents a boolean random variable that is true with probability $p$ and false with probability $1 - p$. A (consistent) combination of the PFs defines a \textit{total choice} $c = \set{p::a, \ldots}$ such that \begin{equation} P(c) = \prod_{a\in c} p \prod_{a \not\in c} (1- p). \end{equation} Our goal is to extend this probability, from total choices, to cover the specification domain. We can foresee two key applications of this extended probability: \begin{enumerate} \item Support any probabilistic reasoning/task on the specification domain. \item Also, given a dataset and a divergence measure, now the specification can be scored (by the divergence w.r.t. the \emph{empiric} distribution of the dataset), and sorted amongst other specifications. This is a key ingredient in algorithms searching, for example, an \textit{optimal specification} of the dataset. \end{enumerate} This goal faces a critical problem concerning situations where \textit{multiple} standard models result from a given total choice, illustrated by the following example. The specification $$ \begin{aligned} 0.3::a&,\cr b \vee c& \leftarrow a. \end{aligned} $$ has three stable models, $\set{\neg a}, \set{a, b}$ and $\set{a, c}$. While it is straightforward to set $P(\neg a)=0.7$, there is \textit{no further information} to assign values to $P(a,b)$ and $P(a,c)$. At best, we can use a parameter $\alpha$ such that $$ \begin{aligned} P(a,b) &= 0.3 \alpha,\cr P(a,c) &= 0.3 (1 - \alpha). \end{aligned} $$ This uncertainty in inherent to the specification, but can be mitigated with the help of a dataset: the parameter $\alpha$ can be estimated from the empirical distribution. In summary, if an ASP specification is intended to describe some system that can be observed then: \begin{enumerate} \item The observations can be used to estimate the value of the parameters (such as $\alpha$ above and others entailed from further clauses). \item With a probability set for the stable models, we want to extend it to all the samples (\textit{i.e.} consistent sets of literals) of the specification. \item This extended probability can then be related to the \textit{empirical distribution}, using a probability divergence, such as Kullback-Leibler; and the divergence value used as a \textit{performance} measure of the specification with respect to the observations. \item If that specification is only but one of many possible candidates then that performance measure can be used, \textit{e.g.} as fitness, by algorithms searching (optimal) specifications of a dataset of observations. \end{enumerate} Currently, we are on the step two above: Extending a probability function (with parameters such as $\alpha$), defined on the stable sets of a specification, to all the samples of the specification. This extension must, of course, respect the axioms of probability so that probabilistic reasoning is consistent with the ASP specification. \section{Extending Probabilities} Given an ASP specification, we look at the \textit{total choices} $\fml{C}$, \textit{stable models} $\fml{S}$, \textit{atoms} and \textit{literals} $\fml{A}, \fml{L}$, \textit{samples} $z \in \fml{Z} \iff z \subseteq \fml{L}$ and \textit{events} $\fml{E}$ (consistent samples). \begin{enumerate} \item \textbf{Given} $\pr{C = c}$ from the specification. Each total choice $C = c$ (together with the specification facts and rules) entails some stable models, $S_c$. \item \textbf{Given} $\pr{S = s \given C = c}$ such that its value is zero if $s \not\in S_c$. \item Each event $E = e$ either: \begin{enumerate} \item Is a stable model. Then \begin{equation} \pr{E = e \given C = c} = \pr{S = e \given C = c}. \end{equation} \item \textit{Is contained} in some stable models. Then \begin{equation} \pr{E = e \given C = c} = \sum_{s \supseteq e}\pr{S = s \given C = c}. \end{equation} \item \textit{Contains} some stable models. Then \begin{equation} \pr{E = e \given C = c} = \prod_{s \subseteq e}\pr{S = s \given C = c}. \end{equation} \item Neither contains nor is contained by a stable model. Then \begin{equation} \pr{E = e \given C = c} = 0. \end{equation} \end{enumerate} \item For each sample $Z = z$ \begin{equation} \pr{Z = z \given C = c} = \begin{cases} \pr{E = z \given C = c} & z \in \fml{E}, \cr 0 & \text{otherwise}. \end{cases} \end{equation} \end{enumerate} \begin{quotation}\marginpar{Todo} Prove the four event cases, and support the sum/product options, with the independence assumptions. \end{quotation} \section*{References} \begin{enumerate} \item Victor Verreet, Vincent Derkinderen, Pedro Zuidberg Dos Martires, Luc De Raedt, Inference and Learning with Model Uncertainty in Probabilistic Logic Programs (2022) \item Andrew Cropper, Sebastijan Dumancic, Richard Evans, Stephen H. Muggleton, Inductive logic programming at 30 (2021) \item Fabio Gagliardi Cozman, Denis Deratani Mauá, The joy of Probabilistic Answer Set Programming: Semantics - complexity, expressivity, inference (2020) \item Fabrizio Riguzzi, Foundations of Probabilistic Logic Programming Languages, Semantics, Inference and Learning. Rivers Publishers (2018) \item Martin Gebser, Roland Kaminski, Benjamin Kaufmann, and Torsten Schaub, Answer Set Solving in Practice, Morgan \& Claypool Publishers (2013) \end{enumerate} \end{document}