Goto

Collaborating Authors

 marquis


Counting and Reasoning with Plans

Speck, David, Hecher, Markus, Gnad, Daniel, Fichte, Johannes K., Corrêa, Augusto B.

arXiv.org Artificial Intelligence

Classical planning asks for a sequence of operators reaching a given goal. While the most common case is to compute a plan, many scenarios require more than that. However, quantitative reasoning on the plan space remains mostly unexplored. A fundamental problem is to count plans, which relates to the conditional probability on the plan space. Indeed, qualitative and quantitative approaches are well-established in various other areas of automated reasoning. We present the first study to quantitative and qualitative reasoning on the plan space. In particular, we focus on polynomially bounded plans. On the theoretical side, we study its complexity, which gives rise to rich reasoning modes. Since counting is hard in general, we introduce the easier notion of facets, which enables understanding the significance of operators. On the practical side, we implement quantitative reasoning for planning. Thereby, we transform a planning task into a propositional formula and use knowledge compilation to count different plans. This framework scales well to large plan spaces, while enabling rich reasoning capabilities such as learning pruning functions and explainable planning.


The ChatGPT Reincarnation of the Marquis de Sade Is Coming

WIRED

The first time I learned about "Loab," it sent shivers down my spine. A strange dead-eyed ghoul that began haunting an AI image generator last year, Loab reminded me of a fiend I'd been tracking for years. This may not seem like an obvious connection to make. The Marquis de Sade, one of the most infamous names in all of writing, was an 18th century French aristocrat, a man known for debauchery and evading authorities, breaking out of prison and eluding his own public execution in 1772. Loab is very much a product of the modern age, the accidental creation of artist Supercomposite, who claimed to have "discovered" her in an AI text-to-image generator in April of last year.


On the Complexity of Enumerating Prime Implicants from Decision-DNNF Circuits

de Colnet, Alexis, Marquis, Pierre

arXiv.org Artificial Intelligence

We consider the problem EnumIP of enumerating prime implicants of Boolean functions represented by decision decomposable negation normal form (dec-DNNF) circuits. We study EnumIP from dec-DNNF within the framework of enumeration complexity and prove that it is in OutputP, the class of output polynomial enumeration problems, and more precisely in IncP, the class of polynomial incremental time enumeration problems. We then focus on two closely related, but seemingly harder, enumeration problems where further restrictions are put on the prime implicants to be generated. In the first problem, one is only interested in prime implicants representing subset-minimal abductive explanations, a notion much investigated in AI for more than three decades. In the second problem, the target is prime implicants representing sufficient reasons, a recent yet important notion in the emerging field of eXplainable AI, since they aim to explain predictions achieved by machine learning classifiers. We provide evidence showing that enumerating specific prime implicants corresponding to subset-minimal abductive explanations or to sufficient reasons is not in OutputP.


Rectifying Mono-Label Boolean Classifiers

Coste-Marquis, Sylvie, Marquis, Pierre

arXiv.org Artificial Intelligence

We elaborate on the notion of rectification of a Boolean classifier $\Sigma$. Given $\Sigma$ and some background knowledge $T$, postulates characterizing the way $\Sigma$ must be changed into a new classifier $\Sigma \star T$ that complies with $T$ have already been presented. We focus here on the specific case of mono-label Boolean classifiers, i.e., there is a single target concept and any instance is classified either as positive (an element of the concept), or as negative (an element of the complementary concept). In this specific case, our main contribution is twofold: (1) we show that there is a unique rectification operator $\star$ satisfying the postulates, and (2) when $\Sigma$ and $T$ are Boolean circuits, we show how a classification circuit equivalent to $\Sigma \star T$ can be computed in time linear in the size of $\Sigma$ and $T$; when $\Sigma$ and $T$ are decision trees, a decision tree equivalent to $\Sigma \star T$ can be computed in time polynomial in the size of $\Sigma$ and $T$.


Intuit Accelerator Combines Fintech For Good With AI

#artificialintelligence

José V. Fernández first got interested in using technology to ramp up financial inclusion when he first arrived in New York City from Spain around 10 years ago. His job working as a trade officer for Spain didn't impress numerous prospective landlords, none of whom would rent him an apartment because he lacked a U.S. credit score. Finally, one company agreed to sign a one-year lease, but only if Fernández paid six months of his pricey Manhattan rent ahead of time. A few years after that, Fernández co-founded a fintech firm to open up microloans to unbanked people in West Africa. Then, last year, he founded Bankuish, which aims to give gig workers and freelancers a way to access banking services they normally wouldn't be able to tap.


On Quantifying Literals in Boolean Logic and Its Applications to Explainable AI

Darwiche, Adnan, Marquis, Pierre

arXiv.org Artificial Intelligence

This extends the reach of Boolean logic by enabling a variety of applications that have been explored over the decades. The existential quantification of literals (variable states) and its applications have also been studied in the literature. In this paper, we complement this by studying universal literal quantification and its applications, particularly to explainable AI. We also provide a novel semantics for quantification, discuss the interplay between variable/literal and existential/universal quantification. We further identify some classes of Boolean formulas and circuits on which quantification can be done efficiently. Literal quantification is more fine-grained than variable quantification as the latter can be defined in terms of the former. This leads to a refinement of quantified Boolean logic with literal quantification as its primitive.


Partition Function Estimation: A Quantitative Study

Agrawal, Durgesh, Pote, Yash, Meel, Kuldeep S

arXiv.org Artificial Intelligence

Probabilistic graphical models have emerged as a powerful modeling tool for several real-world scenarios where one needs to reason under uncertainty. A graphical model's partition function is a central quantity of interest, and its computation is key to several probabilistic reasoning tasks. Given the #P-hardness of computing the partition function, several techniques have been proposed over the years with varying guarantees on the quality of estimates and their runtime behavior. This paper seeks to present a survey of 18 techniques and a rigorous empirical study of their behavior across an extensive set of benchmarks. Our empirical study draws up a surprising observation: exact techniques are as efficient as the approximate ones, and therefore, we conclude with an optimistic view of opportunities for the design of approximate techniques with enhanced scalability. Motivated by the observation of an order of magnitude difference between the Virtual Best Solver and the best performing tool, we envision an exciting line of research focused on the development of portfolio solvers.


On Consensus in Belief Merging

Schwind, Nicolas (National Institute of Advanced Industrial Science and Technology) | Marquis, Pierre (CRIL-CNRS, Université d'Artois, Institut Universitaire de France)

AAAI Conferences

We define a consensus postulate in the propositional belief merging setting. In a nutshell, this postulate imposes the merged base to be consistent with the pieces of information provided by each agent involved in the merging process. The interplay of this new postulate with the IC postulates for belief merging is studied, and an incompatibility result is proved. The maximal sets of IC postulates which are consistent with the consensus postulate are exhibited. When satisfying some of the remaining IC postulates, consensus operators are shown to suffer from a weak inferential power. We then introduce two families of consensus operators having a better inferential power by setting aside some of these postulates.


Prime Compilation of Non-Clausal Formulae

Previti, Alessandro (University College Dublin) | Ignatiev, Alexey (INESC-ID, IST) | Morgado, Antonio (INESC-ID, IST) | Marques-Silva, Joao (INESC-ID, IST and University College Dublin)

AAAI Conferences

Formula compilation by generation of prime implicates or implicants finds a wide range of applications in AI. Recent work on formula compilation by prime implicate/implicant generation often assumes a Conjunctive/Disjunctive Normal Form (CNF/DNF) representation. However, in many settings propositional formulae are naturally expressed in non-clausal form. Despite a large body of work on compilation of non-clausal formulae, in practice existing approaches can only be applied to fairly small formulae, containing at most a few hundred variables. This paper describes two novel approaches for the compilation of non-clausal formulae either with prime implicants or implicates, that is based on propositional Satisfiability (SAT) solving. These novel algorithms also find application when computing all prime implicates of a CNF formula. The proposed approach is shown to allow the compilation of non-clausal formulae of size significantly larger than existing approaches.


Compiling Constraint Networks into Multivalued Decomposable Decision Graphs

Koriche, Frédéric (CRIL-CNRS and Université d'Artois) | Lagniez, Jean-Marie (CRIL-CNRS and Université d'Artois) | Marquis, Pierre (CRIL-CNRS and Université d'Artois) | Thomas, Samuel (CRIL-CNRS and Université d'Artois)

AAAI Conferences

Specifically, we present a top-down algorithm cn2mddg for compiling finite-domain CNs into multivalued decomposable We present and evaluate a top-down algorithm for decision graphs. The input of cn2mddg is a CN compiling finite-domain constraint networks (CNs) represented in the XCSP 2.1 format [Roussel and Lecoutre, into the language MDDG of multivalued decomposable 2009]. The output of our compilation algorithm is a representation decision graphs. Though it includes Decision-of the solutions of the CN in the language MDDG DNNF as a proper subset, MDDG offers the same key of multivalued decomposable decision graphs. MDDG is precisely tractable queries and transformations as Decision-the extension to non-Boolean domains of the language DNNF, which makes it useful for many applications. DDG [Fargier and Marquis, 2006] also known as Decision-Intensive experiments showed that our compiler DNNF [Oztok and Darwiche, 2014]: it is based on decomposable cn2mddg succeeds in compiling CNs which -nodes and (multivalued) decision nodes. Similarly are out of the reach of standard approaches based to Decision-DNNF, the MDDG language offers a number of on a translation of the input network to CNF, followed tractable queries, including (possibly weighted) solution finding by a compilation to Decision-DNNF. Furthermore, and counting, solution enumeration (solutions can be enumerated the sizes of the resulting compiled representations with polynomial delay), and optimization w.r.t. a linear turn out to be much smaller (sometimes by objective function. It also offers tractable transformations, several orders of magnitude).