Probabilistically Estimating Backbones and Variable Bias: Experimental Overview

Similar documents
A Bayesian Approach to Tackling Hard Computational Challenges

CS 4365: Artificial Intelligence Recap. Vibhav Gogate

WEIGHTS STAGNATION IN DYNAMIC LOCAL SEARCH FOR SAT

Procedures. Abstract. Introduction

A Monogenous MBO Approach to Satisfiability

[1] provides a philosophical introduction to the subject. Simon [21] discusses numerous topics in economics; see [2] for a broad economic survey.

Reinforcement Learning : Theory and Practice - Programming Assignment 1

Lec 02: Estimation & Hypothesis Testing in Animal Ecology

Improving implementation of SLS solvers for SAT and new heuristics for k-sat with long clauses

Using Bayesian Networks to Direct Stochastic Search in Inductive Logic Programming

Learning with Rare Cases and Small Disjuncts

MBO: Marriage in Honey Bees Optimization A Haplometrosis Polygynous Swarming Approach

Handling Partial Preferences in the Belief AHP Method: Application to Life Cycle Assessment

LocalSearchCharacteristicsof IncompleteSATProcedures

A Brief Introduction to Bayesian Statistics

Hierarchical Bayesian Modeling of Individual Differences in Texture Discrimination

ERA: Architectures for Inference

Artificial Intelligence Programming Probability

A HMM-based Pre-training Approach for Sequential Data

Learning to Use Episodic Memory

Logistic Regression and Bayesian Approaches in Modeling Acceptance of Male Circumcision in Pune, India

Human and Optimal Exploration and Exploitation in Bandit Problems

Decisions and Dependence in Influence Diagrams

Bayesian Networks in Medicine: a Model-based Approach to Medical Decision Making

Lecture 2: Foundations of Concept Learning

Hybrid HMM and HCRF model for sequence classification

The Dynamics of Two Cognitive Heuristics for Coordination on Networks

Stepwise Knowledge Acquisition in a Fuzzy Knowledge Representation Framework

On the Representation of Nonmonotonic Relations in the Theory of Evidence

CS221 / Autumn 2017 / Liang & Ermon. Lecture 19: Conclusion

Assessing the Foundations of Conscious Computing: A Bayesian Exercise

A Comparison of Collaborative Filtering Methods for Medication Reconciliation

arxiv: v3 [stat.ml] 27 Mar 2018

AN INFERENCE TECHNIQUE FOR INTEGRATING KNOWLEDGE FROM DISPARATE SOURCES. Thomas D. Garvey, John D. Lowrance, and Martin A.

Weight-Enhanced Diversification in Stochastic Local Search for Satisfiability

A Model of Multi-Agent Consensus for Vague and Uncertain Beliefs

Analysis of complex patterns of evidence in legal cases: Wigmore charts vs. Bayesian networks

Models of Information Retrieval

Evolutionary Programming

MS&E 226: Small Data

Predicting Breast Cancer Survivability Rates

Aggressive but Reactive Clause Database Management, Dynamic Restarts. Gilles Audemard Laurent Simon GLUCOSE 2.1. POS 16 june 2012

TEACHING YOUNG GROWNUPS HOW TO USE BAYESIAN NETWORKS.

Visual book review 1 Safe and Sound, AI in hazardous applications by John Fox and Subrata Das

ISC- GRADE XI HUMANITIES ( ) PSYCHOLOGY. Chapter 2- Methods of Psychology

Equilibrium Selection In Coordination Games

Analyzing Spammers Social Networks for Fun and Profit


An Experimental Analysis of Anytime Algorithms for Bayesian Network Structure Learning. Colin Lee and Peter van Beek University of Waterloo

Statistical Audit. Summary. Conceptual and. framework. MICHAELA SAISANA and ANDREA SALTELLI European Commission Joint Research Centre (Ispra, Italy)

Advanced Bayesian Models for the Social Sciences

How Does Analysis of Competing Hypotheses (ACH) Improve Intelligence Analysis?

P(HYP)=h. P(CON HYP)=p REL. P(CON HYP)=q REP

Chapter 1. Introduction

The 29th Fuzzy System Symposium (Osaka, September 9-, 3) Color Feature Maps (BY, RG) Color Saliency Map Input Image (I) Linear Filtering and Gaussian

Minimal Change and Maximal Coherence: A Basis for Belief Revision and Reasoning about Actions

Chapter 2. Knowledge Representation: Reasoning, Issues, and Acquisition. Teaching Notes

Iterative Join Graph Propagation

CHAPTER II CONCEPTUAL FRAMEWORK

Likelihood Ratio Based Computerized Classification Testing. Nathan A. Thompson. Assessment Systems Corporation & University of Cincinnati.

The Lens Model and Linear Models of Judgment

Representation Theorems for Multiple Belief Changes

COHERENCE: THE PRICE IS RIGHT

PERSONALIZED medicine is a healthcare paradigm that. ConfidentCare: A Clinical Decision Support System for Personalized Breast Cancer Screening

MMI 409 Spring 2009 Final Examination Gordon Bleil. 1. Is there a difference in depression as a function of group and drug?

Supplement 2. Use of Directed Acyclic Graphs (DAGs)

Omicron ACO. A New Ant Colony Optimization Algorithm

Sparse Coding in Sparse Winner Networks

Positive and Unlabeled Relational Classification through Label Frequency Estimation

Improving stochastic local search for SAT with a new probability distribution

Advanced Bayesian Models for the Social Sciences. TA: Elizabeth Menninga (University of North Carolina, Chapel Hill)

Fuzzy Decision Tree FID

Citation for published version (APA): Geus, A. F. D., & Rotterdam, E. P. (1992). Decision support in aneastehesia s.n.

International Journal of Software and Web Sciences (IJSWS)

A Decision-Theoretic Approach to Evaluating Posterior Probabilities of Mental Models

A Revisionist View of Blackboard Systems

Improving the Accuracy of Neuro-Symbolic Rules with Case-Based Reasoning

Clustering data based on a measure of

Measuring Performance Of Physicians In The Diagnosis Of Endometriosis Using An Expectation-Maximization Algorithm

The Scientific Method

Approximate Sensor Interpretation

accuracy (see, e.g., Mislevy & Stocking, 1989; Qualls & Ansley, 1985; Yen, 1987). A general finding of this research is that MML and Bayesian

Journal of Political Economy, Vol. 93, No. 2 (Apr., 1985)

Measuring the User Experience

Book review of Herbert I. Weisberg: Bias and Causation, Models and Judgment for Valid Comparisons Reviewed by Judea Pearl

Symbolic CTL Model Checking

Intelligent Techniques for Configuration Knowledge Evolution

Winner s Report: KDD CUP Breast Cancer Identification

The Semantics of Intention Maintenance for Rational Agents

Using Association Rule Mining to Discover Temporal Relations of Daily Activities

Collaborative Filtering with Multi-component Rating for Recommender Systems

Technical Specifications

Science is a way of learning about the natural world by observing things, asking questions, proposing answers, and testing those answers.

Behavioural Economics University of Oxford Vincent P. Crawford Michaelmas Term 2012

Conditional spectrum-based ground motion selection. Part II: Intensity-based assessments and evaluation of alternative target spectra

Progress in Risk Science and Causality

Learning to Identify Irrelevant State Variables

Hoare Logic and Model Checking. LTL and CTL: a perspective. Learning outcomes. Model Checking Lecture 12: Loose ends

1.1 FEATURES OF THOUGHT

Positive and Unlabeled Relational Classification through Label Frequency Estimation

Transcription:

Probabilistically Estimating Backbones and Variable Bias: Experimental Overview Eric I. Hsu, Christian J. Muise, J. Christopher Beck, and Sheila A. McIlraith Department of Computer Science, University of Toronto {eihsu,cjmuise,sheila}@cs.toronto.edu,jcb@mie.utoronto.ca Abstract. Backbone variables have the same assignment in all solutions to a given constraint satisfaction problem; more generally, bias represents the proportion of solutions that assign a variable a particular value. Intuitively such constructs would seem important to efficient search, but their study to date has been from a mostly conceptual perspective, in terms of indicating problem hardness or motivating and interpreting heuristics. Here we summarize a two-phase project where we first measure the ability of both existing and novel probabilistic message-passing techniques to directly estimate bias and identify backbones for the Boolean Satisfiability (SAT) Problem. We confirm that methods like Belief Propagation and Survey Propagation plus Expectation Maximization-based variants do produce good estimates with distinctive properties. The second phase demonstrates the use of bias estimation within a modern SAT solver, exhibiting a correlation between accurate, stable, estimates and successful backtracking search. The same process also yields a family of search heuristics that can dramatically improve search efficiency for the hard random problems considered. 1 Introduction Probabilistic message-passing algorithms like Survey Propagation (SP) and Belief Propagation (BP), plus variants based on Expectation Maximization (EM), have proved very successful for random SAT and CSP problems [2 4]. This success would appear to result from the ability to implicitly sample from the space of solutions and thus estimate variable bias: the percentages of solutions that have a given variable set true or false. However, such bias estimation ability has never been measured directly, and its actual usefulness to heuristic search has also escaped systematic study. Similarly, backbones, or variables that must be set a certain way in any solution to a given problem, have also drawn a good deal of recent interest [5 9], but they have not been directly targeted for discovery within arbitrary problems. Since backbones must have 100% positive or negative bias, bias determination generalizes the task of backbone identification. Isolating the performance of probabilistic techniques when applied to bias estimation improves our understanding of both the estimators and of bias itself, ultimately directing the design of a complete problem-solving system. Thus the first stage of our study compares the basic accuracy of six message-passing techniques and two control methods when applied to hard, random, satisfiable SAT problems as stand-alone bias estimators. The second stage assesses how such comparisons translate when we move the This article summarizes a more detailed description that is available as a technical report [1].

2 Probabilistically Estimating Backbones and Variable Bias: Experimental Overview algorithms into the realm of full-featured search, by embedding them as variable/value ordering heuristics within the MiniSat solver [10]. While it is intuitive that bias should relate to how we set variables during search, it is not obvious that bias should be a key to efficiency in the presence of modern features like restarts and clause learning. 2 Definitions Definition 1 (SAT instance) A (CNF) SAT instance is a set C of m clauses, constraining a set V of n Boolean variables. Each clause c C is a disjunction of literals built from the variables in V. An assignment X {0, 1} n to the variables satisfies the instance if it makes at least one literal true in each clause. Definition 2 (Bias, Survey) For a satisfiable SAT instance F, the estimated bias distribution θ v of a variable v attempts to represent the fraction of solutions to F wherein v appears positively or negatively. Thus it consists of a positive bias θ + v and a negative bias θ v, where θ + v, θ v [0, 1] and θ + v + θ v = 1. A vector of bias distributions, one for each variable in a theory, will be called a survey, denoted Θ(F). Equivalently, it can be useful to think of a variable s bias as the probability of finding the variable set positively or negatively when randomly sampling from the space of satisfying assignments. Less formally, the strength of a bias distribution indicates the margin by which it favors one value over the other. 3 Probabilistic Methods for Estimating Bias We compare six distinct message-passing techniques for measuring variable bias: Belief Propagation (BP), EM Belief Propagation-Local/Global (EMBP-L and EMBP-G), Survey Propagation (SP), and EM Survey Propagation-Local/Global (EMSP-L and EMSP- G). On receiving a SAT instance F, each of the propagation methods begins by formulating an initial survey at random. Each algorithm proceeds to successively refine its estimates over multiple iterations. An iteration consists of a single pass through all variables, where the bias for each variable is updated with respect to the other variables biases, according to the characteristic rule for a method. If no variable s bias has changed between two successive iterations, the process ends with convergence; otherwise a method terminates by timeout or some other parameter. EM-type methods are convergent, or guaranteed to converge naturally, while regular BP and SP are not [4]. BP estimates bias via Pearl s Belief Propagation, also known as the Sum-Product algorithm [11, 12]. SP (Survey Propagation) extends BP to measure not only the probabilities of a variable being positively or negatively constrained in a solution, but also the probability that it could have been set either way [2]. Such extra sensitivity changes the dynamics between iterations, but in the final survey any mass for the third joker state is evenly divided between the positive and the negative for the purposes of estimating bias. Both methods can be re-formulated within the Expectation Maximization framework [13], producing the four EM-based methods. EMBP-L and EMBP-G use the two-state model of BP and differ in exploiting variational approximations [14] based

Probabilistically Estimating Backbones and Variable Bias: Experimental Overview 3 on local (generalized arc-) consistency and global consistency, respectively. Similarly, EMSP-L and EMSP-G apply local or global consistency to the three-state model represented by SP. Global methods embody a tighter bound on the variational approximation, but in general take longer to compute than local methods. For experimental comparison, we created two non-probabilistic control methods. LC ( Literal Count ) greatly simplifies the backbone-inspired heuristic at the core of a highly successful system for refuting unsatisfiable SAT instances [8]. CC ( Clause Count ) is an even simpler baseline method that just counts the number of clauses containing a given variable as a positive literal, and the number wherein it appears negatively. The relative weights of these two counts determine the variable s estimated bias distribution. 4 Summary of Experiments The first phase of experiments compared the eight methods as stand-alone bias estimators for randomly generated problems whose true biases were found via exhaustive model counting. Figure 1 depicts basic root-mean-squared error, with the global EMbased methods performing the best, followed by the controls and the local EM-based methods. Probabilistic methods that do not use EM have the worst basic accuracy, and in each of these bands the SP variants do better than those based on BP s simpler model. 0.48 0.46 0.44 0.42 Root Mean Squared Error 0.4 0.38 0.36 0.34 0.32 0.3 BP EMBPL EMBPG SP EMSPL EMSPG LC CC 0.28 0 20 40 60 80 100 Backbone Size Fig. 1. RMS error over 500 instances of increasing backbone size, n = 100. However, a second pattern arises across experiments that emphasize strong biases and estimates: backbone identification rate and rank of first wrong bias (graphs available in longer presentation [1]). The former measures the proportion of actual backbone variables that an estimator biases toward the correct polarity. The latter examines the variables to which a method assigns the strongest biases, and finds the highestranking estimate that was actually toward the wrong polarity. For both of these mea-

4 Probabilistically Estimating Backbones and Variable Bias: Experimental Overview sures, BP actually performs the best, followed by EMSP-G, LC, SP, EMBP-G, the local EM methods, and CC. These two sets of measures (basic accuracy and accuracy on strong biases) turned out to be the most important predictors of success in the second phase of study. Here, the methods were embedded as variable-ordering heuristics within the MiniSat solver [10]. Part of this process meant determining a good way to use surveys: based on our experimental experiences, the results reflect the intuitive succeed-first approach of setting the most strongly-biased variable in a survey in the direction of its stronger bias. The overall runtime results in Figure 2 exhibit a correlation between good bias estimates and significant improvement in search efficiency. The graph consists average runtimes broken down to show the proportion that was devoted to computing surveys. Ordering Heuristic [Average # Surveys / Run] EMSPG [165.4] EMBPG [78.9] LC [132.8] EMSPL [23.0] EMBPL [12.8] CC [75.1] DEFAULT [0.0] SP [13.0] Computing Surveys All Other Activities 0 0.2 0.4 0.6 0.8 1 1.2 Average Runtime for (seconds) Fig. 2. Total/Survey runtimes averaged over 100 random problems, n = 250 and α = 4.11. The DEFAULT method represents regular MiniSat without bias estimation, and BP performed so badly as to not appear on the graph. When scaling to larger experiments, the relative efficiency of the best method (EMSP-G) exhibits exponential growth. For instance, when n = 450, it requires an average of 5 minutes per problem, while default MiniSat typically requires half an hour, or times out after three hours. Of the many measures examined during the first phase of experiments, basic accuracy seems to be the most important predictor of success as a heuristic: global EM methods and controls outperform local EM methods and SP/BP, and within each of these bands the SP version is a more effective heuristic than the BP version. Exceptions to this correlation indicate the importance of the secondary strength-oriented accuracy measures mentioned above. For instance, CC compares relatively well in basic accuracy, but ranks lower as a heuristic it scored the worst with the accuracy measures that focus on strong biases. Still, the influence of such secondary measures remains dominated by overall accuracy BP is the best estimator of strong biases, but has poor overall accuracy and constitutes the worst heuristic by far.

Probabilistically Estimating Backbones and Variable Bias: Experimental Overview 5 5 Conclusions The main findings of these experiments indicate that probabilistic message-passing techniques can be comparatively successful at estimating variable bias and identifying backbone variables, and that successful bias estimation has a positive effect on heuristic search efficiency within a modern solver. Secondary contributions include a novel family of EM-based bias estimators, and a series of design insights culminating in a fast solver for hard random problems. However, many important issues remain. For instance, structured and unsatisfiable instances have not yet been considered by this bias estimation framework. This may require a finer-grained analysis of accuracy that considers variance across multiple runs with various random seeds. Further, the best way to use surveys for variable ordering cannot be settled conclusively by the limited span of branching strategies that have been studied to date. For instance, we waste some of the SP framework s power when we split the probability mass for the joker state between positive and negative bias; future branching strategies might favor variables with low joker probabilities. References 1. Hsu, E., Muise, C., Beck, J.C., McIlraith, S.: Applying probabilistic inference to heuristic search by estimating variable bias. Technical Report CSRG-577, University of Toronto (2008) 2. Braunstein, A., Mezard, M., Zecchina, R.: Survey propagation: An algorithm for satisfiability. Random Structures and Algorithms 27 (2005) 201 226 3. Dechter, R., Kask, K., Mateescu, R.: Iterative join-graph propagation. In: Proc. of 18th Int l Conference on Uncertainty in A.I. (UAI 02), Edmonton, Canada. (2002) 128 136 4. Hsu, E., Kitching, M., Bacchus, F., McIlraith, S.: Using EM to find likely assignments for solving CSP s. In: Proc. of 22nd National Conference on Artificial Intelligence (AAAI 07), Vancouver, Canada. (2007) 5. Monasson, R., Zecchina, R., Kirkpatrick, S., Selman, B., Troyansky, L.: Determining computational complexity from characteristic phase transitions. Nature 400(7) (1999) 133 137 6. Kilby, P., Slaney, J., Thiébaux, S., Walsh, T.: Backbones and backdoors in satisfiability. In: Proc. of 20th National Conference on A.I. (AAAI 05), Pittsburgh, PA. (2005) 7. Singer, J., Gent, I., Smaill, A.: Backbone fragility and the local search cost peak. Journal of Artificial Intelligence Research 12 (2000) 235 270 8. Dubois, O., Dequen, G.: A backbone-search heuristic for efficient solving of hard 3-SAT formulae. In: Proc. of 17th International Joint Conference on Artificial Intelligence (IJCAI 01), Seattle, WA. (2001) 9. Zhang, W.: Configuration landscape analysis and backbone guided local search. Part I: Satisfiability and maximum satisfiability. Artificial Intelligence 158(1) (2004) 1 26 10. Eén, N., Sörensson, N.: An extensible SAT-solver. In: Proc. of 6th International Conference on Theory and Applications of Satisfiability Testing (SAT 03), Portofino, Italy. (2003) 11. Pearl, J.: Probabilistic Reasoning in Intelligent Systems. Morgan Kaufmann, 1988. 12. Kschischang, F.R., Frey, B.J., Loeliger, H.A.: Factor graphs and the sum-product algorithm. IEEE Transactions on Information Theory 47(2) (2001) 13. Dempster, A., Laird, N., Rubin, D.: Maximum likelihood from incomplete data via the EM algorithm. Journal of the Royal Statistical Society 39(1) (1977) 1 39 14. Jordan, M., Ghahramani, Z., Jaakkola, T., Saul, L.: An introduction to variational methods for graphical models. In Jordan, M., ed.: Learning in Graphical Models. MIT Press (1998)