intro to ai resolution & prolog ruth bergman fall 2002

50
Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Post on 19-Dec-2015

221 views

Category:

Documents


1 download

TRANSCRIPT

Page 1: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Intro to AIResolution & Prolog

Ruth Bergman

Fall 2002

Page 2: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Inference Rules for FOL

• Inference Rules – Modus Ponens, And-Elimination, And-Introduction,

Or-Introduction, Resolution

• Rules for substituting variables– Universal-Elimination, Existential-Elimination,

Existential-Introduction

Page 3: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Theorem Proving Strategies

• Strategy 1: Deduce everything possible from the KB, and hope to derive what you want to prove.– A forward chaining approach

• Strategy 2: Find out what facts you need in order to prove theorem. Try to prove those facts. Repeat until there are no facts to prove.– A backward chaining approach

Page 4: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Forward Chaining

• A “data-driven” strategy• This inference procedure will find a proof for any true sentence• For false sentences this procedure may not halt. This

procedure is semi-decidable.• Generates many irrelevant conclusions.

Forward-chain(KB, 1. If is in KB return yes2. Apply inferences rules on KB3. Add new sentences to KB4. Go to 2

Page 5: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Backward Chaining

• Applied when there is a goal to prove• Apply modus-ponens backwards• Modus Ponens is incomplete

x P(x) Q(x), xP(x) R(x)x Q(x) S(x), x R(x) S(x)– cannot conclude S(A)

back-chain(KB, list

1. If list is empty return yes2. Let First(list)3. for each sentence(p1… pn) in KB

back-chain(KB,[p1, …, pn, First(list)])

Simplified version ignores variable substitution.

Page 6: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Resolution : A Complete Inference Procedure

• Simple resolution (for ground terms):

• Modus ponens derives only atomic conclusions, but resolution derives new implications

, ,

Page 7: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

The Resolution Inference Rule

• Generalized Resolution :– for literals pi and qi, where UNIFY(pj , qk) =

p1 … pj … pm ,q1 … qk … qn

SUBST(, (p1 … pj-1 pj+1 … pm q1 … qk-1 qk+1 … qn

Page 8: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Resolution v.s. Modus Ponens

• Resolution is a generalization of modus ponens

Modus Ponens

,

Resolution

True ,

True

Page 9: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Resolution Proofs

P(w) v Q(w)

R(z) v S(z)

P(w) v S(w) P(x) R(x)

Q(y) v S(y)

S(A)

{y/w}

{x/w}

{w/A, z/A}

S(w) R(w) Premise

Conclusion, resolvent

Still not complete

w P(w) v Q(w)

x P(x) v R(x)

y Q(y) v S(y)

x R(x) v S(x)

Page 10: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Resolution Proof with Refutation

• Complete inference procedure• refutation, proof by contradiction, reductio ad

absurdum• To proof P,

– assume P is false (add P to KB)– prove a contradiction

• (KB P False) (KB P)

Page 11: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Resolution Principle• The resolution procedure:

– To prove p, add P to KB– convert all sentence in KB to cannonical form– use the resolution rule (repeatedly)– try to derive an empty clause

Page 12: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Proof by Refutation

P(w) v Q(w)

R(z) v S(z)

P(w) v S(w) P(x) R(x)

Q(y) v S(y)

S(A)

{y/w}

{x/w}

{w/A, z/A}

S(w) R(w)

w P(w) v Q(w)

x P(x) v R(x)

y Q(y) v S(y)

x R(x) v S(x)

S(A)

S(A)

contradiction

Page 13: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Refutation

P(w) Q(w)

R(z) S(z)

P(w) S(w) True P(x) R(x)

Q(y) S(y)

True S(x)

{y/w}

{w/x}

{z/x}

True S(x) R(x)

S(A) False

True False

{x/A}

Contradiction

Page 14: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Canonical Form for Resolution

• Conjunctive Normal Form (CNF) : conjunctions of disjunctions

• Implied universal quantifier on variables• Remove existential quantifiers

• Quantifiers can be tricky: y x p(x,y) x p(x,A) where A is a new symbol x y p(x,y) ???

• intuitively, x p(x,A) is wrong• we want to capture the idea that the existential quantifier is

somehow dependent on the universal scoped outside of it

CNF P(w) Q(w) P(x) R(x)Q(y) S(y)R(z) S(z)

Page 15: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Conversion to Normal Form

• Any first-order logic sentence can be converted to CNF

• Conversion procedure1. Eliminate implication

2. Reduce the scope of negations () to single predicates

3. Standardize variable

4. Move quantifiers left

5. Skolemize : remove existential quantifiers

6. Distribute over 7. Remove to form separate wff

Page 16: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Conversion to CNF

Example: Convert ( y x p(x,y) x y p(x,y)) to CNF

1. Eliminate implication: Replace A B by A B ( y x p(x,y) v x y p(x,y))

2. Move inwards : negations are only allowed on atoms

x y p(x,y) ^ x y p(x,y)

3. Standardize variables : unique variable name

y x p(x,y) ^ w z p(w,z)

4. Move quantifier left without changing meaning

y x w z p(x,y) ^ p(w,z)

Page 17: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Conversion to CNF: Skolemization

5. Skolemization (named after the Polish logician Skolem)– replace each existentially quantified variable with a new

function () with arguments that are any universally quantified variable scoped outside of it

y x p(x,y) x p(x,sk1) x y p(x,y) x p(x,sk2(x)) x y z p(x,y) ^ q(x,z) x p(x,sk3(x)) ^ q(x,sk4(x)) x y z p(x,y) ^ q(x,z) x y p(x,y) ^ q(x,sk5(x,y))– These function are often referred to as Skolem functions

– We can now remove as implicit. All remaining variables are universally quantified.

p(x,sk1) ^ p(sk2(x),z)

Page 18: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Conversion to CNF

6. Distribute over – (a b) c

7. Remove – p(x,sk1), p(sk2(x),z)

• As a result, we have sets of clauses to which we can apply the resolution rule

– find a set with a positive term that matches a negative term in another set

– to do so, we need to understand how to match things

Page 19: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Unification• The process of matching is called unification:

– p(x) matches p(Jack) with x/Jack– q(fatherof(x),y) matches q(y,z) with y/fatherof(x) and z /y

• note the result of the match is q(fatherof(x),fatherof(x))– p(x) matches p(y) with

• x = Jack and y/Jack• x = John and y/John• or x = y

– The match that makes the least commitment is called the most general unifier (MGU)

Page 20: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Substitution

• We use the notation subst(t,s) to denote the application of the substitution s = {v1/t1,v2/t2 ... vn/tn} to t.

• Composition of substitutions

– subst(COMPOSE(1, 2), p) = subst(2, SUBST(1, p))

Page 21: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Unification Algorithm• Assume input formulas q and q have standardize variables

parentof(fatherof(John),y) parentof(y,z) parentof(fatherof(John),v1) parentof(v2,v3)

• Unification takes two atomic sentences, p and q, and returns a substitution that would make p and q look the same.

Page 22: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Unification Algorithm• Unify(p, q, ) returns a substitution or fail

– If p = q return { }– If either p or q is a constant

if p = q return { } else return fail– If p is a variable return Unify-var(p, q, )– If q is a variable return Unify-var(q, p, )– If the predicate symbols of p and q are different return fail– If p and q have different arity return fail– for each pair of arguments t1, t2 of p, q

• Let S = Unify (t1,t2, )• If S = fail return fail• Apply S to the remaining arguments of p and q• Set = COMPOSE(S,

– Return

Page 23: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Unification Algorithm• Unify-var(var, x, ) returns a substitution or fail

– If {var/val} in return Unify(val, x, – If {x/val} in return Unify(var, val, – If var occurs anywhere in x return fail– Return {x/var} U

• note how this works with skolemization• The occurs check takes time linear in the size of the expression.

Thus time complexity of Unify is O(n2)

Page 24: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Resolution Strategies

• We know that repeated application of the resolution rule will find a proof if one exists

• How efficient is this process?– The resolution procedure may be viewed as a

search problem– The state space is very large

• Some strategies for guiding the search toward a proof …– A resolution strategy is complete if its use will result in a

procedure that will find a contradiction whenever one exists.

Page 25: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Horn Clauses: A Special Case

• Horn clauses are a special subset of logic:– A x P(x)– A x,y P(x) Q(x)^ R(y)– All universal variables– only positive atomic sentences

P(x) v Q(x) v R(x)

• Note that pure backward chaining suffices to prove statements:– To prove P(Y), either

• unify with atomic P(x)• unify with P(x) v Q(x) v R(x); Q(x) v R(x) become

subgoals

Page 26: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Unit Resolution

• Aka Unit preference– prefers to do resolution where one of the

sentence is a single literal (unit clause)– we are trying to prove a sentence of length

0, so it might be a good idea to prefer inferences that produce shorter sentences

– it is a useful heuristic that can be combined with other strategies

– Complete for Horn clauses.

Page 27: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Input Resolution

– every resolution combines one of the input sentences (either query or sentence in KB)

– shape of a diagonal “spine” – input resolution is equivalent to unit resolution– complete for Horn form but incomplete in general

case• Example: Q(u) v P(A) Q(w) v P(w) Q(x) v P(x)Q(y) v P(y)

Page 28: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

An Input Resolution Proof

P(w) v Q(w)

R(z) v S(z)

P(w) v S(w) P(x) R(x)

Q(y) v S(y)

S(A)

{y/w}

{x/w}

{w/A, z/A}

S(w) R(w)

w P(w) v Q(w)

x P(x) v R(x)

y Q(y) v S(y)

x R(x) v S(x)

S(A)

S(A)

contradiction

Page 29: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Linear Resolution

• a generalization of input resolution– allow P and Q to be resolved together if P is an input

sentence or P is an ancestor of Q in the proof tree– linear resolution is complete

Page 30: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

A Linear Resolution Proof

Q(x) v P(x)

Q(u) v P(A)

P(x) Q(w) v P(w)

Q(y) v P(y)

P(A)

Q(w)

Q(u) v P(A)Q(w) v P(w) Q(x) v P(x)Q(y) v P(y)

contradiction

Page 31: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Set of support

• This strategy tries to eliminate many resolutions altogether by identifying a subset of sentences called the set of support– Every resolution combines a sentence from set of support

with another sentence– add resolvent into set of support– if set of support is small relative to KB, reduce search space

significantly– A bad choice for the set of support will make the algorithm

incomplete• If the remainder of the sentences are jointly satisfiable

– Use the negated query as the set of support• goal-directed : easy to understand• Assumes the KB is consistent

Page 32: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

An Set-of-Support Resolution Proof

w P(w) v Q(w)

x P(x) v R(x)

y Q(y) v S(y)

x R(x) v S(x)

S(A)

Q(A)

P(A)

R(A)

P(A)

contradiction

Page 33: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Simplification Strategies

• Elimination of Tautologies– A clause containing P(x) v P(x) may be removed

• Subsumption– eliminate all sentences that are subsumed by an existing

sentence in KB– remove P(A), P(A) Q(B), P(y) v Q(z) if P(x) is in KB– subsumption keeps the KB smaller thus reducing the

branching factor

Page 34: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Resolution Examples x y p(x,y) y x p(x,y) y x p(x,y) x y p(x,y)• From our previous axiomitization

x male(x) v female(x) ^ (male(x) ^ female(x)) x y z parentof(x,y) ^ ancesterof(y,z) ancesterof(x,z) x y parentof(x,y) ancesterof(x,y) x parentof(fatherof(x),x) ^ parentof(motherof(x),x) ^ male(fatherof(x)) ^

female(motherof(x)) x y1, y2 parentof(y1,x) ^ parentof(y2,x) ^ (y1 = y2) x,y childof(y,x) parentof(x,y)– male(john), female(annika), parentof(annika,john) .....

• show x y ancesterof(y,x) ^ female(y)– childof(john, annika) childof(annika,john)

Page 35: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Example

• Relevant clausal forms from KB– {parentof(x,y), ancesterof(x,y)}– {parentof(motherof(x),x)}– {female(motherof(x))}

• negated goal x y ancesterof(y,x) ^ female(y)– {ancesterof(y,A),female(y)}

• Proof .... (done in class)

Page 36: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

West Wessex Marathon Race

I. A racer medals iff he receives a gold, silver or bronze medalx medal(x) (gold(x) v silver(x) v bronze(x))

II. A racer can only receive one medal

. x gold(x) medal(x) ^ (silver(x) v bronze(x))

. x silver(x) medal(x) ^ (gold(x) v bronze(x))

c. x bronze(x) medal(x) ^ (silver(x) v gold(x))III. Only one race can receive a gold (silver, bronze) medal

. x,y gold(x) ^ x != y gold(y)

. x,y silver(x) ^ x != y silver(y)

c. x,y bronze(x) ^ x != y bronze(y)IV. Some participant must win a gold (silver, bronze) medal

. x gold(x). x silver(x)c. x bronze(x)

Page 37: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

West Wessex Marathon Race

1. A will not win the gold, nor B the silver is falsegold(A) v silver(B)

2. C will win a medal, and D will not is false

medal(C) v medal(D)3. D and E will both win medals is false

medal(D) v medal(E)4. D will not win the silver, nor E the bronze is false

silver(D) v bronze(E)5. A will win a medal, and C will not is false

medal(A) v medal(C)• Who won which of the medals? Proof in class.

Page 38: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Prolog: a Logic Programming Language

• A program is a sequence of sentences• Sentences are represented as Horn clauses,

no negated antecedents• The Prolog representation has the

consequent, or head, on the left hand side, and the antecedents, or body, on the right.

• A query is a conjunction of terms; it is called the goal

Page 39: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Prolog: Example

• fatherof(x,y) :- male(x), parentof(x,y).• motherof(x,y) :- female(x), parentof(x,y).• ancesterof(x,y) :- parentof(x,y).• ancesterof(x,y) :- parentof(x,z), ancesterof(z,y).• childof(x,y) :- parentof(y,x).• sonof(x,y) :- male(x), childof(x,y).• daughterof(x,y) :- female(x), childof(x,y).• male(john), male(joe),female(mary), female(jane)• parentof(john,mary), parentof(mary,jane),

parentof(joe, john)

Page 40: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

How does it work

• male(john), male(joe),female(mary), female(jane)• parentof(john,mary), parentof(mary,jane), parentof(joe, john)

• How would ?- ancesterof(x,y) be answered?– ancesterof(john,mary)– ancesterof(mary,jane)– ancesterof(joe,john)– ancesterof(john,jane) – .– .– .

Page 41: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

How Does it Work?

• Queries are answered by the following algorithm:– The first clause with a head that matches is chosen

• if it has subgoals, they become new queries, answered in left to right order

• when all goals are satisfied, an answer is returned• if another answer is requested, the search continues where

it left off– When no more matches are possible with chosen clause, next

clause is chosen– This is effectively a depth-first search of the solution tree

• note that we need to avoid left recursion!!

Page 42: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Some Real Prolog examples

• ...... running code in prolog.lisp ....

Page 43: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

The Difference Between Logic and Prolog

• Handling of negation:– Prolog handles negation as failure to prove, e.g.

• not(fatherof(bill,mike)) will return true in Prolog, but not provable in formal logic

• To save time, Prolog often runs without the occurs check for speed

• Prolog programs can side-effect as they run, and possibly do computations (e.g. arithmetic)

• Clause ordering affects what can be shown; ordering doesn’t matter if logic

Page 44: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Rules in Production Systems

• A rule based system– A rule has the format

if condition then action

“if the condition is met in the world then take the action” - the condition states when the rule applies- the action describes the result of applying the rule

Page 45: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Production System

A production system consists of• A knowledge base

– Working memory : short-term memory• contains set of positive literals with no variable

– Rule memory : long-term memory• contains set of inference rules

• A control strategy• A rule interpreter

Page 46: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Forward-Chaining Production Systems

The forward-chaining production system loop:

Repeat• Match phase

– computes subset of rules whose left-hand side is satisfied by current content of working memory

• Conflict resolution phase– decide which rule(s) should be executed

• Act phase– modifies the knowledge base, usually working memory

Page 47: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Match phase

• A unification exercise– c cycles, w elements in WM, r rules with n

elements in LHS c*w*r*n unifications

• The RETE algorithm – used in the OPS-5 production system– Construct a RETE network using the working

memory and the rule base– eliminate duplication btn. Rules– eliminate duplication over time

Page 48: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Conflict Resolution Phase

• More than one rules pass the match phase• Do every actions or select one (some)• Control strategy

– No duplication– Recency– Specificity

• Mammal(x) add Legs(x,4)• Mammal(x) Human(x) add Legs(x,2)

– Operation priority : Action(Dust(p)) v.s. Action(Evacuate)

More on production systems when we discuss uncertain reasoning

Page 49: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Frame Systems & Semantic Networks

Semantic networks has– Graphical notation

easy for people to visualize

– very simple execution modeleasy to predict the behavior of the inference engine

simple query language

Whether the language uses strings or nodes and links, and whether it is called a semantic network of a logic, has no effect on its meaning or an its implementation

Page 50: Intro to AI Resolution & Prolog Ruth Bergman Fall 2002

Some Uses of Logic

• Otter theorem prover– current system 4th generation, dating back

to the 60’s– first order logic plus equality– used as a proving assistant

• To see what’s been done, go to the otter home page!