Agents that Reason Logically

Slides:



Advertisements
Similar presentations
Propositional and First-Order Logic
Advertisements

Methods of Proof Chapter 7, second half.. Proof methods Proof methods divide into (roughly) two kinds: Application of inference rules: Legitimate (sound)
Agents That Reason Logically Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 7 Spring 2004.
Propositional Logic CMSC 471 Chapter , 7.7 and Chuck Dyer
Logic CPSC 386 Artificial Intelligence Ellen Walker Hiram College.
3/30/00 Agents that Reason Logically by Chris Horn Jiansui Yang Xiaojing Wu.
Artificial Intelligence Knowledge-based Agents Russell and Norvig, Ch. 6, 7.
1 Knowledge-Based Agents Chapter CMSC 471 Adapted from slides by Tim Finin and Marie desJardins. Some material adopted from notes by Andreas Geyer-Schulz.
Logical Agents Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 7 Fall 2005.
Class Project Due at end of finals week Essentially anything you want, so long as its AI related and I approve Any programming language you want In pairs.
Logical Agents Chapter 7. Why Do We Need Logic? Problem-solving agents were very inflexible: hard code every possible state. Search is almost always exponential.
Computing & Information Sciences Kansas State University Lecture 11 of 42 CIS 530 / 730 Artificial Intelligence Lecture 11 of 42 William H. Hsu Department.
Logical Agents Chapter 7. Why Do We Need Logic? Problem-solving agents were very inflexible: hard code every possible state. Search is almost always exponential.
Knowledge in intelligent systems So far, we’ve used relatively specialized, naïve agents. How can we build agents that incorporate knowledge and a memory?
Methods of Proof Chapter 7, second half.
Knoweldge Representation & Reasoning
LOGICAL AGENTS Yılmaz KILIÇASLAN. Definitions Logical agents are those that can:  form representations of the world,  use a process of inference to.
Let remember from the previous lesson what is Knowledge representation
Logical Agents Chapter 7 Feb 26, Knowledge and Reasoning Knowledge of action outcome enables problem solving –a reflex agent can only find way from.
Logical Agents Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 7 Spring 2005.
Logical Agents Copyright, 1996 © Dale Carnegie & Associates, Inc. Chapter 7 Spring 2008.
Agents that Reason Logically Logical agents have knowledge base, from which they draw conclusions TELL: provide new facts to agent ASK: decide on appropriate.
1 Agents that Reason Logically Some material adopted from notes by Andreas Geyer-Schulz and Chuck Dyer Tim Finin Knowledge-based agents Propositional logic.
Logical Agents Chapter 7 (based on slides from Stuart Russell and Hwee Tou Ng)
Artificial Intelligence Lecture No. 9 Dr. Asad Ali Safi ​ Assistant Professor, Department of Computer Science, COMSATS Institute of Information Technology.
Knowledge-Based Agents Chapter Some material adopted from notes by Andreas Geyer-Schulz and Chuck Dyer.
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes Jan 19, 2012.
A knowledge-based agent
Pattern-directed inference systems
1 Logical Agents CS 171/271 (Chapter 7) Some text and images in these slides were drawn from Russel & Norvig’s published material.
Logical Agents Logic Propositional Logic Summary
1 CMSC 471 Fall 2002 Class #10/12–Wednesday, October 2 / Wednesday, October 9.
Class Project Due at end of finals week Essentially anything you want, so long as its AI related and I approve Any programming language you want In pairs.
An Introduction to Artificial Intelligence – CE Chapter 7- Logical Agents Ramin Halavati
1 CMSC 671 Fall 2001 Class #9 – Tuesday, October 2.
1 CMSC 471 Fall 2002 Class #9– Monday, Sept Knowledge Representation Chapter Some material adopted from notes by Andreas Geyer-Schulz and.
CS 4100 Artificial Intelligence Prof. C. Hafner Class Notes Jan 17, 2012.
Logical Agents Chapter 7. Knowledge bases Knowledge base (KB): set of sentences in a formal language Inference: deriving new sentences from the KB. E.g.:
1 Logical Agents CS 171/271 (Chapter 7) Some text and images in these slides were drawn from Russel & Norvig’s published material.
Knowledge Representation Lecture # 17, 18 & 19. Motivation (1) Up to now, we concentrated on search methods in worlds that can be relatively easily represented.
For Friday Read chapter 8 Homework: –Chapter 7, exercise 1.
Logical Agents Chapter 7. Outline Knowledge-based agents Logic in general Propositional (Boolean) logic Equivalence, validity, satisfiability.
1 Knowledge-Based Agents Chapter Some material adopted from notes by Andreas Geyer-Schulz and Chuck Dyer.
Knowledge-based agents: they keep track of the world by means of an internal state which is constantly updated by the agent itself A Knowledge-Based Agent.
1 The Wumpus Game StenchBreeze Stench Gold Breeze StenchBreeze Start  Breeze.
Logical Agents Chapter 7 Andreas Geyer-Schulz and Chuck Dyer
Dr. Shazzad Hosain Department of EECS North South Universtiy Lecture 04 – Part B Propositional Logic.
Knowledge-Based Agents Chapter Some material adopted from notes by Andreas Geyer-Schulz and Chuck Dyer.
Logical Agents Chapter 7. Outline Knowledge-based agents Propositional (Boolean) logic Equivalence, validity, satisfiability Inference rules and theorem.
1 UNIT-3 KNOWLEDGE REPRESENTATION. 2 Agents that reason logically(Logical agents) A Knowledge based Agent The Wumpus world environment Representation,
Propositional and First-Order Logic Chapter 7.4─7.8, 8.1─8.3, 8.5 Some material adopted from notes by Andreas Geyer-Schulz and Chuck Dyer.
Logical Agents Chapter 7 Part I. 2 Outline Knowledge-based agents Wumpus world Logic in general - models and entailment Propositional (Boolean) logic.
1 Knowledge Representation Logic and Inference Propositional Logic Vumpus World Knowledge Representation Logic and Inference Propositional Logic Vumpus.
Knowledge-Based Agents Chapter Some material adopted from notes by Andreas Geyer-Schulz and Chuck Dyer.
1 Intro to AI Logical Agents. 2 Intro to AI Outline Logic Efficient satisfiability testing by backtracking search Efficient satisfiability testing by.
Artificial Intelligence Logical Agents Chapter 7.
Artificial Intelligence Logical Agents Chapter 7.
Logical Agents. Outline Knowledge-based agents Logic in general - models and entailment Propositional (Boolean) logic Equivalence, validity, satisfiability.
Class #10– Wednesday, October 1
Knowledge and reasoning – second part
EA C461 – Artificial Intelligence Logical Agent
Learning and Knowledge Acquisition
Artificial Intelligence
Class #9– Thursday, September 29
Knowledge and reasoning – second part
Logical Agents Chapter 7.
Propositional Logic CMSC 471 Chapter , 7.7 and Chuck Dyer
CMSC 471 Fall 2011 Class #10 Tuesday, October 4 Knowledge-Based Agents
Logical Agents Chapter 7 Andreas Geyer-Schulz and Chuck Dyer
Logical Agents Prof. Dr. Widodo Budiharto 2018
Presentation transcript:

Agents that Reason Logically Knowledge-based agents Propositional logic Tim Finin Some material adopted from notes by Andreas Geyer-Schulz and Chuck Dyer

A knowledge-based agent A knowledge-based agent includes a knowledge base and an inference system. A knowledge base is a set of representations of facts of the world. Each individual representation is called a sentence. The sentences are expressed in a knowledge representation language. The agent operates as follows: 1. It TELLs the knowledge base what it perceives. 2. It ASKs the knowledge base what action it should perform. 3. It performs the chosen action. Examples of sentences The moon is made of green cheese If A is true then B is true A is false All humans are mortal Confucius is a human

Architecture of a knowledge-based agent Knowledge Level. The most abstract level: describe agent by saying what it knows. Example: A taxi agent might know that the Golden Gate Bridge connects San Francisco with the Marin County. Logical Level. The level at which the knowledge is encoded into sentences. Example: Links(GoldenGateBridge, SanFrancisco, MarinCounty). Implementation Level. The physical representation of the sentences in the logical level. Example: ‘(links goldengatebridge sanfrancisco marincounty)

The Inference engine derives new sentences from the input and KB The inference mechanism depends on representation in KB The agent operates as follows: 1. It receives percepts from environment 2. It computes what action it should perform (by IE and KB) 3. It performs the chosen action (some actions are simply inserting inferred new facts into KB). Input from environment Inference Engine Output (actions) Learning (KB update) Knowledge Base

The Wumpus World environment The Wumpus computer game The agent explores a cave consisting of rooms connected by passageways. Lurking somewhere in the cave is the Wumpus, a beast that eats any agent that enters its room. Some rooms contain bottomless pits that trap any agent that wanders into the room. Occasionally, there is a heap of gold in a room. The goal of the player is: to collect the gold and exit the world without being eaten

Jargon file on “Hunt the Wumpus” WUMPUS /wuhm'p*s/ n. The central monster (and, in many versions, the name) of a famous family of very early computer games called “Hunt The Wumpus,” dating back at least to 1972 The wumpus lived somewhere in a cave with the topology of a dodecahedron's edge/vertex graph (later versions supported other topologies, including an icosahedron and Mobius strip). The player started somewhere at random in the cave, or in a selected state like (1,1) He had five “crooked arrows”; these could be shot through up to three connected rooms, and would kill the wumpus on a hit (later versions introduced the wounded wumpus, which got very angry).

Jargon file on “Hunt the Wumpus” (cont) This slide has only historical information or info useful for extensions and can be skipped Unfortunately for players, the movement necessary to map the maze was made hazardous not merely by the wumpus (which would eat you if you stepped on him) There are also bottomless pits and colonies of super bats that would pick you up and drop you at a random location (later versions added: “anaerobic termites” that ate arrows, bat migrations, and earthquakes that randomly change pit locations). This game appears to have been the first to use a non-random graph-structured map (as opposed to a rectangular grid like the even older Star Trek games). In this respect, as in the dungeon-like setting and its terse, amusing messages, it prefigured ADVENT and Zork. It was directly ancestral to both. (Zork acknowledged this heritage by including a super-bat colony.) Today, a port is distributed with SunOS and as freeware for the Mac. A C emulation of the original Basic game is in circulation as freeware on the net.

A typical Wumpus world The agent always starts in the field [1,1]. The task of the agent is to find the gold, return to the field [1,1] and climb out of the cave.

Agent in a Wumpus world: Percepts The agent perceives a stench in the square containing the wumpus and in the adjacent squares (not diagonally) a breeze in the squares adjacent to a pit a glitter in the square where the gold is a bump, if it walks into a wall a woeful scream everywhere in the cave, if the wumpus is killed The percepts will be given as a five-symbol list: If there is a stench, and a breeze, but no glitter, no bump, and no scream, the percept is [Stench, Breeze, None, None, None] The agent can not perceive its own location.

The actions of the agent in Wumpus game are: go forward turn right 90 degrees turn left 90 degrees grab means pick up an object that is in the same square as the agent shoot means fire an arrow in a straight line in the direction the agent is looking. The arrow continues until it either hits and kills the wumpus or hits the wall. The agent has only one arrow. Only the first shot has any effect. climb is used to leave the cave. Only effective in start field. die, if the agent enters a square with a pit or a live wumpus. (No take-backs!)

The agent’s goal The agent’s goal is to find the gold and bring it back to the start as quickly as possible, without getting killed. 1000 points reward for climbing out of the cave with the gold 1 point deducted for every action taken 10000 points penalty for getting killed

The Wumpus agent’s first step

Later Agent takes a risk to go here but finds gold Now Agent has to safely withdraw Agent takes a risk to go here

World-wide web wumpuses http://scv.bu.edu/wcl http://216.246.19.186 http://www.cs.berkeley.edu/~russell/code/doc/overview-AGENTS.html

Representation, reasoning, and logic The object of knowledge representation is to express knowledge in a computer-tractable form, so that agents can perform well. A knowledge representation language is defined by: its syntax, which defines all possible sequences of symbols that constitute sentences of the language. Examples: Sentences in a book, bit patterns in computer memory. its semantics, which determines the facts in the world to which the sentences refer. Each sentence makes a claim about the world. An agent is said to believe a sentence about the world.

The connection between sentences and facts Entails Follows Semantics maps sentences in logic to facts in the world. The property of one fact following from another is mirrored by the property of one sentence being entailed by another.

Logic as a Knowledge-Representation (KR) language Propositional Logic First Order Higher Order Modal Fuzzy Logic Multi-valued Probabilistic Temporal Non-monotonic

Ontology and epistemology Ontology is the study of what there is, an inventory of what exists. An ontological commitment is a commitment to an existence claim. Epistemology is major branch of philosophy that concerns the forms, nature, and preconditions of knowledge.

Propositional logic  ...and  ...or ...implies ..is equivalent Logical constants: true, false Propositional symbols: P, Q, S, ... Wrapping parentheses: ( … ) Sentences are combined by connectives:  ...and  ...or ...implies ..is equivalent  ...not

Propositional logic (PL) A simple language useful for showing key ideas and definitions User defines a set of propositional symbols, like P and Q. User defines the semantics of each of these symbols, e.g.: P means "It is hot" Q means "It is humid" R means "It is raining" A sentence (aka formula, well-formed formula, wff) defined as: A symbol If S is a sentence, then ~S is a sentence (e.g., "not”) If S is a sentence, then so is (S) If S and T are sentences, then (S v T), (S ^ T), (S => T) , and (S <=> T) are sentences (e.g., "or," "and," "implies," and "if and only if”) A finite number of applications of the above

Examples of PL sentences (P ^ Q) => R “If it is hot and humid, then it is raining” Q => P “If it is humid, then it is hot” Q “It is humid.” A better way: Ho = “It is hot” Hu = “It is humid” R = “It is raining”

A BNF grammar of sentences in propositional logic S := <Sentence> ; <Sentence> := <AtomicSentence> | <ComplexSentence> ; <AtomicSentence> := "TRUE" | "FALSE" | "P" | "Q" | "S" ; <ComplexSentence> := "(" <Sentence> ")" | <Sentence> <Connective> <Sentence> | "NOT" <Sentence> ; <Connective> := "NOT" | "AND" | "OR" | "IMPLIES" | "EQUIVALENT" ;

Terms: semantics, interpretation, model, tautology, contradiction, entailment The meaning or semantics of a sentence determines its interpretation. Given the truth values of all of symbols in a sentence, it can be “evaluated” to determine its truth value (True or False). A model for a KB is a “possible world” in which each sentence in the KB is True. A valid sentence or tautology is a sentence that is True under all interpretations, no matter what the world is actually like or what the semantics is. Example: “It’s raining or it’s not raining.”

Terms: semantics, interpretation, model, tautology, contradiction, entailment An inconsistent sentence or contradiction is a sentence that is False under all interpretations. The world is never like what it describes, as in “It’s raining and it's not raining.” P entails Q, written P |= Q, means that whenever P is True, so is Q. In other words, all models of P are also models of Q. Sentence, relation P |= Q Entails On(a,b) a b Follows Fact of the world

Truth tables

Truth tables II The five logical connectives: A complex sentence:

Models of complex sentences

Agents have no independent access to the world The reasoning agent often gets its knowledge about the facts of the world as a sequence of logical sentences. It must draw conclusions only from them , without independent access to the world. Thus it is very important that the agent’s reasoning is sound! reasoning agent Remember symbol hypothesis and Brooks?

Inference rules Logical inference is used to create new sentences that logically follow from a given set of predicate calculus sentences (KB). An inference rule is sound if every sentence X produced by an inference rule operating on a KB logically follows from the KB. (That is, the inference rule does not create any contradictions) An inference rule is complete if it is able to produce every expression that logically follows from the KB. We also say - “expression is entailed by KB”. Please note the analogy to complete search algorithms.

Sound rules of inference Here are some examples of sound rules of inference. Each can be shown to be sound using a truth table: A rule is sound if its conclusion is true whenever the premise is true. RULE PREMISE CONCLUSION Modus Ponens A, A => B B And Introduction A, B A ^ B And Elimination A ^ B A Double Negation ~~A A Unit Resolution A v B, ~B A Resolution A v B, ~B v C A v C

Sound Inference Rules (deductive rules) Here are some examples of sound rules of inference. Each can be shown to be sound using a truth table -- a rule is sound if it’s conclusion is true whenever the premise is true. RULE PREMISE CONCLUSION Modus Tollens ~B, A => B ~A Or Introduction A A v B Chaining A => B, B => C A => C

Soundness of modus ponens premise premise conclusion A B A → B OK? True  False

Soundness of the resolution inference rule (~B + C) (A + B) Whenever premise is true, the conclusion is also true (A + C) But it may be also in other cases

Resolution rule Unit Resolution A v B, ~A B Resolution A v B, ~B v C A v C Operates on two disjunctions of literals The pair of two opposite literals cancel each other, all other literals from the two disjuncts are combined to form a new disjunct as the inferred sentence Resolution rule can replace all other inference rules Modus Ponens A, ~A v B B Modus Tollens ~B, ~A v B ~A Chaining ~A v B, ~B v C ~A v C These old rules are special cases of resolution

Proving things: what are proofs and theorems? A proof is a sequence of sentences, where each sentence is either a premise or a sentence derived from earlier sentences in the proof by one of the rules of inference. The last sentence is the theorem (also called goal or query) that we want to prove. Example for the “weather problem” given above. 1 Hu Premise “It is humid” 2 Hu=>Ho Premise “If it is humid, it is hot” 3 Ho Modus Ponens(1,2) “It is hot” 4 (Ho^Hu)=>R Premise “If it’s hot & humid, it’s raining” 5 Ho^Hu And Introduction(1,2) “It is hot and humid” 6 R Modus Ponens(4,5) “It is raining”

Theorem proving as search Proof by resolution Q ~Q v P ~P v ~Q v R premises P ~Q v R R theorem Theorem proving as search Start node: the set of given premises/axioms (KB + Input) Operator: inference rule (add a new sentence into parent node) Goal: a state that contains the theorem asked to prove Solution: a path from start node to a goal

Normal forms of PL sentences Disjunctive normal form (DNF) Any sentence can be written as a disjunction of conjunctions of literals. Examples: P ^ Q ^ ~R; A^B v C^D v P^Q^R; P Widely used in logical circuit design (simplification) Conjunctive normal form (CNF) Any sentence can be written as a conjunction of disjunctions of literals. Examples: P v Q v ~R; (A v B) ^ (C v D) ^ (P v Q v R); P Normal forms can be obtained by applying equivalence laws [(A v B) => (C v D)] => P  ~[~(A v B) v (C v D)] v P // law for implication  [~~(A v B) ^ ~(C v D)] v P // de Morgan’s law  [(A v B)^(~C ^ ~D)] v P // double negation and de Morgan’s law  (A v B v P)^(~C^~D v P) // distribution law  (A v B v P)^(~C v P)^(~D v P) // a CNF

Horn sentences A Horn sentence or Horn clause has the form: P1 ^ P2 ^ P3 ... ^ Pn => Q or alternatively ~P1 v ~P2 v ~P3 ... V ~Pn v Q where Ps and Q are non-negated atoms To get a proof for Horn sentences, apply Modus Ponens repeatedly until nothing can be done We will use the Horn clause form later (P => Q) = (~P v Q)

Entailment and derivation Entailment: KB |= Q Q is entailed by KB (a set of premises or assumptions) if and only if there is no logically possible world in which Q is false while all the premises in KB are true. Or, stated positively, Q is entailed by KB if and only if the conclusion is true in every logically possible world in which all the premises in KB are true. Derivation: KB |- Q We can derive Q from KB if there is a proof consisting of a sequence of valid inference steps starting from the premises in KB and resulting in Q Entailment is about what is true Derivation is about what can be proved

Two important properties for inference entailment Soundness: If KB |- Q then KB |= Q If Q is derived from a set of sentences KB using a given set of rules of inference, then Q is entailed by KB. Hence, inference produces only real entailments, or any sentence that follows deductively from the premises is valid. Completeness: If KB |= Q then KB |- Q If Q is entailed by a set of sentences KB, then Q can be derived from KB using the rules of inference. Hence, inference produces all entailments, or all valid sentences can be proved from the premises. completeness What can be derived What is true What can be derived What is true soundness

Why propositional logic is a weak language? 1. Because it is hard to identify "individuals." E.g., Mary, 3 Individuals cannot be PL sentences themselves. 2. Because we can not directly talk about properties of individuals or relations between individuals. (hard to connect individuals to class properties). E.g., property of being a human implies property of being mortal E.g. “Bill is tall” 3. Because generalizations, patterns, regularities can not be easily represented. E.g., all triangles have 3 sides All members of a class have this property Some members of a class have this property A better representation is needed to capture the relationship (and distinction) between objects and classes, including properties belonging to classes and individuals.

Confusius Example: weakness of PL Consider the problem of representing the following information: Every person is mortal. Confucius is a person. Confucius is mortal. How can these sentences be represented so that we can infer the third sentence from the first two?

PL is Too Weak a Representational Language Consider the problem of representing the following information: Every person is mortal. (S1) Confucius is a person. (S2) Confucius is mortal. (S3) S3 is clearly a logical consequence of S1 and S2. But how can these sentences be represented using PL so that we can infer the third sentence from the first two? We can use symbols P, Q, and R to denote the three propositions, but this leads us to nowhere because knowledge important to infer R from P and Q (i.e., relationship between being a human and mortality, and the membership relation between Confucius and human class) is not expressed in a way that can be used by inference rules

Example continued Every person is mortal. Confucius is a person. In PL we have to create propositional symbols to stand for all or part of each sentence. For example, we might do: P = “person”; Q = “mortal”; R = “Confucius” so the above 3 sentences are represented as: P => Q; R => P; R => Q Although the third sentence is entailed by the first two, we needed an explicit symbol, R, to represent an individual, Confucius, who is a member of the classes “person” and “mortal.” To represent other individuals we must introduce separate symbols for each one. This means, for representing the fact that all individuals who are “people” are also "mortal.” P => Q Every person is mortal. Confucius is a person. Confucius is mortal.

Bad semantics “Confucius” (and “person” and “mortal”) are not PL sentences (not a declarative statement) and cannot have a truth value. What does P => M mean? We need infinite distinct symbols X for individual persons, and infinite implications to connect these X with P (person) and M (mortal) because we need a unique symbol for each individual. Person_1 => P; person_1 => M; Person_2 => P; person_2 => M; ... ... Person_n => P; person_n => M Requiring unique symbol for each individual is bad, no generalization possible!

…. therefore first order logic was invented……. First-Order Logic (abbreviated FOL or FOPC) is expressive enough to concisely represent this kind of situation by separating classes and individuals 1. Explicit representation of individuals and classes, x, Mary, 3, persons. 2. Adds: relations, variables, and quantifiers, e.g., “Every person is mortal” Forall X: person(X) => mortal(X) “There is a white alligator” There exists some X: Alligator(X) ^ white(X)

The “Hunt the Wumpus” agent Some Atomic Propositions S12 = There is a stench in cell (1,2) B34 = There is a breeze in cell (3,4) W22 = The Wumpus is in cell (2,2) V11 = We have visited cell (1,1) OK11 = Cell (1,1) is safe. etc Some rules (R1) ~S11 => ~W11 ^ ~W12 ^ ~W21 (R2) ~S21 => ~W11 ^ ~W21 ^ ~W22 ^ ~W31 (R3) ~S12 => ~W11 ^ ~W12 ^ ~W22 ^ ~W13 (R4) S12 => W13 v W12 v W22 v W11 Note that the lack of variables requires us to give similar rules for each cell.

After the third move We can prove that the Wumpus is in (1,3) using the four rules given. See book of Russell and Norvig (R&N) section 6.5

Proving W13 ~S11 (R1) ~S11 => ~W11 ^ ~W12 ^ ~W21 Apply MP with ~S11 and R1: ~W11 ^ ~W12 ^ ~W21 Apply And-Elimination to this we get 3 sentences: ~W11, ~W12, ~W21 Apply MP to ~S21 and R2, then applying And-elimination: ~W22, ~W21, ~W31 Apply MP to S12 and R4 we obtain: W13 v W12 v W22 v W11 Apply Unit resolution on (W13 v W12 v W22 v W11) and ~W11 W13 v W12 v W22 Apply Unit Resolution with (W13 v W12 v W22) and ~W22 W13 v W12 Apply UR with (W13 v W12) and ~W12 W13 QED ~W11 ^ ~W12 ^ ~W21 ~S21 R2 ~W22, ~W21, ~W31 S12 R4 (W13 v W12 v W22 v W11) ~W11 (W13 v W12 v W22) ~W22 ~W12 We can prove that the Wumpus is in (1,3) W13 v W12 W13

Problems with the propositional Wumpus hunter Lack of variables prevents stating more general rules. E.g., we need a set of similar rules for each cell Change of the KB over time is difficult to represent Standard technique is to index facts with the time when they’re true This means we have a separate KB for every time point. We need to add predicates and time

Summary Intelligent agents need knowledge about the world for making good decisions. The knowledge of an agent is stored in a knowledge base in the form of sentences in a knowledge representation language. A knowledge-based agent needs a knowledge base and an inference mechanism. It operates by storing sentences in its knowledge base, inferring new sentences with the inference mechanism, and using them to deduce which actions to take. A representation language is defined by its syntax and semantics, which specify the structure of sentences and how they relate to the facts of the world. The interpretation of a sentence is the fact to which it refers. If this fact is part of the actual world, then the sentence is true.

Summary II The process of deriving new sentences from old one is called inference. Sound inference processes derives true conclusions given true premises. Complete inference processes derive all true conclusions from a set of premises. A valid sentence is true in all worlds under all interpretations. If an implication sentence can be shown to be valid, then - given its premise - its consequent can be derived. Different logics make different commitments about what the world is made of and what kind of beliefs we can have regarding the facts. Logics are useful for the commitments they do not make because lack of commitment gives the knowledge base write more freedom. Propositional logic commits only to the existence of facts that may or may not be the case in the world being represented. It has a simple syntax and a simple semantic. It suffices to illustrate the process of inference. Propositional logic quickly becomes impractical, even for very small worlds.