Ontologies and Knowledge-based Systems Is there a flexible way to - - PowerPoint PPT Presentation

ontologies and knowledge based systems
SMART_READER_LITE
LIVE PREVIEW

Ontologies and Knowledge-based Systems Is there a flexible way to - - PowerPoint PPT Presentation

Ontologies and Knowledge-based Systems Is there a flexible way to represent relations? How can knowledge bases be made to interoperate semantically? D. Poole and A. Mackworth 2017 c Artificial Intelligence, Lecture 14.1, Page 1 1 / 12


slide-1
SLIDE 1

Ontologies and Knowledge-based Systems

Is there a flexible way to represent relations? How can knowledge bases be made to interoperate semantically?

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.1, Page 1 1 / 12

slide-2
SLIDE 2

Choosing Individuals and Relations

How to represent: “Pen #7 is red.”

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.1, Page 2 2 / 12

slide-3
SLIDE 3

Choosing Individuals and Relations

How to represent: “Pen #7 is red.” red(pen7). It’s easy to ask “What’s red?” Can’t ask “what is the color of pen7?”

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.1, Page 3 2 / 12

slide-4
SLIDE 4

Choosing Individuals and Relations

How to represent: “Pen #7 is red.” red(pen7). It’s easy to ask “What’s red?” Can’t ask “what is the color of pen7?” color(pen7, red). It’s easy to ask “What’s red?” It’s easy to ask “What is the color of pen7?” Can’t ask “What property of pen7 has value red?”

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.1, Page 4 2 / 12

slide-5
SLIDE 5

Choosing Individuals and Relations

How to represent: “Pen #7 is red.” red(pen7). It’s easy to ask “What’s red?” Can’t ask “what is the color of pen7?” color(pen7, red). It’s easy to ask “What’s red?” It’s easy to ask “What is the color of pen7?” Can’t ask “What property of pen7 has value red?” prop(pen7, color, red). It’s easy to ask all these questions.

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.1, Page 5 2 / 12

slide-6
SLIDE 6

Choosing Individuals and Relations

How to represent: “Pen #7 is red.” red(pen7). It’s easy to ask “What’s red?” Can’t ask “what is the color of pen7?” color(pen7, red). It’s easy to ask “What’s red?” It’s easy to ask “What is the color of pen7?” Can’t ask “What property of pen7 has value red?” prop(pen7, color, red). It’s easy to ask all these questions. prop(Individual, Property, Value) is the only relation needed: called individual-property-value representation

  • r triple representation

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.1, Page 6 2 / 12

slide-7
SLIDE 7

Universality of prop

To represent “a is a parcel”

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.1, Page 7 3 / 12

slide-8
SLIDE 8

Universality of prop

To represent “a is a parcel” prop(a, type, parcel), where type is a special property prop(a, parcel, true), where parcel is a Boolean property

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.1, Page 8 3 / 12

slide-9
SLIDE 9

Reification

To represent scheduled(cs422, 2, 1030, cc208). “section 2 of course cs422 is scheduled at 10:30 in room cc208.”

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.1, Page 9 4 / 12

slide-10
SLIDE 10

Reification

To represent scheduled(cs422, 2, 1030, cc208). “section 2 of course cs422 is scheduled at 10:30 in room cc208.” Let b123 name the booking: prop(b123, course, cs422). prop(b123, section, 2). prop(b123, time, 1030). prop(b123, room, cc208). We have reified the booking. Reify means: to make into an individual. What if we want to add the year?

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.1, Page 10 4 / 12

slide-11
SLIDE 11

Semantic Networks / Knowledge Graphs

When you only have one relation, prop, it can be omitted without loss of information. Logic: prop(Individual, Property, Value) triple: Individual, Property, Value simple sentence: Individual Property Value. graphically:

Obj Val Prop

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.1, Page 11 5 / 12

slide-12
SLIDE 12

An Example Semantic Network / Knowledge Graph

comp_2347

  • wned_by

craig room r107 building comp_sci deliver_to ming room building r117 model lemon_laptop_10000 brand lemon_computer logo lemon_disc color brown size medium weight light packing cardboard_box

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.1, Page 12 6 / 12

slide-13
SLIDE 13

Equivalent Logic Program

prop(comp 2347, owned by, craig). prop(comp 2347, deliver to, ming). prop(comp 2347, model, lemon laptop 10000). prop(comp 2347, brand, lemon computer). prop(comp 2347, logo, lemon disc). prop(comp 2347, color, brown). prop(craig, room, r107). prop(r107, building, comp sci). . . .

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.1, Page 13 7 / 12

slide-14
SLIDE 14

A Structured Semantic Network / Knowledge Graph

comp_2347

  • wned_by

craig room r107 building comp_sci deliver_to ming room building r117 type logo lemon_disc color brown size medium weight light packing cardboard_box subClassOf subClassOf computer lemon_computer lemon_laptop_10000

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.1, Page 14 8 / 12

slide-15
SLIDE 15

Logic of Property

An arc c p − → v from a class c with a property p to value v means every individual in the class has value v on property p: prop(Obj, p, v) ← prop(Obj, type, c). Example: prop(X, weight, light) ← prop(X, type, lemon laptop 10000). prop(X, packing, cardboard box) ← prop(X, type, computer).

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.1, Page 15 9 / 12

slide-16
SLIDE 16

Logic of Property Inheritance

You can do inheritance through the subclass relationship: prop(X, type, T) ← prop(S, subClassOf , T) ∧ prop(X, type, S).

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.1, Page 16 10 / 12

slide-17
SLIDE 17

Multiple Inheritance

An individual is usually a member of more than one class. For example, the same person may be a wine expert, a teacher, a football coach,. . . . The individual can inherit the properties of all of the classes it is a member of: multiple inheritance. With default values,what is an individual inherits conflicting defaults from the different classes? multiple inheritance problem.

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.1, Page 17 11 / 12

slide-18
SLIDE 18

Choosing Primitive and Derived Properties

Associate an property value with the most general class with that property value. Don’t associate contingent properties of a class with the class. For example, if all of current computers just happen to be brown.

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.1, Page 18 12 / 12

slide-19
SLIDE 19

Knowledge Sharing

A conceptualization is a map from the problem domain into the representation. A conceptualization specifies:

◮ What sorts of individuals are being modeled ◮ The vocabulary for specifying individuals, relations and

properties

◮ The meaning or intention of the vocabulary

If more than one person is building a knowledge base, they must be able to share the conceptualization. An ontology is a specification of a conceptualization. An ontology specifies the meanings of the symbols in an information system.

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 1 1 / 20

slide-20
SLIDE 20

Mapping from a conceptualization to a symbol

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 2 2 / 20

slide-21
SLIDE 21

Semantic Web

Ontologies are published on the web in machine readable form. Builders of knowledge bases or web sites adhere to and refer to a published ontology:

◮ a symbol defined by an ontology means the same thing across

web sites that obey the ontology.

◮ if someone wants to refer to something not defined, they

publish an ontology defining the terminology. Others adopt the terminology by referring to the new ontology. In this way, ontologies evolve.

◮ Separately developed ontologies can have mappings between

them published.

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 3 3 / 20

slide-22
SLIDE 22

Challenges of building ontologies

They can be huge: finding the appropriate terminology for a concept may be difficult.

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 4 4 / 20

slide-23
SLIDE 23

Challenges of building ontologies

They can be huge: finding the appropriate terminology for a concept may be difficult. How one divides the world can depend on the application. Different ontologies describe the world in different ways. People can fundamentally disagree about an appropriate structure.

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 5 4 / 20

slide-24
SLIDE 24

Challenges of building ontologies

They can be huge: finding the appropriate terminology for a concept may be difficult. How one divides the world can depend on the application. Different ontologies describe the world in different ways. People can fundamentally disagree about an appropriate structure. Different knowledge bases can use different ontologies. To allow KBs based on different ontologies to inter-operate, there must be mapping between ontologies. It has to be in user’s interests to use an ontology.

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 6 4 / 20

slide-25
SLIDE 25

Challenges of building ontologies

They can be huge: finding the appropriate terminology for a concept may be difficult. How one divides the world can depend on the application. Different ontologies describe the world in different ways. People can fundamentally disagree about an appropriate structure. Different knowledge bases can use different ontologies. To allow KBs based on different ontologies to inter-operate, there must be mapping between ontologies. It has to be in user’s interests to use an ontology. The computer doesn’t understand the meaning of the symbols. The formalism can constrain the meaning, but can’t define it.

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 7 4 / 20

slide-26
SLIDE 26

Semantic Web Technologies

XML the Extensible Markup Language provides generic syntax. tag . . . / or tag . . . . . . /tag. URI a Uniform Resource Identifier is a name of an individual (resource). This name can be shared. Often in the form of a URL to ensure uniqueness. RDF the Resource Description Framework is a language of triples OWL the Web Ontology Language, defines some primitive properties that can be used to define terminology. (Doesn’t define a syntax).

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 8 5 / 20

slide-27
SLIDE 27

Main Components of an Ontology

Individuals the things / objects in the world (not usually specified as part of the ontology) Classes sets of individuals Properties between individuals and their values

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 9 6 / 20

slide-28
SLIDE 28

Individuals

Individuals are things in the world that can be named. (Concrete, abstract, concepts, reified). Unique names assumption (UNA): different names refer to different individuals. The UNA is not an assumption we can universally make: “The Queen”, “Elizabeth Windsor”, etc. Without the determining equality, we can’t count! In OWL we can specify:

  • wl:SameIndividual(i1, i2)
  • wl:DifferentIndividuals(i1, i3)

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 10 7 / 20

slide-29
SLIDE 29

Classes

A class is a set of individuals. E.g., house, building,

  • fficeBuilding

One class can be a subclass of another

  • wl:SubClassOf(house, building)
  • wl:SubClassOf(officeBuilding, building)

The most general class is owl:Thing. Classes can be declared to be the same or to be disjoint:

  • wl:EquivalentClasses(house, singleFamilyDwelling)
  • wl:DisjointClasses(house, officeBuilding)

Different classes are not necessarily disjoint. E.g., a building can be both a commercial building and a residential building.

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 11 8 / 20

slide-30
SLIDE 30

Properties

A property is between an individual and a value. A property has a domain and a range. rdfs:domain(livesIn, person) rdfs:range(livesIn, placeOfResidence)

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 12 9 / 20

slide-31
SLIDE 31

Properties

A property is between an individual and a value. A property has a domain and a range. rdfs:domain(livesIn, person) rdfs:range(livesIn, placeOfResidence) An ObjectProperty is a property whose range is an individual. A DatatypeProperty is one whose range isn’t an individual, e.g., is a number or string.

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 13 9 / 20

slide-32
SLIDE 32

Properties

A property is between an individual and a value. A property has a domain and a range. rdfs:domain(livesIn, person) rdfs:range(livesIn, placeOfResidence) An ObjectProperty is a property whose range is an individual. A DatatypeProperty is one whose range isn’t an individual, e.g., is a number or string. There can also be property hierarchies:

  • wl:subPropertyOf(livesIn, enclosure)
  • wl:subPropertyOf(principalResidence, livesIn)

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 14 9 / 20

slide-33
SLIDE 33

Properties (Cont.)

One property can be inverse of another

  • wl:InverseObjectProperties(livesIn, hasResident)

Properties can be declared to be transitive, symmetric, functional, or inverse-functional.

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 15 10 / 20

slide-34
SLIDE 34

Properties (Cont.)

One property can be inverse of another

  • wl:InverseObjectProperties(livesIn, hasResident)

Properties can be declared to be transitive, symmetric, functional, or inverse-functional. (Which of these are only applicable to object properties?)

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 16 10 / 20

slide-35
SLIDE 35

Properties (Cont.)

One property can be inverse of another

  • wl:InverseObjectProperties(livesIn, hasResident)

Properties can be declared to be transitive, symmetric, functional, or inverse-functional. (Which of these are only applicable to object properties?) We can also state the minimum and maximal cardinality of a property.

  • wl:minCardinality(principalResidence, 1)
  • wl:maxCardinality(principalResidence, 1)

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 17 10 / 20

slide-36
SLIDE 36

Property and Class Restrictions

We can define complex descriptions of classes in terms of restrictions of other classes and properties. E.g., A homeowner is a person who owns a house.

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 18 11 / 20

slide-37
SLIDE 37

Property and Class Restrictions

We can define complex descriptions of classes in terms of restrictions of other classes and properties. E.g., A homeowner is a person who owns a house. homeOwner ⊆ person∩{x : ∃h ∈ house such that x owns h}

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 19 11 / 20

slide-38
SLIDE 38

Property and Class Restrictions

We can define complex descriptions of classes in terms of restrictions of other classes and properties. E.g., A homeowner is a person who owns a house. homeOwner ⊆ person∩{x : ∃h ∈ house such that x owns h}

  • wl:subClassOf(homeOwner,person)
  • wl:subClassOf(homeOwner,
  • wl:ObjectSomeValuesFrom(owns, house))

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 20 11 / 20

slide-39
SLIDE 39

OWL Class Constructors

  • wl:Thing ≡ all individuals
  • wl:Nothing ≡ no individuals
  • wl:ObjectIntersectionOf(C1, . . . , Ck) ≡ C1 ∩ · · · ∩ Ck
  • wl:ObjectUnionOf(C1, . . . , Ck) ≡ C1 ∪ · · · ∪ Ck
  • wl:ObjectComplementOf(C) ≡ Thing \ C
  • wl:ObjectOneOf(I1, . . . , Ik) ≡ {I1, . . . , Ik}
  • wl:ObjectHasValue(P, I) ≡ {x : x P I}
  • wl:ObjectAllValuesFrom(P, C) ≡ {x : x P y → y ∈ C}
  • wl:ObjectSomeValuesFrom(P, C) ≡

{x : ∃y ∈ C such that x P y}

  • wl:ObjectMinCardinality(n, P, C) ≡

{x : #{y|xPy and y ∈ C} ≥ n}

  • wl:ObjectMaxCardinality(n, P, C) ≡

{x : #{y|xPy and y ∈ C} ≤ n}

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 21 12 / 20

slide-40
SLIDE 40

OWL Predicates

rdf:type(I, C) ≡ I ∈ C rdfs:subClassOf(C1, C2) ≡ C1 ⊆ C2

  • wl:EquivalentClasses(C1, C2) ≡ C1 ≡ C2
  • wl:DisjointClasses(C1, C2) ≡ C1 ∩ C2 = {}

rdfs:domain(P, C) ≡ if xPy then x ∈ C rdfs:range(P, C) ≡ if xPy then y ∈ C rdfs:subPropertyOf(P1, P2) ≡ xP1y implies xP2y

  • wl:EquivalentObjectProperties(P1, P2) ≡ xP1y if and only if xP2y
  • wl:DisjointObjectProperties(P1, P2) ≡ xP1y implies not xP2y
  • wl:InverseObjectProperties(P1, P2) ≡ xP1y if and only if yP2x
  • wl:SameIndividual(I1, . . . , In) ≡∀j∀k Ij = Ik
  • wl:DifferentIndividuals(I1, . . . , In) ≡ ∀j∀k j = k implies Ij = Ik
  • wl:FunctionalObjectProperty(P) ≡ if xPy1 and xPy2 then y1 = y2
  • wl:InverseFunctionalObjectProperty(P) ≡

if x1Py and x2Py then x1 = x2

  • wl:TransitiveObjectProperty(P) ≡ if xPy and yPz then xPz
  • wl:SymmetricObjectProperty ≡ if xPy then yPx

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 22 13 / 20

slide-41
SLIDE 41

Knowledge Sharing

One ontology typically imports and builds on other ontologies. OWL provides facilities for version control. Tools for mapping one ontology to another allow inter-operation of different knowledge bases. The semantic web promises to allow two pieces of information to be combined if

◮ they both adhere to an ontology ◮ these are the same ontology or there is a mapping between

them.

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 23 14 / 20

slide-42
SLIDE 42

Example: Apartment Building

An apartment building is a residential building with more than two units and they are rented.

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 24 15 / 20

slide-43
SLIDE 43

Example: Apartment Building

An apartment building is a residential building with more than two units and they are rented. Declaration(ObjectProperty(:numberOfunits)) FunctionalObjectProperty(:numberOfunits) ObjectPropertyDomain(:numberOfunits :ResidentialBuilding) ObjectPropertyRange(:numberOfunits ObjectOneOf(:two :one :moreThanTwo)) Declaration(Class(:ApartmentBuilding)) EquivalentClasses(:ApartmentBuilding ObjectIntersectionOf( :ResidentialBuilding ObjectHasValue(:numberOfunits :moreThanTwo) ObjectHasValue(:ownership :rental)))

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 25 15 / 20

slide-44
SLIDE 44

Aristotelian definitions

Aristotle [350 B.C.] suggested the definition if a class C in terms

  • f:

Genus: the super-class Differentia: the attributes that make members of the class C different from other members of the super-class “If genera are different and co-ordinate, their differentiae are themselves different in kind. Take as an instance the genus ’animal’ and the genus ’knowledge’. ’With feet’, ’two-footed’, ’winged’, ’aquatic’, are differentiae of ’animal’; the species of knowledge are not distinguished by the same differentiae. One species of knowledge does not differ from another in being ’two-footed’.” Aristotle, Categories, 350 B.C.

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 26 16 / 20

slide-45
SLIDE 45

Example: hotel ontology

Define the following: Room BathRoom StandardRoom - what is rented as a room in a hotel Suite RoomOnly

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 27 17 / 20

slide-46
SLIDE 46

Example: hotel ontology

Define the following: Room BathRoom StandardRoom - what is rented as a room in a hotel Suite RoomOnly Hotel HasForRent AllSuitesHotel NoSuitesHotel HasSuitesHotel

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 28 17 / 20

slide-47
SLIDE 47

Top-Level Ontology — Basic Formal Ontology (BFO)

1: if entity continues to exist through time then 2:

it is a continuant

3:

if it doesn’t need another entity for its existence then

4:

it is an independent continuant

5:

if it has matter as a part then

6:

it is a material entity

7:

if it is a single coherent whole then

8:

it is an object

9:

else it is an immaterial entity

10:

else it is a dependent continuant

11:

if it a region in space then

12:

it is a spatial region

13:

else if it is a property then

14:

if it is a property all objects have then

15:

it is a quality

16:

. . . role . . . disposition . . . function . . .

17: else

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 29 18 / 20

slide-48
SLIDE 48

Continuants vs Occurrents

A continuant exists in an instance of time and maintains its identity through time. An occurrent has temporal parts. Continuants participate in occurrents. a person, a life, a finger, infancy: what is part of what?

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 30 19 / 20

slide-49
SLIDE 49

Continuants vs Occurrents

A continuant exists in an instance of time and maintains its identity through time. An occurrent has temporal parts. Continuants participate in occurrents. a person, a life, a finger, infancy: what is part of what? a holiday, the end of a lecture, an email, the sending of an email, the equator, earthquake, a smile, a laugh, the smell of a flower

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 31 19 / 20

slide-50
SLIDE 50

Continuants

a pen, a person, Newtonian mechanics, the memory of a past event:

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 32 20 / 20

slide-51
SLIDE 51

Continuants

a pen, a person, Newtonian mechanics, the memory of a past event: objects

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 33 20 / 20

slide-52
SLIDE 52

Continuants

a pen, a person, Newtonian mechanics, the memory of a past event: objects a flock of birds, the students in CS422, a card collection:

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 34 20 / 20

slide-53
SLIDE 53

Continuants

a pen, a person, Newtonian mechanics, the memory of a past event: objects a flock of birds, the students in CS422, a card collection:

  • bject aggregates

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 35 20 / 20

slide-54
SLIDE 54

Continuants

a pen, a person, Newtonian mechanics, the memory of a past event: objects a flock of birds, the students in CS422, a card collection:

  • bject aggregates

a city, a room, a mouth, the hole of a doughnut:

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 36 20 / 20

slide-55
SLIDE 55

Continuants

a pen, a person, Newtonian mechanics, the memory of a past event: objects a flock of birds, the students in CS422, a card collection:

  • bject aggregates

a city, a room, a mouth, the hole of a doughnut: site

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 37 20 / 20

slide-56
SLIDE 56

Continuants

a pen, a person, Newtonian mechanics, the memory of a past event: objects a flock of birds, the students in CS422, a card collection:

  • bject aggregates

a city, a room, a mouth, the hole of a doughnut: site the dangerous part of a city, part of Grouse Mountain with the best view:

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 38 20 / 20

slide-57
SLIDE 57

Continuants

a pen, a person, Newtonian mechanics, the memory of a past event: objects a flock of birds, the students in CS422, a card collection:

  • bject aggregates

a city, a room, a mouth, the hole of a doughnut: site the dangerous part of a city, part of Grouse Mountain with the best view: fiat part of an object.

c

  • D. Poole and A. Mackworth 2017

Artificial Intelligence, Lecture 14.2, Page 39 20 / 20

slide-58
SLIDE 58

Knowledge Engineering

Overview: Roles of people involved in a knowledge-based system How representation and reasoning systems interact with humans. Knowledge-based interaction and debugging tools Building representation and reasoning systems

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.3, Page 1

slide-59
SLIDE 59

Knowledge-based system architecture

Domain Expert Knowledge Engineer Knowledge Base Inference Engine User Interface

User

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.3, Page 2

slide-60
SLIDE 60

Roles for people in a KBS

Software engineers build the inference engine and user interface. Knowledge engineers design, build, and debug the knowledge base in consultation with domain experts. Domain experts know about the domain, but nothing about particular cases or how the system works. Users have problems for the system, know about particular cases, but not about how the system works or the domain.

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.3, Page 3

slide-61
SLIDE 61

Users

How can users provide knowledge when they don’t know the internals of the system they aren’t experts in the domain they don’t know what information is relevant they don’t know the syntax of the system but they have essential information about the particular case

  • f interest?

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.3, Page 4

slide-62
SLIDE 62

Querying the User

The system can determine what information is relevant and ask the user for the particular information. A top-down derivation can determine what information is

  • relevant. There are three types of goals:

◮ Goals for which the user isn’t expected to know the answer, so

the system never asks.

◮ Goals for which the user should know the answer, and for

which they have not already provided an answer.

◮ Goals for which the user has already provided an answer. c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.3, Page 5

slide-63
SLIDE 63

Yes/No questions

The simplest form of a question is a ground query. Ground queries require an answer of “yes” or “no”. The user is only asked a question if

◮ the question is askable, and ◮ the user hasn’t previously answered the question.

When the user has answered a question, the answer needs to be recorded.

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.3, Page 6

slide-64
SLIDE 64

Electrical Domain

In the electrical domain: The designer of a house:

◮ will know how switches and lights are connected by wires, ◮ won’t know if the light switches are up or down.

A new resident in a house:

◮ won’t know how switches and lights are connected by wires, ◮ will know (or can observe) if the light switches are up or down. c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.3, Page 7

slide-65
SLIDE 65

Functional Relations

You probably don’t want to ask ?age(fred, 0), ?age(fred, 1), ?age(fred, 2), . . . You probably want to ask for Fred’s age once, and succeed for queries for that age and fail for other queries. This exploits the fact that age is a functional relation. Relation r(X, Y ) is functional if, for every X there exists a unique Y such that r(X, Y ) is true.

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.3, Page 8

slide-66
SLIDE 66

Getting information from a user

The user may not know the vocabulary that is expected by the knowledge engineer. Either:

◮ The system designer provides a menu of items from which the

user has to select the best fit.

◮ The user can provide free-form answers. The system needs a

large dictionary to map the responses into the internal forms expected by the system.

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.3, Page 9

slide-67
SLIDE 67

More General Questions

Example: For the subgoal p(a, X, f (Z)) the user can be asked: for which X, Z is p(a, X, f (Z)) true? Should users be expected to give all instances which are true,

  • r should they give the instances one at a time, with the

system prompting for new instances? Example: For which S, C is enrolled(S, C) true? Psychological issues are important.

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.3, Page 10

slide-68
SLIDE 68

Re-asking Questions

For the case when a user provides instances one at a time: When should the system repeat a question or not ask a question? Example: Query Ask? Response ?p(X) yes p(f (Z)) ?p(f (c)) no ?p(a) yes yes ?p(X) yes no ?p(c) no

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.3, Page 11

slide-69
SLIDE 69

When to ask the user

Don’t ask a question that is an instance of a positive answer that has already been given or

  • r instance of a query to which the user has replied

no.

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.3, Page 12

slide-70
SLIDE 70

Delaying Asking the User

Should the system ask the question as soon as it’s encountered, or should it delay the goal until more variables are bound? Example consider query ?p(X) & q(X), where p(X) is askable.

◮ If p(X) succeeds for many instances of X and q(X) succeeds

for few (or no) instances of X it’s better to delay asking p(X) and prove q(X) first.

◮ If p(X) succeeds for few instances of X and q(X) succeeds for

many instances of X, don’t delay.

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.3, Page 13

slide-71
SLIDE 71

Multiple Information Sources

Asking the user is just one instance of using multiple information

  • sources. There are many types of subgoals:

those the system has rules about those the system has facts about those that the user should be able to answer those that a web site may be able to answer (e.g., flight arrival times) those that a database may be able to answer (e.g., someone’s phone number, or the meaning of a word) Each information source has its own characteristics.

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.3, Page 14

slide-72
SLIDE 72

Assumptions

Some subgoals you don’t know if they are true; they are assumptions or hypotheses. You want to collect the assumptions needed to prove the goal. Example: in the electrical domain, ok may be assumable.

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.3, Page 15

slide-73
SLIDE 73

Explanation

The system must be able to justify that its answer is correct, particularly when it is giving advice to a human. The same features can be used for explanation and for debugging the knowledge base. There are three main mechanisms:

◮ Ask HOW a goal was derived. ◮ Ask WHYNOT a goal wasn’t derived. ◮ Ask WHY a subgoal is being proved. c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.4, Page 1

slide-74
SLIDE 74

How did the system prove a goal?

If g is derived, there must be a rule instance g ⇐ a1 & . . . & ak. where each ai is derived. If the user asks HOW g was derived, the system can display this rule. The user can then ask

HOW i.

to give the rule that was used to prove ai. The HOW command moves down the proof tree.

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.4, Page 2

slide-75
SLIDE 75

Why Did the System Ask a Question?

It is useful to find out why a question was asked. Knowing why a question was asked will increase the user’s confidence that the system is working sensibly. It helps the knowledge engineer optimize questions asked of the user. An irrelevant question can be a symptom of a deeper problem. The user may learn something from the system by knowing why the system is doing something.

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.4, Page 3

slide-76
SLIDE 76

WHY question

When the system asks the user a question g, the user can reply with

WHY

This gives the instance of the rule h ⇐ · · · & g & · · · that is being tried to prove h. When the user asks WHY again, it explains why h was proved.

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.4, Page 4

slide-77
SLIDE 77

Debugging Knowledge Bases

There are four types of nonsyntactic errors that can arise in rule-based systems: An incorrect answer is produced; that is, some atom that is false in the intended interpretation was derived. Some answer wasn’t produced; that is, the proof failed when it should have succeeded, or some particular true atom wasn’t derived. The program gets into an infinite loop. The system asks irrelevant questions.

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.4, Page 5

slide-78
SLIDE 78

Debugging Incorrect Answers

An incorrect answer is a derived answer which is false in the intended interpretation. An incorrect answer means a clause in the KB is false in the intended interpretation. If g is false in the intended interpretation, there is a proof for g using g ⇐ a1 & . . . & ak. Either:

◮ Some ai is false: debug it. ◮ All ai are true. This rule is buggy. c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.4, Page 6

slide-79
SLIDE 79

Debugging Missing Answers

WHYNOT g. g fails when it should have succeeded.

Either:

◮ There is an atom in a rule that succeeded with the wrong

answer, use HOW to debug it.

◮ There is an atom in a body that failed when it should have

succeeded, debug it using WHYNOT.

◮ There is a rule missing for g. c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.4, Page 7

slide-80
SLIDE 80

Debugging Infinite Loops

There is no automatic way to debug all such errors: halting problem. There are many errors that can be detected:

◮ If a subgoal is identical to an ancestor in the proof tree, the

program is looping.

◮ Define a well-founded ordering that is reduced each time

through a loop.

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.4, Page 8

slide-81
SLIDE 81

Implementing Knowledge-based Systems

To build an interpreter for a language, we need to distinguish Base language the language of the RRS being implemented. Metalanguage the language used to implement the system. They could even be the same language!

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.5, Page 1

slide-82
SLIDE 82

Implementing the base language

Let’s use the definite clause language as the base language and the metalanguage. We need to represent the base-level constructs in the metalanguage. We represent base-level terms, atoms, and bodies as meta-level terms. We represent base-level clauses as meta-level facts. In the non-ground representation base-level variables are represented as meta-level variables.

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.5, Page 2

slide-83
SLIDE 83

Representing the base level constructs

Base-level atom p(t1, . . . , tn) is represented as the meta-level term p(t1, . . . , tn). Meta-level term oand(e1, e2) denotes the conjunction of base-level bodies e1 and e2. Meta-level constant true denotes the object-level empty body. The meta-level atom clause(h, b) is true if “h if b” is a clause in the base-level knowledge base.

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.5, Page 3

slide-84
SLIDE 84

Example representation

The base-level clauses connected to(l1, w0). connected to(w0, w1) ← up(s2). lit(L) ← light(L) ∧ ok(L) ∧ live(L). can be represented as the meta-level facts clause(connected to(l1, w0), true). clause(connected to(w0, w1), up(s2)). clause(lit(L), oand(light(L), oand(ok(L), live(L)))).

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.5, Page 4

slide-85
SLIDE 85

Making the representation pretty

Use the infix function symbol “&” rather than oand.

◮ instead of writing oand(e1, e2), you write e1 & e2.

Instead of writing clause(h, b) you can write h ⇐ b, where ⇐ is an infix meta-level predicate symbol.

◮ Thus the base-level clause “h ← a1 ∧ · · · ∧ an” is represented

as the meta-level atom h ⇐ a1 & · · · & an.

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.5, Page 5

slide-86
SLIDE 86

Example representation

The base-level clauses connected to(l1, w0). connected to(w0, w1) ← up(s2). lit(L) ← light(L) ∧ ok(L) ∧ live(L). can be represented as the meta-level facts connected to(l1, w0) ⇐ true. connected to(w0, w1) ⇐ up(s2). lit(L) ⇐ light(L) & ok(L) & live(L).

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.5, Page 6

slide-87
SLIDE 87

Vanilla Meta-interpreter

prove(G) is true when base-level body G is a logical consequence

  • f the base-level KB.

prove(true). prove((A & B)) ← prove(A) ∧ prove(B). prove(H) ← (H ⇐ B) ∧ prove(B).

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.5, Page 7

slide-88
SLIDE 88

Example base-level KB

live(W ) ⇐ connected to(W , W1) & live(W1). live(outside) ⇐ true. connected to(w6, w5) ⇐ ok(cb2). connected to(w5, outside) ⇐ true.

  • k(cb2) ⇐ true.

?prove(live(w6)).

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.5, Page 8

slide-89
SLIDE 89

Expanding the base-level

Adding clauses increases what can be proved. Disjunction Let a; b be the base-level representation for the disjunction of a and b. Body a; b is true when a is true, or b is true, or both a and b are true. Built-in predicates You can add built-in predicates such as N is E that is true if expression E evaluates to number N.

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.5, Page 9

slide-90
SLIDE 90

Expanded meta-interpreter

prove(true). prove((A & B)) ← prove(A) ∧ prove(B). prove((A; B)) ← prove(A). prove((A; B)) ← prove(B). prove((N is E)) ← N is E. prove(H) ← (H ⇐ B) ∧ prove(B).

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.5, Page 10

slide-91
SLIDE 91

Depth-Bounded Search

Adding conditions reduces what can be proved. % bprove(G, D) is true if G can be proved with a proof tree of depth less than or equal to number D. bprove(true, D). bprove((A & B), D) ← bprove(A, D) ∧ bprove(B, D). bprove(H, D) ← D ≥ 0 ∧ D1 is D − 1 ∧ (H ⇐ B) ∧ bprove(B, D1).

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.5, Page 11

slide-92
SLIDE 92

Ask-the-user meta-interpreter

% aprove(G) is true if G is a logical consequence of the base-level KB and yes/no answers provided by the user. aprove(true). aprove((A & B)) ← aprove(A) ∧ aprove(B). aprove(H) ← askable(H) ∧ answered(H, yes). aprove(H) ← askable(H) ∧ unanswered(H) ∧ ask(H, Ans) ∧ record(answered(H, Ans)) ∧ Ans = yes. aprove(H) ← (H ⇐ B) ∧ aprove(B).

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.6, Page 1

slide-93
SLIDE 93

Meta-interpreter to collect rules for WHY

% wprove(G, A) is true if G follows from base-level KB, and A is a list of ancestor rules for G. wprove(true, Anc). wprove((A & B), Anc) ← wprove(A, Anc) ∧ wprove(B, Anc). wprove(H, Anc) ← (H ⇐ B) ∧ wprove(B, [(H ⇐ B)|Anc]).

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.6, Page 2

slide-94
SLIDE 94

Delaying Goals

Some goals, rather than being proved, can be collected in a list. To delay subgoals with variables, in the hope that subsequent calls will ground the variables. To delay assumptions, so that you can collect assumptions that are needed to prove a goal. To create new rules that leave out intermediate steps. To reduce a set of goals to primitive predicates.

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.6, Page 3

slide-95
SLIDE 95

Delaying Meta-interpreter

% dprove(G, D0, D1) is true if D0 is an ending of list of delayable atoms D1 and KB ∧ (D1 − D0) | = G. dprove(true, D, D). dprove((A & B), D1, D3) ← dprove(A, D1, D2) ∧ dprove(B, D2, D3). dprove(G, D, [G|D]) ← delay(G). dprove(H, D1, D2) ← (H ⇐ B) ∧ dprove(B, D1, D2).

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.6, Page 4

slide-96
SLIDE 96

Example base-level KB

live(W ) ⇐ connected to(W , W1) & live(W1). live(outside) ⇐ true. connected to(w6, w5) ⇐ ok(cb2). connected to(w5, outside) ⇐ ok(outside connection). delay(ok(X)). ?dprove(live(w6), [ ], D).

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.6, Page 5

slide-97
SLIDE 97

Meta-interpreter that builds a proof tree

% hprove(G, T) is true if G can be proved from the base-level KB, with proof tree T. hprove(true, true). hprove((A & B), (L & R)) ← hprove(A, L) ∧ hprove(B, R). hprove(H, if (H, T)) ← (H ⇐ B) ∧ hprove(B, T).

c

  • D. Poole and A. Mackworth 2016

Artificial Intelligence, Lecture 14.6, Page 6