Integrating Logical Representations with Probabilistic Information - - PowerPoint PPT Presentation

integrating logical representations with probabilistic
SMART_READER_LITE
LIVE PREVIEW

Integrating Logical Representations with Probabilistic Information - - PowerPoint PPT Presentation

Integrating Logical Representations with Probabilistic Information using Markov Logic Dan Garrette, Katrin Erk, and Raymond Mooney The University of Texas at Austin 1 Overview Some phenomena best modeled through logic , others statistically


slide-1
SLIDE 1

Integrating Logical Representations with Probabilistic Information using Markov Logic

Dan Garrette, Katrin Erk, and Raymond Mooney The University of Texas at Austin

1

slide-2
SLIDE 2

Overview

Some phenomena best modeled through logic, others statistically Aim: a unified framework for both We present first steps towards this goal Basic framework: Markov Logic Technical solutions for phenomena

2

slide-3
SLIDE 3

Introduction

3

slide-4
SLIDE 4

Semantics

Represent the meaning of language Logical Models Probabilistic Models

4

slide-5
SLIDE 5

Phenomena Modeled with Logic

Standard first-order logic concepts

  • Negation
  • Quantification: universal, existential

Implicativity / factivity

5

slide-6
SLIDE 6

Implicativity / Factivity

Presuppose truth or falsity of complement Influenced by polarity of environment

6

slide-7
SLIDE 7

Implicativity / Factivity

“Ed knows Mary left.”

➡ Mary left

“Ed refused to lock the door.”

➡ Ed did not lock the door

7

slide-8
SLIDE 8

“Ed did not forget to ensure that Dave failed.”

➡ Dave failed

“Ed hopes that Dave failed.”

➡ ??

8

Implicativity / Factivity

slide-9
SLIDE 9

Word Similarity Synonyms Hypernyms / hyponyms

9

Phenomena Modeled Statistically

slide-10
SLIDE 10

10

Synonymy

“The wine left a stain.”

➡ paraphrase: “result in”

“He left the children with the nurse.”

➡ paraphrase: “entrust”

slide-11
SLIDE 11

11

Hypernymy

“The bat flew out of the cave.”

➡ hypernym: “animal”

“The player picked up the bat.”

➡ hypernym: “stick”

slide-12
SLIDE 12

Hypernymy and Polarity

12

“John does not own a vehicle”

➡ John does not own a car

“John owns a car”

➡ John owns a vehicle

vehicle boat car truck vehicle boat car truck

slide-13
SLIDE 13

Our Goal

A unified semantic representation incorporate logic and probabilities interaction between the two Ability to reason with this representation

13

slide-14
SLIDE 14

Our Solution

Markov Logic “Softened” first order logic: weighted formulas Judge likelihood of inference

14

slide-15
SLIDE 15

Evaluating Understanding

How can we tell if our semantic representation is correct? Need a way to measure comprehension Textual Entailment: determine whether

  • ne text implies another

15

slide-16
SLIDE 16

Textual Entailment

16

premise: iTunes software has seen strong

sales in Europe. Yes

hypothesis: Strong sales for iTunes in Europe.

Yes

premise: Oracle had fought to keep the

forms from being released No

hypothesis: Oracle released a confidential

document No

slide-17
SLIDE 17

Textual Entailment

Requires deep understanding of text Allows us to construct test data that targets our specific phenomena

17

slide-18
SLIDE 18

Motivation

18

slide-19
SLIDE 19

Bos-style Logical RTE

Generate rules linking all possible paraphrases Unable to distinguish between good and bad paraphrases

19

slide-20
SLIDE 20

Bos-style Logical RTE

“The player picked up the bat.”

20

⊧ “The player picked up the animal” ⊧ “The player picked up the stick”

slide-21
SLIDE 21

Distributional-Only

Able to judge similarity Unable to properly handle logical phenomena

21

slide-22
SLIDE 22

Our Approach

Handle logical phenomena discretely Handle probabilistic phenomena with weighted formulas Do both simultaneously, allowing them to influence each other

22

slide-23
SLIDE 23

Background

23

slide-24
SLIDE 24

Logical Semantics

Semanticists have traditionally represented meaning with formal logic We use Boxer (Bos et al., 2004) to generate Discourse Representation Structures (Kamp and Reyle, 1993)

24

slide-25
SLIDE 25

Logical Semantics

25

“John did not manage to leave”

x0 x0 name named(x0 med(x0, john, per) r) e1 l2 e1 l2

¬

mana even agen theme prop manage(e1) event(e1) agent(e1, x0) theme(e1, l2) proposition(l2)

¬

l2: e3 leave(e3) event(e3) agent(e3, x0)

slide-26
SLIDE 26

Logical Semantics

26

“John did not manage to leave”

x0 x0 name named(x0 med(x0, john, per) r) e1 l2 e1 l2

¬

mana even agen theme prop manage(e1) event(e1) agent(e1, x0) theme(e1, l2) proposition(l2)

¬

l2: e3 leave(e3) event(e3) agent(e3, x0)

Boxes have existentially quantified variables ...and atomic formulas ...and logical operators

slide-27
SLIDE 27

Logical Semantics

27

“John did not manage to leave”

x0 x0 name named(x0 med(x0, john, per) r) e1 l2 e1 l2

¬

mana even agen theme prop manage(e1) event(e1) agent(e1, x0) theme(e1, l2) proposition(l2)

¬

l2: e3 leave(e3) event(e3) agent(e3, x0)

Box structure shows scope Labels allow reference to entire boxes

slide-28
SLIDE 28

Logical Semantics

Powerful, flexible representation Straightforward inference procedure Why use First Order Logic?

28

Unable to handle uncertainty Natural language is not discrete Why Not?

slide-29
SLIDE 29

Distributional Semantics

Describe word meaning by its context Representation is a continuous function

29

slide-30
SLIDE 30

30

Distributional Semantics

“leave”

“result in” “entrust” “The wine left a stain” “He left the children with the nurse”

slide-31
SLIDE 31

Can predict word-in-context similarity Can be learned in an unsupervised fashion

Why use Distributional Models?

31

Distributional Semantics

Incomplete representation of semantics No concept of negation, quantification, etc

Why Not?

slide-32
SLIDE 32

Approach

32

slide-33
SLIDE 33

Flatten DRS into first order representation Add weighted word-similarity constraints

33

Approach

slide-34
SLIDE 34

Standard FOL Conversion

34

∃ x0.(ne_per_john(x0) & ∃ e1 l2.(manage(e1) &

event(e1) & agent(e1, x0) & theme(e1, l2) & proposition(l2) &

∃ e3.(leave(e3) &

event(e3) & agent(e3, x0)))) x0 x0 name named(x0 med(x0, john, per) r) e1 l2 e1 l2

¬

mana even agen theme prop manage(e1) event(e1) agent(e1, x0) theme(e1, l2) proposition(l2)

¬

l2: e3 leave(e3) event(e3) agent(e3, x0)

¬

“John did not manage to leave”

slide-35
SLIDE 35

Standard FOL Conversion

35

∃ x0.(ne_per_john(x0) & ∃ e1 l2.(manage(e1) &

event(e1) & agent(e1, x0) & theme(e1, l2) & proposition(l2) &

∃ e3.(leave(e3) &

event(e3) & agent(e3, x0)))) x0 x0 name named(x0 med(x0, john, per) r) e1 l2 e1 l2

¬

mana even agen theme prop manage(e1) event(e1) agent(e1, x0) theme(e1, l2) proposition(l2)

¬

l2: e3 leave(e3) event(e3) agent(e3, x0)

¬

DRT allows the theme proposition to be labeled as “l2” The conversion loses track of what “l2” labels

“John did not manage to leave”

slide-36
SLIDE 36

Standard FOL Conversion

36

∃ x0 e1 l2.(ne_per_john(x0) &

forget(e1) & event(e1) & agent(e1, x0) & theme(e1, l2) & proposition(l2) &

∃ e3.(leave(e3) &

event(e3) & agent(e3, x0)))

“John forgot to leave” “John left”

∃ x0 e3.(ne_per_john(x0) &

leave(e3) & event(e3) & agent(e3, x0))

slide-37
SLIDE 37

Standard FOL Conversion

37

“John left”

∃ x0 e3.(ne_per_john(x0) &

leave(e3) & event(e3) & agent(e3, x0))

∃ x0 e1 l2 e3.(ne_per_john(x0) &

forget(e1) & event(e1) & agent(e1, x0) & theme(e1, l2) & proposition(l2) & leave(e3) & event(e3) & agent(e3, x0))

“John forgot to leave”

⊧ ⊧

slide-38
SLIDE 38

l0:

38

named(l0, ne_per_john, x0) l1:

Our FOL Conversion

pred(l2, leave, e3) event(l2, e3) rel(l2, agent, e3, x0) not(l0, l1) x0 x0 name named(x0 med(x0, john, per) r) e1 l2 e1 l2

¬

mana even agen theme prop manage(e1) event(e1) agent(e1, x0) theme(e1, l2) proposition(l2)

¬

l2: e3 leave(e3) event(e3) agent(e3, x0) pred(l1, manage, e1) event(l1, e1) rel(l1, agent, e1, x0) rel(l1, theme, e1, l2) prop(l1, l2)

label “l2” is maintained

true(l0)

slide-39
SLIDE 39

39

∀ p c.[(true(p) ∧ not(p,c)) → false(c)]] ∀ p c.[(false(p) ∧ not(p,c)) → true(c)]]

Our FOL Conversion

With “connectives” as predicates, rules are needed to capture relationships:

slide-40
SLIDE 40

40

∀ l1 l2 e.[(pred(l1, “forget”, e) ∧ true(l1) ∧ rel(l1, “theme”, e, l2)) → false(l2)]

Implicativity / Factivity

Calculate truth values of nested propositions For example, “forget to” is downward entailing in positive contexts:

slide-41
SLIDE 41

Word-Similarity

41

sweep

brush move sail broom wipe embroil tangle drag involve traverse span cover extend clean win continue swing wield handle manage

“A stadium craze is sweeping the country”

synset1: synset2: synset3: synset4: synset5: synset6: synset7: synset8: synset9:

slide-42
SLIDE 42

Word-Similarity

42

sweep

brush move sail broom wipe embroil tangle drag involve traverse span cover extend clean win continue swing wield handle manage

“A stadium craze is sweeping the country”

slide-43
SLIDE 43

Word-Similarity

43

paraphrase continue move win cover clean handle embroil wipe brush traverse sail, span, ...

“A stadium craze is sweeping the country”

rank 1 2 3 4 5 6 7 8 9 10 11 P = 1/rank 0.50 0.33 0.25 0.20 0.17 0.14 0.13 0.11 0.10 0.09 0.08 W = log(P/(1-P)) 0.00

  • 1.00
  • 1.58
  • 2.00
  • 2.32
  • 2.58
  • 2.81
  • 3.00
  • 3.17
  • 3.32
  • 3.46

penalties increase with rank

slide-44
SLIDE 44

Word-Similarity

44

“A stadium craze is sweeping the country” ∀ l x.[pred(l, “sweep”, x) ↔ pred(l, “ ”, x)] ∀ l x.[pred(l, “sweep”, x) ↔ pred(l, “ ”, x)] Inject a rule for every possible paraphrase MLN decides which to use

  • 2.00
  • 3.17

cover brush

slide-45
SLIDE 45

Evaluation

45

slide-46
SLIDE 46

Executed over 100 hand-written examples Hand-write examples instead of using RTE data to target specific phenomena Examples discussed in this talk are handled correctly by the system

46

Evaluation

slide-47
SLIDE 47

Example

47

p: South Korea fails to honor U.S. patents hgood: South Korea does not observe U.S. patents hbad*: South Korea does not reward U.S. patents

“fail to” is negatively entailing in positive environments In context, “observe” is a better paraphrase than “reward”

slide-48
SLIDE 48

Conclusion

48

slide-49
SLIDE 49

Conclusion

Presented unified logical/statistical framework for semantics Markov Logic Allows interaction between logic and probabilities Technical solutions for phenomena

49

slide-50
SLIDE 50

Large-scale evaluation Address a larger number of phenomena

50

Next Steps

slide-51
SLIDE 51

Thank You!

51