Machine Translation Spring 2020 Adapted from slides from Chris - - PowerPoint PPT Presentation

machine translation
SMART_READER_LITE
LIVE PREVIEW

Machine Translation Spring 2020 Adapted from slides from Chris - - PowerPoint PPT Presentation

SFU NatLangLab CMPT 825: Natural Language Processing Machine Translation Spring 2020 Adapted from slides from Chris Manning, Abigail See, Matthew Lamm, Danqi Chen and Karthik Narasimhan Translation One of the holy grail problems in


slide-1
SLIDE 1

Machine Translation

Spring 2020 CMPT 825: Natural Language Processing

SFU NatLangLab

Adapted from slides from Chris Manning, Abigail See, Matthew Lamm, Danqi Chen and Karthik Narasimhan

slide-2
SLIDE 2
  • One of the “holy grail” problems in artificial intelligence
  • Practical use case: Facilitate communication between people in the

world

  • Extremely challenging (especially for low-resource languages)

Translation

slide-3
SLIDE 3

Easy and not so easy translations

  • Easy:
  • I like apples

ich mag Äpfel (German)

  • Not so easy:
  • I like apples

J'aime les pommes (French)

  • I like red apples

J'aime les pommes rouges (French)

  • les

the but les pommes apples

↔ ↔ ↔ ↔ ↔

slide-4
SLIDE 4

MT basics

  • Goal: Translate a sentence

in a source language (input) to a sentence in the target language (output)

  • Can be formulated as an optimization problem:
  • where is a scoring function over source and target sentences
  • Requires two components:
  • Learning algorithm to compute parameters of
  • Decoding algorithm for computing the best translation

w(s) ̂ w(t) = arg max

w(t) ψ (w(s), w(t))

ψ ψ ̂ w(t)

slide-5
SLIDE 5

Why is MT challenging?

  • Single words may be replaced with multi-word phrases
  • I like apples

J'aime les pommes

  • Reordering of phrases
  • I like red apples

J'aime les pommes rouges

  • Contextual dependence
  • les

the but les pommes apples

↔ ↔ ↔ ↔

Extremely large output space Decoding is NP-hard

slide-6
SLIDE 6

Vauquois Pyramid

  • Hierarchy of concepts and distances between them in different languages
  • Lowest level: individual words/characters
  • Higher levels: syntax, semantics
  • Interlingua: Generic language-agnostic representation of meaning
slide-7
SLIDE 7

Evaluating translation quality

  • Two main criteria:
  • Adequacy: Translation

should adequately reflect the linguistic content of

  • Fluency: Translation

should be fluent text in the target language

w(t) w(s) w(t)

Different translations of A Vinay le gusta Python

slide-8
SLIDE 8

Evaluation metrics

  • Manual evaluation is most accurate, but expensive
  • Automated evaluation metrics:
  • Compare system hypothesis with reference translations
  • BiLingual Evaluation Understudy (BLEU) (Papineni et al., 2002):
  • Modified n-gram precision
slide-9
SLIDE 9

BLEU

Two modifications:

  • To avoid

, all precisions are smoothed

  • Each n-gram in reference can be used at most once
  • Ex. Hypothesis: to to to to to vs Reference: to be or not to be

should not get a unigram precision of 1 Precision-based metrics favor short translations

  • Solution: Multiply score with a brevity penalty for translations

shorter than reference,

BLEU = exp 1

N

N

n=1

log pn

log 0 e1−r/h

slide-10
SLIDE 10

BLEU

  • Correlates somewhat well with human judgements

(G. Doddington, NIST)

slide-11
SLIDE 11

BLEU scores

Sample BLEU scores for various system outputs

  • Alternatives have been proposed:
  • METEOR: weighted F-measure
  • Translation Error Rate (TER): Edit distance between hypothesis

and reference

Issues?

slide-12
SLIDE 12

Machine Translation (MT)

task of translating a sentence from one language (the source language) to a sentence in another language (the target language)

slide-13
SLIDE 13

History

  • Started in the 1950s: rule-based, tightly linked to formal linguistics theories
  • 1980s: Statistical MT
  • 2000s-2015: Statistical Phrase-Based MT
  • 2015-Present: Neural Machine Translation

13

  • Russian → English (motivated by the Cold War!)
  • Systems were mostly rule-based, using a bilingual

dictionary to map Russian words to their English counterparts

1 minute video showing 1954 MT: https://youtu.be/K-HfpsHPmvw

slide-14
SLIDE 14

History

  • Started in the 1950s: rule-based, tightly linked to formal linguistics theories
  • (late) 1980s to 2000s: Statistical MT
  • 2000s-2014: Statistical Phrase-Based MT
  • 2014-Present: Neural Machine Translation

14

slide-15
SLIDE 15

History

  • Started in the 1950s: rule-based, tightly linked to formal linguistics theories
  • 1980s: Statistical MT
  • 2000s-2015: Statistical Phrase-Based MT
  • 2015-Present: Neural Machine Translation

15

slide-16
SLIDE 16

Statistical MT

  • Key Idea: Learn probabilistic model from data
  • To find the best English sentence y, given French sentence x:

̂ y = arg max

y

P(y|x) ̂ y = arg max

y

P(x|y)P(y)

  • Decompose using Bayes Rule:

Translation/Alignment Model Models how words and phrases should be translated (adequacy/fidelity). Learn from parallel data Language Model Models how to write good English (fluency) Learn from monolingual data

slide-17
SLIDE 17

Noisy channel model

  • Generative process for source sentence
  • Use Bayes rule to recover

that is maximally likely under the conditional distribution (which is what we want)

w(t) pT|S

̂ w(t) = arg max

w(t) ψ (w(s), w(t))

Decoder

pS|T

Noisy Channel

w(s)

Input Model

pT w(t)

ψ (w(s), w(t)) = ψA (w(s), w(t)) + ψF (w(t)) log PS,T(w(s), w(t)) = log PS|T(w(s)|w(t)) + log PT(w(t))

slide-18
SLIDE 18

Data

  • Statistical MT relies requires lots of parallel corpora


(Europarl, Koehn, 2005)

  • Not available for many low-resource languages in the world
slide-19
SLIDE 19

How to define the translation model?

Introduce latent variable modeling the alignment (word-level correspondence) between the source sentence x and the target sentence y

P(x|y) = P(x, A|y)

slide-20
SLIDE 20

What is alignment?

slide-21
SLIDE 21

Alignment is complex

Alignment can be many-to-one

slide-22
SLIDE 22

Alignment is complex

Alignment can be one-to-many

slide-23
SLIDE 23

Alignment is complex

Some words are very fertile!

slide-24
SLIDE 24

Alignment is complex

Alignment can be many-to-many (phrase-level)

slide-25
SLIDE 25

How to define the translation model?

Given the alignment, how do we incorporate in our model?

P(x|y) = P(x, A|y)

slide-26
SLIDE 26

Incorporating alignments

  • Joint probability of alignment and translation can be defined as:
  • are the number of words in source and target

sentences

  • is the alignment of the

word in the source sentence, i.e. it specifies that the word is aligned to the word in target

M(s), M(t)

am mth mth amth

Is this sufficient?

slide-27
SLIDE 27

Incorporating alignments

a1 = 2, a2 = 3, a3 = 4,...

Multiple source words may align to the same target word! (source) (target)

slide-28
SLIDE 28

Reordering and word insertion

(Slide credit: Brendan O’Connor)

Assume extra NULL token

slide-29
SLIDE 29

Independence assumptions

  • Two independence assumptions:
  • Alignment probability factors across tokens:
  • Translation probability factors across tokens:
slide-30
SLIDE 30

How do we translate?

  • We want:
  • Sum over all possible alignments:
  • Alternatively, take the max over alignments

arg max

w(t) p(w(t)|w(s)) = arg max w(t)

p(w(s), w(t)) p(w(s))

slide-31
SLIDE 31

Alignments

  • Key question: How should we align words in source to

words in target? good bad

slide-32
SLIDE 32

IBM Model 1

  • Assume
  • Is this a good assumption?


p(am|m, M(s), M(t)) = 1 M(t)

Every alignment is equally likely!

slide-33
SLIDE 33
  • Each source word is aligned to at most one target word
  • Further, assume
  • We then have: 

  • How do we estimate

?

p(am|m, M(s), M(t)) = 1 M(t) p(w(s), w(t)) = p(w(t))∑

A

( 1 M(t) )M(s) p(w(s)|w(t)) p(w(s) = v|w(t) = u)

IBM Model 1

slide-34
SLIDE 34
  • If we had word-to-word alignments, we could compute the

probabilities using the MLE:

  • where

= #instances where word was aligned to word in the training set

  • However, word-to-word alignments are often hard to come by

p(v|u) = count(u, v) count(u) count(u, v) u v

IBM Model 1

What can we do?

slide-35
SLIDE 35

EM for Model 1

  • (E-Step) If we had an accurate translation model, we can

estimate likelihood of each alignment as:

  • (M Step) Use expected count to re-estimate translation

parameters:
 p(v|u) =

Eq[count(u, v)] count(u)

slide-36
SLIDE 36

Independence assumptions allow for Viterbi decoding

In general, use greedy or beam decoding

slide-37
SLIDE 37

Model 1: Decoding

  • Pick target sentence length
  • Decode:

M(t) arg max

w(t) p(w(t)|w(s)) = arg max w(t) p(w(s), w(t))

p(w(s), w(t)) = p(w(t))∑

A

1 M(t) p(w(s)|w(t))

slide-38
SLIDE 38

Model 1: Decoding

(source) (target) At every step , pick target word to maximize product of:

  • 1. Language model:
  • 2. Translation model:

where is the inverse alignment from target to source

m pLM(w(t)

m |w(t) <m)

p(w(s)

bm |w(t) m )

bm

slide-39
SLIDE 39

IBM Model 2

  • Slightly relaxed assumption:
  • is also estimated, not set to

constant

p(am|m, M(s), M(t))

  • Original independence assumptions still required:
  • Alignment probability factors across tokens:
  • Translation probability factors across tokens:
slide-40
SLIDE 40

Other IBM models

  • Models 3 - 6 make successively weaker assumptions
  • But get progressively harder to optimize
  • Simpler models are often used to ‘initialize’ complex ones
  • e.g train Model 1 and use it to initialize Model 2 parameters
slide-41
SLIDE 41

Phrase-based MT

  • Word-by-word translation is not sufficient in many cases
  • Solution: build alignments and translation tables between

multiword spans or “phrases”

(literal) (actual)

slide-42
SLIDE 42

Phrase-based MT

  • Solution: build alignments and translation tables between

multiword spans or “phrases”

  • Translations condition on multi-word units and assign

probabilities to multi-word units

  • Alignments map from spans to spans
slide-43
SLIDE 43

Syntactic MT

(Slide credit: Greg Durrett)

slide-44
SLIDE 44

Syntactic MT

slide-45
SLIDE 45

Vauquois Pyramid

  • Hierarchy of concepts and distances between them in different languages
  • Lowest level: individual words/characters
  • Higher levels: syntax, semantics
  • Interlingua: Generic language-agnostic representation of meaning
slide-46
SLIDE 46

Statistical MT

  • 1990s to 2010s: huge research area
  • Extremely complex systems
  • Many separately-designed subcomponents
  • Lots of feature engineering
  • Needed to compile/maintain extra resources (phrase

tables)

  • Lots of human effort to maintain
slide-47
SLIDE 47

History

  • Started in the 1950s: rule-based, tightly linked to formal linguistics theories
  • 1980s: Statistical MT
  • 2000s-2015: Statistical Phrase-Based MT
  • 2015-Present: Neural Machine Translation
  • ~2018-Present: Neural Machine Translation + PBMT Hybrid

47