Tuning SMT Systems on the Training Set Chris Dyer, Patrick Simianer, - - PowerPoint PPT Presentation

tuning smt systems on the training set
SMART_READER_LITE
LIVE PREVIEW

Tuning SMT Systems on the Training Set Chris Dyer, Patrick Simianer, - - PowerPoint PPT Presentation

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler Tuning SMT Systems on the Training Set Chris Dyer, Patrick Simianer, Stefan Riezler, Phil Blunsom, Eva Hasler Project Report MT Marathon 2011 FBK Trento Tuning SMT Systems on the Training Set


slide-1
SLIDE 1

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Tuning SMT Systems on the Training Set

Chris Dyer, Patrick Simianer, Stefan Riezler, Phil Blunsom, Eva Hasler

Project Report MT Marathon 2011 FBK Trento

slide-2
SLIDE 2

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Tuning SMT Systems on the Training Set

Goal: Discriminative training using sparse features on the full training set

slide-3
SLIDE 3

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Tuning SMT Systems on the Training Set

Goal: Discriminative training using sparse features on the full training set Approach: Picky-picky / elitist learning:

slide-4
SLIDE 4

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Tuning SMT Systems on the Training Set

Goal: Discriminative training using sparse features on the full training set Approach: Picky-picky / elitist learning: Stochastic learning with true random selection of examples.

slide-5
SLIDE 5

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Tuning SMT Systems on the Training Set

Goal: Discriminative training using sparse features on the full training set Approach: Picky-picky / elitist learning: Stochastic learning with true random selection of examples. Feature selection according to various regularization criteria.

slide-6
SLIDE 6

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Tuning SMT Systems on the Training Set

Goal: Discriminative training using sparse features on the full training set Approach: Picky-picky / elitist learning: Stochastic learning with true random selection of examples. Feature selection according to various regularization criteria. Leave-one-out estimation: Leave out sentence/shard currently being trained on when extracting rules/features in training.

slide-7
SLIDE 7

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

SMT Framework + Data

cdec decoder (https://github.com/redpony/cdec)

slide-8
SLIDE 8

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

SMT Framework + Data

cdec decoder (https://github.com/redpony/cdec) Hiero SCFG grammars

slide-9
SLIDE 9

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

SMT Framework + Data

cdec decoder (https://github.com/redpony/cdec) Hiero SCFG grammars WMT11 news-commentary corpus

slide-10
SLIDE 10

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

SMT Framework + Data

cdec decoder (https://github.com/redpony/cdec) Hiero SCFG grammars WMT11 news-commentary corpus

132,755 parallel sentences

slide-11
SLIDE 11

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

SMT Framework + Data

cdec decoder (https://github.com/redpony/cdec) Hiero SCFG grammars WMT11 news-commentary corpus

132,755 parallel sentences German-to-English

slide-12
SLIDE 12

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Learning Framework: SGD for Pairwise Ranking

slide-13
SLIDE 13

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Constraint Selection = Sampling of Pairs

Random sampling of pairs from full chart for pairwise ranking:

slide-14
SLIDE 14

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Constraint Selection = Sampling of Pairs

Random sampling of pairs from full chart for pairwise ranking:

First sample translations according to their model score.

slide-15
SLIDE 15

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Constraint Selection = Sampling of Pairs

Random sampling of pairs from full chart for pairwise ranking:

First sample translations according to their model score. Then sample pairs.

slide-16
SLIDE 16

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Constraint Selection = Sampling of Pairs

Random sampling of pairs from full chart for pairwise ranking:

First sample translations according to their model score. Then sample pairs.

Sampling will diminish problem of learning to discriminate translations that are too close (in terms of sentence-wise

  • approx. BLEU) to each other.
slide-17
SLIDE 17

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Constraint Selection = Sampling of Pairs

Random sampling of pairs from full chart for pairwise ranking:

First sample translations according to their model score. Then sample pairs.

Sampling will diminish problem of learning to discriminate translations that are too close (in terms of sentence-wise

  • approx. BLEU) to each other.

Sampling will also speed up learning.

slide-18
SLIDE 18

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Constraint Selection = Sampling of Pairs

Random sampling of pairs from full chart for pairwise ranking:

First sample translations according to their model score. Then sample pairs.

Sampling will diminish problem of learning to discriminate translations that are too close (in terms of sentence-wise

  • approx. BLEU) to each other.

Sampling will also speed up learning. Lots of variations on sampling possible ...

slide-19
SLIDE 19

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Candidate Features

Efficient computation of features from local rule context:

slide-20
SLIDE 20

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Candidate Features

Efficient computation of features from local rule context:

Hiero SCFG rule identifier

slide-21
SLIDE 21

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Candidate Features

Efficient computation of features from local rule context:

Hiero SCFG rule identifier target n-grams within rule

slide-22
SLIDE 22

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Candidate Features

Efficient computation of features from local rule context:

Hiero SCFG rule identifier target n-grams within rule target n-gram with gaps (X) within rule

slide-23
SLIDE 23

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Candidate Features

Efficient computation of features from local rule context:

Hiero SCFG rule identifier target n-grams within rule target n-gram with gaps (X) within rule binned rule counts in full training set

slide-24
SLIDE 24

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Candidate Features

Efficient computation of features from local rule context:

Hiero SCFG rule identifier target n-grams within rule target n-gram with gaps (X) within rule binned rule counts in full training set rule length features

slide-25
SLIDE 25

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Candidate Features

Efficient computation of features from local rule context:

Hiero SCFG rule identifier target n-grams within rule target n-gram with gaps (X) within rule binned rule counts in full training set rule length features rule shape features

slide-26
SLIDE 26

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Candidate Features

Efficient computation of features from local rule context:

Hiero SCFG rule identifier target n-grams within rule target n-gram with gaps (X) within rule binned rule counts in full training set rule length features rule shape features word alignments in rules

slide-27
SLIDE 27

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Candidate Features

Efficient computation of features from local rule context:

Hiero SCFG rule identifier target n-grams within rule target n-gram with gaps (X) within rule binned rule counts in full training set rule length features rule shape features word alignments in rules

... and many more!

slide-28
SLIDE 28

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Feature Selection

ℓ1/ℓ2-regularization

slide-29
SLIDE 29

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Feature Selection

ℓ1/ℓ2-regularization

Compute ℓ2-norm of column vectors (= vector of examples/shards for each of n features), then ℓ1-norm of resulting n-dimensional vector.

slide-30
SLIDE 30

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Feature Selection

ℓ1/ℓ2-regularization

Compute ℓ2-norm of column vectors (= vector of examples/shards for each of n features), then ℓ1-norm of resulting n-dimensional vector.

slide-31
SLIDE 31

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Feature Selection

ℓ1/ℓ2-regularization

Compute ℓ2-norm of column vectors (= vector of examples/shards for each of n features), then ℓ1-norm of resulting n-dimensional vector.

Effect is to choose small subset of features that are useful across all examples/shards

slide-32
SLIDE 32

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Feature Selection, done properly

Incremental gradient-based selection of column vectors (Obozinski, Taskar, Jordan: Joint covariant selection and joint subspace selection for multiple classification

  • problems. Stat Comput (2010))
slide-33
SLIDE 33

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Feature Selection, done properly

Incremental gradient-based selection of column vectors (Obozinski, Taskar, Jordan: Joint covariant selection and joint subspace selection for multiple classification

  • problems. Stat Comput (2010))
slide-34
SLIDE 34

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Feature Selection, quick and dirty

Combine feature selection with averaging:

slide-35
SLIDE 35

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Feature Selection, quick and dirty

Combine feature selection with averaging:

Keep only those features with large enough ℓ2-norm computed over examples/shards.

slide-36
SLIDE 36

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Feature Selection, quick and dirty

Combine feature selection with averaging:

Keep only those features with large enough ℓ2-norm computed over examples/shards. Then average feature values over examples/shards.

slide-37
SLIDE 37

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

How far did we get in a few days?

First full training run finished!

slide-38
SLIDE 38

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

How far did we get in a few days?

First full training run finished!

150k parallel sentences from news commentary data, German-to-English

slide-39
SLIDE 39

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

How far did we get in a few days?

First full training run finished!

150k parallel sentences from news commentary data, German-to-English pairwise ranking perceptron

slide-40
SLIDE 40

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

How far did we get in a few days?

First full training run finished!

150k parallel sentences from news commentary data, German-to-English pairwise ranking perceptron sample 100 translations from chart, use all 100*(99)/2 pairs

slide-41
SLIDE 41

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

How far did we get in a few days?

First full training run finished!

150k parallel sentences from news commentary data, German-to-English pairwise ranking perceptron sample 100 translations from chart, use all 100*(99)/2 pairs OR: use n-best list sparse rule-id features AND/OR dense features

slide-42
SLIDE 42

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

How far did we get in a few days?

First full training run finished!

150k parallel sentences from news commentary data, German-to-English pairwise ranking perceptron sample 100 translations from chart, use all 100*(99)/2 pairs OR: use n-best list sparse rule-id features AND/OR dense features 200 shards (25 machines with 8 cores)

slide-43
SLIDE 43

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Results

Still a lot of bugs due to integration of code from different sources

slide-44
SLIDE 44

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Results

Still a lot of bugs due to integration of code from different sources Infrastructure is working

slide-45
SLIDE 45

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Results

Still a lot of bugs due to integration of code from different sources Infrastructure is working Experiments still running

slide-46
SLIDE 46

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Results

Still a lot of bugs due to integration of code from different sources Infrastructure is working Experiments still running Sensible things happening:

Best rule X → X1 , dass X2, X1 that X2 Bad rule X → X1 oder X2, X1 and X2

slide-47
SLIDE 47

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Results

Still a lot of bugs due to integration of code from different sources Infrastructure is working Experiments still running Sensible things happening:

Best rule X → X1 , dass X2, X1 that X2 Bad rule X → X1 oder X2, X1 and X2

At the moment still trailing behind MERT ...

slide-48
SLIDE 48

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Results

Still a lot of bugs due to integration of code from different sources Infrastructure is working Experiments still running Sensible things happening:

Best rule X → X1 , dass X2, X1 that X2 Bad rule X → X1 oder X2, X1 and X2

At the moment still trailing behind MERT ... We’ll catch up!

slide-49
SLIDE 49

ToTS Dyer, Simianer, Riezler, Blunsom, Hasler

Thanks

Thanks to organizers for great

  • pportunity to learn/chat/hobnob!