Learning Unified Multi-Document Summarization From Collaborative - - PowerPoint PPT Presentation

learning unified multi document summarization from
SMART_READER_LITE
LIVE PREVIEW

Learning Unified Multi-Document Summarization From Collaborative - - PowerPoint PPT Presentation

Learning Unified Multi-Document Summarization From Collaborative Journalism Masters Thesis by Yasar Naci Gndz First Referee : Prof.Dr.Benno Stein Second Referee : Prof.Dr.Andreas Jakoby 1 INTRODUCTION: New age, new habits 2


slide-1
SLIDE 1

Learning Unified Multi-Document Summarization From Collaborative Journalism

Master’s Thesis by Yasar Naci Gündüz First Referee : Prof.Dr.Benno Stein Second Referee : Prof.Dr.Andreas Jakoby

1

slide-2
SLIDE 2

2

INTRODUCTION: New age, new habits

slide-3
SLIDE 3

3

INTRODUCTION: New age, new habits

slide-4
SLIDE 4

Introduction : How about journalism?

Several research reported:

  • Reading attention span is getting shorter
  • Young generation is the least informed…
  • ...and more interested in social media

4

slide-5
SLIDE 5

Introduction : How about journalism?

Several research reported:

  • Reading attention span is getting shorter
  • Young generation is the least informed…
  • ...and more interested in social media

Information Pollution:

  • Reliable sources are more important than ever

5

slide-6
SLIDE 6

Introduction : Our proposal

Make the content:

  • Less time consuming
  • Yet still adequately informing

Solution: Automatic Summarization

6

slide-7
SLIDE 7

Introduction : Automatic Summarization for Journalism

“Journalism is the activity of gathering, assessing, creating, and presenting news and information.”

7

American Press Institute

slide-8
SLIDE 8

Introduction : Automatic Summarization for Journalism

“Journalism is the activity of gathering, assessing, creating, and presenting news and information.”

  • Whole
  • Extensive
  • Unbiased

8

slide-9
SLIDE 9

Introduction : Automatic Summarization for Journalism

“Journalism is the activity of gathering, assessing, creating, and presenting news and information.”

  • Whole
  • Extensive
  • Unbiased

Solution: Multi-document Summarization

9

slide-10
SLIDE 10

Introduction : Automatic Summarization for Journalism

“Journalism is the activity of gathering, assessing, creating, and presenting news and information.”

  • Extractive and Abstractive

10

slide-11
SLIDE 11

Introduction : Automatic Summarization for Journalism

“Journalism is the activity of gathering, assessing, creating, and presenting news and information.”

  • Extractive and Abstractive
  • Neural Abstractive Summarization

○ Methods are generally for Single-Document

11

slide-12
SLIDE 12

Introduction : Automatic Summarization for Journalism

“Journalism is the activity of gathering, assessing, creating, and presenting news and information.”

  • Extractive and Abstractive
  • Neural Abstractive Summarization

○ Methods are generally for Single-Document

  • Unified Model : Extractive + Abstractive

○ Content Selection ○ Multi-Document -> Single Document

12

slide-13
SLIDE 13

13

Dataset Unified Summarization Pipeline Experiments&Evaluation

slide-14
SLIDE 14

14

Dataset

slide-15
SLIDE 15

Dataset: What do we need?

Neural Abstractive:

  • Typically needs a dataset of thousands of documents
  • i.e. CNN/Dailymail > 90k/197k (single-document dataset)

15

slide-16
SLIDE 16

Dataset: What do we have?

  • Multi-Document datasets are typically small
  • One of the most well-known does not contain more than 60 cluster and

600 documents

16

Data Source Cluster/Sample Summaries Documents DUC 2001 30 309 DUC 2002 59 567 DUC 2004 50 500 Total 139 1,376

slide-17
SLIDE 17

Dataset: Solution

  • We created Webis-wikinews-corpus
  • One of the first of its kind...

○ Large-scale ○ Multi-document ○ For the news domain

17

slide-18
SLIDE 18

Dataset: Source

  • Wikimedia Projects : Wikinews & Wikipedia

○ Unbiased ○ Open-source ○ Up-to-date ○ Clustered news from reliable sources

18

slide-19
SLIDE 19

Dataset: Construction

Extract the useful information from Dump File:

  • Article, source links, auxiliary information
  • Only the pages with news sources for the Wikipedia

19

slide-20
SLIDE 20

Dataset: Construction

Retrieval:

20

slide-21
SLIDE 21

Dataset: Size & Folder Structure

21

Data Source Cluster/Sample Summaries Documents Wikinews 9,514 21,314 Wikipedia 2,174 17,807 Total 11,688 39,121

slide-22
SLIDE 22

22

Unified Summarization Pipeline

slide-23
SLIDE 23

Unified Summarization

  • Extractive Summarization: Wikisummarizer
  • Abstractive Summarization: Pointer-Generator Network [See et al., 2017]

23

slide-24
SLIDE 24

Unified Summarization

  • Extractive Summarization: Wikisummarizer

○ A Google Brain project [Liu et al. ,2018] : Extraction from similar source (Wikipedia)

  • Abstractive Summarization: Pointer-Generator Network [See et al., 2017]

24

slide-25
SLIDE 25

Unified Summarization

  • Extractive Summarization: Wikisummarizer

○ A Google Brain project [Liu et al. ,2018] : Extraction from similar source (Wikipedia) ○ CST: Filter out the duplication [Radev and Zhang, 2004]

  • Abstractive Summarization: Pointer-Generator Network [See et al., 2017]

25

slide-26
SLIDE 26

Unified Summarization

  • Extractive Summarization: Wikisummarizer

○ A Google Brain project [Liu et al. ,2018] : Extraction from similar source (Wikipedia) ○ CST: Filter out the duplication [Radev and Zhang, 2004]

  • Abstractive Summarization: Pointer-Generator Network [See et al., 2017]

26

slide-27
SLIDE 27

Unified Summarization

  • Extractive Summarization: Wikisummarizer

○ A Google Brain project [Liu et al. ,2018] : Extraction from similar source (Wikipedia) ○ CST: Filter out the duplication [Radev and Zhang, 2004]

  • Abstractive Summarization: Pointer-Generator Network [See et al., 2017]

○ Solves the problems of earlier approaches such as repetitiveness, senseless sentences and inaccurate facts

27

slide-28
SLIDE 28

28

Experiments&Evaluation

slide-29
SLIDE 29

Experiments and Evaluation: Training Models

29

  • Double-abstractive
  • Extractive + Abstractive Full Target
  • Extractive + Abstractive Short Target
slide-30
SLIDE 30

Experiments and Evaluation: Training Models

30

Double-abstractive

  • Trivial method
  • To examine the unified model
slide-31
SLIDE 31

Experiments and Evaluation: Training Models

31

Unified Models: Extractive + Abstractive

  • ea-full-target - Target document size : Full size
  • ea-short-target - Target document size : 3 sentences
  • To examine the effects of different ratio between

input and target

slide-32
SLIDE 32

Introduction : Automatic Summarization for Journalism

“Journalism is the activity of gathering, assessing, creating, and presenting news and information.”

32

slide-33
SLIDE 33

Experiments and Evaluation: Aspects

33

“Journalism is the activity of gathering, assessing, creating, and presenting news and information.”

  • Aspects :

○ Content ○ Readability

slide-34
SLIDE 34

Experiments and Evaluation: Aspects

34

  • Aspects :

○ Content ■ Automatic > a state-of-the-art method exist ○ Readability

slide-35
SLIDE 35

Experiments and Evaluation: ROUGE

35

Computer Generated Summary : the cat was found under the bed Ground-truth Summary : the cat was under the bed

slide-36
SLIDE 36

Experiments and Evaluation: ROUGE

36

Computer Generated Summary : the cat was found under the bed Ground-truth Summary : the cat was under the bed

slide-37
SLIDE 37

Experiments and Evaluation: ROUGE

37

Computer Generated Summary : the cat was found under the bed Ground-truth Summary : the cat was under the bed

slide-38
SLIDE 38

Experiments and Evaluation: ROUGE

38

Computer Generated Summary : the cat was found under the bed Ground-truth Summary : the cat was under the bed

slide-39
SLIDE 39

Experiments and Evaluation: ROUGE

39

  • ROUGE-N(ROUGE-1) : Overlapping n-grams > Word wise similarity
  • ROUGE-L : Longest Common Subsequence > Sequence wise similarity
slide-40
SLIDE 40

Experiments and Evaluation: Results

40

  • Aspects :

○ Content: ■ Automatic > a state-of-the-art method exist ROUGE double-abstractive ea-full-target ROUGE-1 0.23 0.29 ROUGE-L 0.16 0.21

slide-41
SLIDE 41

Experiments and Evaluation: Results

41

  • Aspects :

○ Content ■ Automatic > a state-of-the-art method exist ROUGE double-abstractive ea-full-target ea-short-target ROUGE-1 0.23 0.29 0.54 ROUGE-L 0.16 0.21 0.49

slide-42
SLIDE 42

Experiments and Evaluation: Aspects

42

  • Aspects :

○ Content ■ Automatic > a state-of-the-art method exist ○ Readability

slide-43
SLIDE 43

Experiments and Evaluation: ROUGE for readability?

43 Computer Generated Summary : was the found under the cat Ground-truth Summary : the cat was found under the bed 1 ROUGE-1 Average_R: 0.83333 1 ROUGE-1 Average_P: 0.83333 1 ROUGE-1 Average_F: 0.83333 1 ROUGE-L Average_R: 0.50000 1 ROUGE-L Average_P: 0.50000 1 ROUGE-L Average_F: 0.50000

slide-44
SLIDE 44

Experiments and Evaluation: ROUGE for readability?

44 Computer Generated Summary : he found no lights on Ground-truth Summary : all of the lamps were off already when he walked into the room 1 ROUGE-1 Average_R: 0.07692 1 ROUGE-1 Average_P: 0.20000 1 ROUGE-1 Average_F: 0.11111 1 ROUGE-L Average_R: 0.07692 1 ROUGE-L Average_P: 0.20000 1 ROUGE-L Average_F: 0.11111 Computer Generated Summary : was the found under the cat Ground-truth Summary : the cat was found under the bed 1 ROUGE-1 Average_R: 0.83333 1 ROUGE-1 Average_P: 0.83333 1 ROUGE-1 Average_F: 0.83333 1 ROUGE-L Average_R: 0.50000 1 ROUGE-L Average_P: 0.50000 1 ROUGE-L Average_F: 0.50000

slide-45
SLIDE 45

Experiments and Evaluation: Aspects

45

  • Aspects :

○ Content ■ Automatic > a state-of-the-art method exist ○ Readability ■ ROUGE is not reliable for readability ■ Manual > There are not many automatic methods, mostly manual

slide-46
SLIDE 46

Experiments and Evaluation: Readability Aspects by DUC

46

  • Grammaticality
  • Non-redundancy
  • Referential clarity
  • Focus
  • Structure and coherence
slide-47
SLIDE 47

Experiments and Evaluation: Survey

47

  • Grammaticality
  • Non-redundancy
  • Referential clarity
  • Focus
  • Structure and coherence

First Survey

slide-48
SLIDE 48

Experiments and Evaluation: Survey

48

  • Grammaticality
  • Non-redundancy
  • Referential clarity
  • Focus
  • Structure and coherence

First Survey

slide-49
SLIDE 49

Experiments and Evaluation: Survey

49

  • Grammaticality
  • Non-redundancy
  • Referential clarity
  • Focus
  • Structure and coherence

First Survey Second Survey

slide-50
SLIDE 50

Experiments and Evaluation: Results

50

  • Aspects :

○ Content: ■ Automatic > a state-of-the-art method exist ○ Readability ■ ROUGE is not reliable for readability ■ Manual > There are not many automatic methods, mostly manual Training Model Mean Score double-abstractive 2.15 ea-full-target 2.67

slide-51
SLIDE 51

Experiments and Evaluation: Results

51

  • Aspects :

○ Content: ■ Automatic > a state-of-the-art method exist ○ Readability ■ ROUGE is not reliable for readability ■ Manual > There are not many automatic methods, mostly manual Training Model Mean Score double-abstractive 2.15 ea-full-target 2.67 ea-short-target 4.18

slide-52
SLIDE 52

Experiments and Evaluation: Comparison of Evaluation Aspects

52

Training Model Mean Score double-abstractive 2.15 ea-full-target 2.67 ea-short-target 4.18

ROUGE double-abstractive ea-full-target ea-short-target ROUGE-1 0.23 0.29 0.54 ROUGE-L 0.16 0.21 0.49

slide-53
SLIDE 53

Observations

53

slide-54
SLIDE 54

Observations : UNK Error

54 Ground-truth Summary: yesterday san francisco giants lf barry bonds hit a 435-foot home run , his 756th , off a pitch from mike bacsick of the washington nationals , breaking the all-time career home run record , formerly held by hank aaron.the pitch , the seventh of the at-bat , was a 3-2 pitch , which bonds hit into the right-center field bleachers.matt murphy , a 22-year-old from queens in new york city , got the ball and was promptly protected and escorted away from the mayhem by a group of san francisco police officers . Computer Generated Summary: yesterday san francisco giants [UNK] barry bonds hit a [UNK]home run , his 756th , off a pitch from mike [UNK] of the washington nationals , breaking the [UNK] home run in 1974 .

slide-55
SLIDE 55

Observations : Inaccurate Facts

55 Input … Hello Kitty was first introduced by Japanese company Sanrio in 1974.The cute round-faced ... Ground-truth summary: the armband , which features "hello kitty" sitting on top of two hearts , will be worn by police officers who commit minor offences.these include , and parking in a prohibited area.the officers will also be forced to stay with the deputy chief all day in division office and will be forbidden to disclose their offences. Computer-generated summary: the armband , which features "hello kitty" sitting on top of two hearts , will be worn by police officers who commit minor [UNK] include , and parking in a prohibited area.the officers will also be forced to stay with the deputy police in 1974.

slide-56
SLIDE 56

Observations : Repetition

56 Computer-generated summary: a court in , kenya has sentenced a group of seven somali pirates to five years each in jail , according to a statement by the european [UNK] mission eu [UNK] said that the men , “ i have concrete proof that you attacked a vessel in the high seas and i order you to serve five years in prison seas and i order you to serve five years in prison seas and i

  • rder you to serve five years in prison seas and i order you to serve five years in prison

seas and i order you to serve

slide-57
SLIDE 57

Observations : Successfvl Summaries

57 Ground-truth summary: At least ten attackers with knives, dressed in black, attacked a train station in , China yesterday.At least 28 victims were killed, with 113 more wounded by knives, Chinese state news agency reported.The local municipal government accuses " separatist forces" for the attack. Computer-generated summary: At least ten attackers with knives, dressed in black, attacked a train station in , China yesterday.At least 28 victims were killed, with 113 more wounded by knives,Chinese state news agency reported.The local municipal government accuses " separatist forces" for the attack. Ground-truth summary: the united states navy has successfully destroyed a crippled spy satellite in a decaying orbit , by intercepting it with a missile.a modified sm-3 missile was launched from the uss lake erie at 03:26 gmt this morning , and intercepted the usa-193 satellite around three minutes later.it has been reported that the satellite has broken into around 80 pieces , some of which have already re-entered the earth ’s atmosphere . Computer-generated summary: the united states navy has successfully destroyed a crippled spy satellite in a decaying orbit , by intercepting it with a missile.a modified sm-3 missile was launched from the uss lake erie at 03:26 gmt this morning , and intercepted the usa-193 satellite around three minutes later.it has been reported that the satellite has been damaged.

slide-58
SLIDE 58

Recap, Conclusion & Future Work

58

  • Contributions

○ One of the first large-scale multi-document summarization dataset for news domain

slide-59
SLIDE 59

Recap, Conclusion & Future Work

59

  • Contributions

○ One of the first large-scale multi-document summarization dataset for news domain ○ Wikisummarizer

slide-60
SLIDE 60

Recap, Conclusion & Future Work

60

  • Contributions

○ One of the first large-scale multi-document summarization dataset for news domain ○ Wikisummarizer ○ Unified multi-document summarization pipeline

slide-61
SLIDE 61

Recap, Conclusion & Future Work

61

  • Contributions

○ One of the first large-scale multi-document summarization dataset for news domain ○ Wikisummarizer ○ Unified multi-document summarization pipeline

  • Conclusion
slide-62
SLIDE 62

Recap, Conclusion & Future Work

62

  • Contributions

○ One of the first large-scale multi-document summarization dataset for news domain ○ Wikisummarizer ○ Unified multi-document summarization pipeline

  • Conclusion

○ Extractive Summarization proved to be a good method to transfer from multi-document to single-document

slide-63
SLIDE 63

Recap, Conclusion & Future Work

63

  • Contributions

○ One of the first large-scale multi-document summarization dataset for news domain ○ Wikisummarizer ○ Unified multi-document summarization pipeline

  • Conclusion

○ Extractive Summarization proved to be a good method to transfer from multi-document to single-document ○ Better content selection resulted in better readability

slide-64
SLIDE 64

Recap, Conclusion & Future Work

64

  • Contributions

○ One of the first large-scale multi-document summarization dataset for news domain ○ Wikisummarizer ○ Unified multi-document summarization pipeline

  • Conclusion

○ Extractive Summarization proved to be a good method to transfer from multi-document to single-document ○ Better content selection resulted in better readability ○ Even though there is a room for improvement, the idea behind the framework is promising

slide-65
SLIDE 65

Recap, Conclusion & Future Work

65

  • Contributions

○ One of the first large-scale multi-document summarization dataset for news domain ○ Wikisummarizer ○ Unified multi-document summarization pipeline

  • Conclusion

○ Extractive Summarization proved to be a good method to transfer from multi-document to single-document ○ Better content selection resulted in better readability ○ Even though there is a room for improvement, the idea behind the framework is promising

  • Future Work
slide-66
SLIDE 66

Recap, Conclusion & Future Work

66

  • Contributions

○ One of the first large-scale multi-document summarization dataset for news domain ○ Wikisummarizer ○ Unified multi-document summarization pipeline

  • Conclusion

○ Extractive Summarization proved to be a good method to transfer from multi-document to single-document ○ Better content selection resulted in better readability ○ Even though there is a room for improvement, the idea behind the framework is promising

  • Future Work

○ Extending the dataset

slide-67
SLIDE 67

Recap, Conclusion & Future Work

67

  • Contributions

○ One of the first large-scale multi-document summarization dataset for news domain ○ Wikisummarizer ○ Unified multi-document summarization pipeline

  • Conclusion

○ Extractive Summarization proved to be a good method to transfer from multi-document to single-document ○ Better content selection resulted in better readability ○ Even though there is a room for improvement, the idea behind the framework is promising

  • Future Work

○ Extending the dataset ■ Distant Supervision for clustering other datasets

slide-68
SLIDE 68

Recap, Conclusion & Future Work

68

  • Contributions

○ One of the first large-scale multi-document summarization dataset for news domain ○ Wikisummarizer ○ Unified multi-document summarization pipeline

  • Conclusion

○ Extractive Summarization proved to be a good method to transfer from multi-document to single-document ○ Better content selection resulted in better readability ○ Even though there is a room for improvement, the idea behind the framework is promising

  • Future Work

○ Extending the dataset ■ Distant Supervision for clustering other datasets ■ Merge with Multi-News

slide-69
SLIDE 69

Recap, Conclusion & Future Work

69

  • Contributions

○ One of the first large-scale multi-document summarization dataset for news domain ○ Wikisummarizer ○ Unified multi-document summarization pipeline

  • Conclusion

○ Extractive Summarization proved to be a good method to transfer from multi-document to single-document ○ Better content selection resulted in better readability ○ Even though there is a room for improvement, the idea behind the framework is promising

  • Future Work

○ Extending the dataset ■ Distant Supervision for clustering other datasets ■ Merge with Multi-News ○ Other setups of Wikisummarizer

slide-70
SLIDE 70

Thanks for coming...

70

slide-71
SLIDE 71

References

[Liu et al. ,2018] : Peter J. Liu, Mohammad Saleh, Etienne Pot, Ben Goodrich, Ryan Sepassi,Lukasz Kaiser, and Noam Shazeer. Generating wikipedia by summarizing long sequences. arXiv:1801.10198 [cs], 2018. URL http://arxiv.org/abs/1801.10198 [Radev and Zhang, 2004] : Dragomir R. Radev and Zhu Zhang. Cross-document relationship classification for text summarization. 2004. [See et al., 2017] : Abigail See, Peter J. Liu, and Christopher D. Manning. Get to the point: Summarization with pointer-generator networks. In ACL, 2017. 71

slide-72
SLIDE 72

Image references https://img.buzzfeed.com/buzzfeed-static/static/enhanced/webdr05/2013/6/28/10/enhanced-buzz-29020-137243101 4-2.jpg?downsize=800:*&output-format=auto&output-quality=auto https://tr.pinterest.com/pin/451345193878807678/?lp=true https://images.sadhguru.org/sites/default/files/media_files/iso/en/48257-confusion-clarity-spiritual-path.jpg https://www.timeshighereducation.com/sites/default/files/styles/the_breaking_news_image_style/public/Pictures/web /n/c/o/numbers_on_podium.jpg?itok=-nVlhkPx

72