Daizong Ding 1 Mi Zhang 1 Xudong Pan 1 Min - - PowerPoint PPT Presentation

daizong ding 1 mi zhang 1 xudong pan 1 min yang 1
SMART_READER_LITE
LIVE PREVIEW

Daizong Ding 1 Mi Zhang 1 Xudong Pan 1 Min - - PowerPoint PPT Presentation

2019 2019 Daizong Ding 1 Mi Zhang 1 Xudong Pan 1 Min Yang 1 Xiangnan He 2 1. School of Computer Science, Fudan University 2. School of Data Science, University of Science and Technology of


slide-1
SLIDE 1

我们毕业啦

其 实 是 答 辩 的 标 题 地 方

Daizong Ding1 Mi Zhang1 Xudong Pan1 Min Yang1 Xiangnan He2

  • 1. School of Computer Science, Fudan University
  • 2. School of Data Science, University of Science and Technology of China

2019 2019

slide-2
SLIDE 2

Background Problem Analysis Proposed Model Extreme Value Loss

Time Series Prediction

Experiments Conclusion

?

Training Inputs: 𝑌1:𝑈 = 𝑦1, ⋯ , 𝑦𝑈 Labels: 𝑍

1:𝑈 = 𝑧1, ⋯ , 𝑧𝑈

Outputs: 𝑃1:𝑈 = 𝑝1, ⋯ , 𝑝𝑈 Goal: min σ𝑢=1

𝑈

𝑝𝑢 − 𝑧𝑢 2 Testing Inputs: 𝑌1:𝑈+𝐿 = 𝑦1, ⋯ , 𝑦𝑈, 𝑦𝑈+1, ⋯ , 𝑦𝑈+𝐿 Outputs: 𝑃1:𝑈+𝐿 = 𝑝1, ⋯ , 𝑝𝑈, 𝑝𝑈+1, ⋯ , 𝑝𝑈+𝐿 Length=𝑈 Length=𝐿

slide-3
SLIDE 3

Background Problem Analysis Proposed Model Extreme Value Loss

Recurrent Neural Network

Experiments Conclusion

Training For 𝑢 = 1, ⋯ , 𝑈: ℎ𝑢 = 𝐻𝑆𝑉 𝑦1, ⋯ , 𝑦𝑢 𝑝𝑢 = 𝑋

𝑝 𝑈ℎ𝑢 + 𝑐𝑝

min σ𝑢=1

𝑈

𝑝𝑢 − 𝑧𝑢 2 Testing For 𝑢 = 1, ⋯ , 𝑈 + 𝐿: ℎ𝑢 = 𝐻𝑆𝑉 𝑦1, ⋯ , 𝑦𝑢 𝑝𝑢 = 𝑋

𝑝 𝑈ℎ𝑢 + 𝑐𝑝

𝑦1 𝑝1 𝑦2 𝑝2 𝑦𝑈+𝐿 𝑝𝑈+𝐿 𝑦𝑈 𝑝𝑈 𝑧1 𝑧2 𝑧𝑈 Train Test Results GRU GRU GRU GRU FC FC FC FC … … … … … … ℎ1 ℎ2 ℎ𝑈+𝐿 ℎ𝑈

slide-4
SLIDE 4

Background Problem Analysis Proposed Model Extreme Value Loss

Underfitting Phenomenon

Experiments Conclusion

slide-5
SLIDE 5

Background Problem Analysis Proposed Model Extreme Value Loss

Overfitting Phenomenon

Experiments Conclusion

slide-6
SLIDE 6

Extreme Events in Time Series Data

Characteristic

  • Extremely small or large values
  • Irregular
  • Rare occurrences
  • Light-tailed distributions (Gaussian, Poisson, etc.) cannot

model them well Problem

  • Why Deep Neural Network could suffer extreme event

problem in time series prediction?

  • How can we improve the performance on the prediction
  • f extreme events?

Background Problem Analysis Proposed Model Extreme Value Loss Experiments Conclusion

slide-7
SLIDE 7

Background Problem Analysis Proposed Model Extreme Value Loss

Estimated Distribution of Labels 𝒛𝒖

Experiments Conclusion

  • With Bayes Theorem,

𝑄 𝑍 𝑌, 𝜄 = 𝑄 𝑌 𝑍, 𝜄 𝑄 𝑍 𝑄 𝑌|𝜄 Likelihood Posterior Estimated distribution of labels ෠ 𝑄 𝑍 = 1

𝑈 σ𝑢=1 𝑈

𝒪(𝑧𝑢, Ƹ 𝜐2)

  • DNN will internally estimate the distribution of 𝑧𝑢 according to the sampled data.
  • The optimization of deep neural network under probability perspective:

min σt=1

𝑈

𝑝𝑢 − 𝑧𝑢 2 max ς𝑢=1

𝑈

𝒪 𝑧𝑢 𝑝𝑢, Ƹ 𝜐2 ⟺ max ς𝑢=1

𝑈

𝑄 𝑧𝑢 𝑦𝑢, 𝜄

Bregman Divergence

slide-8
SLIDE 8

Extreme Event Problem in DNN

Underfitting Phenomenon

  • For those normal points, e.g., 𝑧1,

𝑄 𝑧1 𝑌, 𝜄 = 𝑄 𝑌 𝑧1, 𝜄 ෠ 𝑄 𝑧1 𝑄 𝑌, 𝜄 ≥ 𝑄 𝑌 𝑧1, 𝜄 𝑄𝑢𝑠𝑣𝑓 𝑧1 𝑄 𝑌, 𝜄 = 𝑄𝑢𝑠𝑣𝑓 𝑧1 𝑌, 𝜄

  • For those rarely occurred extreme events, e.g., 𝑧2,

𝑄 𝑧2 𝑌, 𝜄 = 𝑄 𝑌 𝑧2, 𝜄 ෠ 𝑄 𝑧2 𝑄 𝑌, 𝜄 ≤ 𝑄 𝑌 𝑧2, 𝜄 𝑄𝑢𝑠𝑣𝑓 𝑧2 𝑄 𝑌, 𝜄 = 𝑄𝑢𝑠𝑣𝑓 𝑧2 𝑌, 𝜄

  • Therefore model commonly lacks the ability of predicting extreme events

𝑧2 𝑧1 𝑧3

Background Problem Analysis Proposed Model Extreme Value Loss Experiments Conclusion

slide-9
SLIDE 9

Extreme Event Problem in DNN

Overfitting Phenomenon

  • If we add weights of extreme events during the training
  • For those normal points, e.g., 𝑧1,

𝑄 𝑧1 𝑌, 𝜄 = 𝑄 𝑌 𝑧1, 𝜄 ෠ 𝑄 𝑧1 𝑄 𝑌, 𝜄 ≤ 𝑄 𝑌 𝑧1, 𝜄 𝑄𝑢𝑠𝑣𝑓 𝑧1 𝑄 𝑌, 𝜄 = 𝑄𝑢𝑠𝑣𝑓 𝑧1 𝑌, 𝜄

  • For those rarely occurred extreme events, e.g., 𝑧3,

𝑄 𝑧3 𝑌, 𝜄 = 𝑄 𝑌 𝑧3, 𝜄 ෠ 𝑄 𝑧3 𝑄 𝑌, 𝜄 ≥ 𝑄 𝑌 𝑧3, 𝜄 𝑄𝑢𝑠𝑣𝑓 𝑧3 𝑄 𝑌, 𝜄 = 𝑄𝑢𝑠𝑣𝑓 𝑧3 𝑌, 𝜄

  • The estimated distribution is not accurate
  • The performance on test data is poor

𝑧2 𝑧1 𝑧3

Background Problem Analysis Proposed Model Extreme Value Loss Experiments Conclusion

slide-10
SLIDE 10

Problem Analysis

Extreme Event Problem in DNN mainly because:

  • Extreme events are extremely large or small values with

rare occurrence. Therefore it is hard to estimate the true distribution of them given limited samples.

  • Usually DNN learns time series data from light-tailed

likelihood, which further increases the difficulty of estimating the distribution of extreme events.

Background Problem Analysis Proposed Model Extreme Value Loss Experiments Conclusion

𝑧2 𝑧1 𝑧3

slide-11
SLIDE 11

Motivation: Find the regularity inside irregular extreme events

According to previous research:

  • Extreme events in time-series data often show some

form of temporal regularity.

  • Randomness of extreme events have limited degrees of

freedom (DOF). The pattern of extreme events after a window could be memorized !

Background Proposed Model Extreme Value Loss Experiments Conclusion Problem Analysis

S&P 500

slide-12
SLIDE 12

Recalling Extreme Events in History

We propose to use Memory Network to recall extreme events in history:

  • For each time step 𝑢, we sample 𝑁 windows.
  • For window 𝑘 , we propose to use GRU to calculate

the feature 𝑡

𝑘 of the window.

  • Meanwhile, we also record the occurrence of

extreme events 𝑟𝑘 = −1,0,1 by setting threshold previously at the next time step of window 𝑘.

Background Proposed Model Extreme Value Loss Experiments Conclusion Problem Analysis

Memory Module

slide-13
SLIDE 13

Attention Mechanism

We propose to use attention to incorporate memory module with the prediction:

  • At time 𝑢, we first calculate the output from GRU:
  • Then we construct the memory module, and calculate

the similarity between the current and the history:

  • The final output from our model is,

Background Proposed Model Extreme Value Loss Experiments Conclusion Problem Analysis

slide-14
SLIDE 14

Extreme Value Theory

If we still use Gaussian likelihood, the improved model still suffer extreme event problem:

  • We should use a heavy-tailed likelihood to fit the

distribution of extreme events given limited samples. It is hard to predict the values of extreme events, however, the DOF of extreme events are easier to be modelled.

  • We could propose a heavy-tailed likelihood for predicting

the occurrence of extreme events.

Background Proposed Model Extreme Value Loss Experiments Conclusion Problem Analysis

slide-15
SLIDE 15

Extreme Value Loss

Background Proposed Model Extreme Value Loss Experiments Conclusion Problem Analysis

  • Through Extreme Value Theory (EVT), the approximation of 𝑧𝑢 from EVT can be written as,
  • 𝑤𝑢 = 0,1 is the indicator of whether a large value will happen or not.
  • If we pay our attention to predict whether there is an extremely large value at 𝑢 by outputting 𝑣𝑢 = 0,1 ,

we can add the weights of extreme events on binary cross entropy loss:

  • It is easy to extend the binary classification to ut, vt = −1,0,1 .

Scale function Binary cross entropy loss

slide-16
SLIDE 16

Optimization

Background Proposed Model Extreme Value Loss Experiments Conclusion Problem Analysis

The final loss function can be written as: For the two challenges in DNN:

  • We predict the labels from both GRU and memory

module, which memorizes the regularity inside extreme events given limited samples.

  • We propose to minimize a heavy-tailed classification

loss (EVL) for detecting the occurrence of extreme events.

𝑧𝑢 𝑤𝑢 EVL Square Loss

= −1,0,1

slide-17
SLIDE 17

Background Proposed Model Extreme Value Loss Experiments Conclusion Problem Analysis

Experimental Settings

  • Dataset:
  • Stock Dataset: 564 corporations in Nasdaq Stock Market with one sample per week
  • Climate Dataset: Green Gas Observing Network dataset and Atmospheric Co2 Dataset
  • Pseudo Periodic Synthetic Dataset
  • Baselines:
  • LSTM
  • GRU
  • Time-LSTM
  • Research questions:
  • RQ1: Is our proposed framework effective in time series prediction?
  • RQ2: Is our proposed loss function EVL worked in detecting extreme events?
  • RQ3: What is the influence of hyper-parameters in the framework?
slide-18
SLIDE 18

Background Proposed Model Extreme Value Loss Experiments Conclusion Problem Analysis

Time Series Prediction (RMSE)

slide-19
SLIDE 19

Background Proposed Model Extreme Value Loss Experiments Conclusion Problem Analysis

Time Series Prediction (Visualization)

slide-20
SLIDE 20

Background Proposed Model Extreme Value Loss Experiments Conclusion Problem Analysis

Extreme Events Prediction (F1 Value)

slide-21
SLIDE 21

Background Proposed Model Extreme Value Loss Experiments Conclusion Problem Analysis

Influence of hyper-parameters

slide-22
SLIDE 22

Background Proposed Model Extreme Value Loss Experiments Conclusion Problem Analysis

Conclusion

  • In this paper, we pay our attention to extreme events in time series data.
  • We first analysis why DNN is innately weak in predicting extreme events:
  • Extreme events are extremely large or small values with rare occurrence, it is hard to model them with

limited samples.

  • The commonly used Gaussian likelihood is a light-tailed distribution.
  • We further propose a framework to improve the performance on time series prediction:
  • For the first challenge, we propose to use Memory Network to recall extreme events in history.
  • For the second challenge, we propose a new loss function called extreme value loss (EVL).
  • Empirical results show the effectiveness of our proposed framework.
slide-23
SLIDE 23

Background Proposed Model Extreme Value Loss Experiments Conclusion Problem Analysis

Future Work

  • Extending our work to multi-dimensional time series data.
  • Applying EVL to more kinds of tasks.
  • …………
slide-24
SLIDE 24

Background Proposed Model Extreme Value Loss Experiments Conclusion Problem Analysis

Thank you for listening!

If you have any questions, please contact Daizong Ding Email: 17110240010@fudan.edu.cn Wechat: