我们毕业啦
其 实 是 答 辩 的 标 题 地 方
Daizong Ding1 Mi Zhang1 Xudong Pan1 Min Yang1 Xiangnan He2
- 1. School of Computer Science, Fudan University
- 2. School of Data Science, University of Science and Technology of China
Daizong Ding 1 Mi Zhang 1 Xudong Pan 1 Min - - PowerPoint PPT Presentation
2019 2019 Daizong Ding 1 Mi Zhang 1 Xudong Pan 1 Min Yang 1 Xiangnan He 2 1. School of Computer Science, Fudan University 2. School of Data Science, University of Science and Technology of
其 实 是 答 辩 的 标 题 地 方
Daizong Ding1 Mi Zhang1 Xudong Pan1 Min Yang1 Xiangnan He2
Background Problem Analysis Proposed Model Extreme Value Loss
Experiments Conclusion
?
Training Inputs: 𝑌1:𝑈 = 𝑦1, ⋯ , 𝑦𝑈 Labels: 𝑍
1:𝑈 = 𝑧1, ⋯ , 𝑧𝑈
Outputs: 𝑃1:𝑈 = 𝑝1, ⋯ , 𝑝𝑈 Goal: min σ𝑢=1
𝑈
𝑝𝑢 − 𝑧𝑢 2 Testing Inputs: 𝑌1:𝑈+𝐿 = 𝑦1, ⋯ , 𝑦𝑈, 𝑦𝑈+1, ⋯ , 𝑦𝑈+𝐿 Outputs: 𝑃1:𝑈+𝐿 = 𝑝1, ⋯ , 𝑝𝑈, 𝑝𝑈+1, ⋯ , 𝑝𝑈+𝐿 Length=𝑈 Length=𝐿
Background Problem Analysis Proposed Model Extreme Value Loss
Experiments Conclusion
Training For 𝑢 = 1, ⋯ , 𝑈: ℎ𝑢 = 𝐻𝑆𝑉 𝑦1, ⋯ , 𝑦𝑢 𝑝𝑢 = 𝑋
𝑝 𝑈ℎ𝑢 + 𝑐𝑝
min σ𝑢=1
𝑈
𝑝𝑢 − 𝑧𝑢 2 Testing For 𝑢 = 1, ⋯ , 𝑈 + 𝐿: ℎ𝑢 = 𝐻𝑆𝑉 𝑦1, ⋯ , 𝑦𝑢 𝑝𝑢 = 𝑋
𝑝 𝑈ℎ𝑢 + 𝑐𝑝
𝑦1 𝑝1 𝑦2 𝑝2 𝑦𝑈+𝐿 𝑝𝑈+𝐿 𝑦𝑈 𝑝𝑈 𝑧1 𝑧2 𝑧𝑈 Train Test Results GRU GRU GRU GRU FC FC FC FC … … … … … … ℎ1 ℎ2 ℎ𝑈+𝐿 ℎ𝑈
Background Problem Analysis Proposed Model Extreme Value Loss
Experiments Conclusion
Background Problem Analysis Proposed Model Extreme Value Loss
Experiments Conclusion
Characteristic
model them well Problem
problem in time series prediction?
Background Problem Analysis Proposed Model Extreme Value Loss Experiments Conclusion
Background Problem Analysis Proposed Model Extreme Value Loss
Experiments Conclusion
𝑄 𝑍 𝑌, 𝜄 = 𝑄 𝑌 𝑍, 𝜄 𝑄 𝑍 𝑄 𝑌|𝜄 Likelihood Posterior Estimated distribution of labels 𝑄 𝑍 = 1
𝑈 σ𝑢=1 𝑈
𝒪(𝑧𝑢, Ƹ 𝜐2)
min σt=1
𝑈
𝑝𝑢 − 𝑧𝑢 2 max ς𝑢=1
𝑈
𝒪 𝑧𝑢 𝑝𝑢, Ƹ 𝜐2 ⟺ max ς𝑢=1
𝑈
𝑄 𝑧𝑢 𝑦𝑢, 𝜄
Bregman Divergence
Underfitting Phenomenon
𝑄 𝑧1 𝑌, 𝜄 = 𝑄 𝑌 𝑧1, 𝜄 𝑄 𝑧1 𝑄 𝑌, 𝜄 ≥ 𝑄 𝑌 𝑧1, 𝜄 𝑄𝑢𝑠𝑣𝑓 𝑧1 𝑄 𝑌, 𝜄 = 𝑄𝑢𝑠𝑣𝑓 𝑧1 𝑌, 𝜄
𝑄 𝑧2 𝑌, 𝜄 = 𝑄 𝑌 𝑧2, 𝜄 𝑄 𝑧2 𝑄 𝑌, 𝜄 ≤ 𝑄 𝑌 𝑧2, 𝜄 𝑄𝑢𝑠𝑣𝑓 𝑧2 𝑄 𝑌, 𝜄 = 𝑄𝑢𝑠𝑣𝑓 𝑧2 𝑌, 𝜄
𝑧2 𝑧1 𝑧3
Background Problem Analysis Proposed Model Extreme Value Loss Experiments Conclusion
Overfitting Phenomenon
𝑄 𝑧1 𝑌, 𝜄 = 𝑄 𝑌 𝑧1, 𝜄 𝑄 𝑧1 𝑄 𝑌, 𝜄 ≤ 𝑄 𝑌 𝑧1, 𝜄 𝑄𝑢𝑠𝑣𝑓 𝑧1 𝑄 𝑌, 𝜄 = 𝑄𝑢𝑠𝑣𝑓 𝑧1 𝑌, 𝜄
𝑄 𝑧3 𝑌, 𝜄 = 𝑄 𝑌 𝑧3, 𝜄 𝑄 𝑧3 𝑄 𝑌, 𝜄 ≥ 𝑄 𝑌 𝑧3, 𝜄 𝑄𝑢𝑠𝑣𝑓 𝑧3 𝑄 𝑌, 𝜄 = 𝑄𝑢𝑠𝑣𝑓 𝑧3 𝑌, 𝜄
𝑧2 𝑧1 𝑧3
Background Problem Analysis Proposed Model Extreme Value Loss Experiments Conclusion
Extreme Event Problem in DNN mainly because:
rare occurrence. Therefore it is hard to estimate the true distribution of them given limited samples.
likelihood, which further increases the difficulty of estimating the distribution of extreme events.
Background Problem Analysis Proposed Model Extreme Value Loss Experiments Conclusion
𝑧2 𝑧1 𝑧3
According to previous research:
form of temporal regularity.
freedom (DOF). The pattern of extreme events after a window could be memorized !
Background Proposed Model Extreme Value Loss Experiments Conclusion Problem Analysis
S&P 500
We propose to use Memory Network to recall extreme events in history:
the feature 𝑡
𝑘 of the window.
extreme events 𝑟𝑘 = −1,0,1 by setting threshold previously at the next time step of window 𝑘.
Background Proposed Model Extreme Value Loss Experiments Conclusion Problem Analysis
Memory Module
We propose to use attention to incorporate memory module with the prediction:
the similarity between the current and the history:
Background Proposed Model Extreme Value Loss Experiments Conclusion Problem Analysis
If we still use Gaussian likelihood, the improved model still suffer extreme event problem:
distribution of extreme events given limited samples. It is hard to predict the values of extreme events, however, the DOF of extreme events are easier to be modelled.
the occurrence of extreme events.
Background Proposed Model Extreme Value Loss Experiments Conclusion Problem Analysis
Background Proposed Model Extreme Value Loss Experiments Conclusion Problem Analysis
we can add the weights of extreme events on binary cross entropy loss:
Scale function Binary cross entropy loss
Background Proposed Model Extreme Value Loss Experiments Conclusion Problem Analysis
The final loss function can be written as: For the two challenges in DNN:
module, which memorizes the regularity inside extreme events given limited samples.
loss (EVL) for detecting the occurrence of extreme events.
𝑧𝑢 𝑤𝑢 EVL Square Loss
= −1,0,1
Background Proposed Model Extreme Value Loss Experiments Conclusion Problem Analysis
Background Proposed Model Extreme Value Loss Experiments Conclusion Problem Analysis
Background Proposed Model Extreme Value Loss Experiments Conclusion Problem Analysis
Background Proposed Model Extreme Value Loss Experiments Conclusion Problem Analysis
Background Proposed Model Extreme Value Loss Experiments Conclusion Problem Analysis
Background Proposed Model Extreme Value Loss Experiments Conclusion Problem Analysis
limited samples.
Background Proposed Model Extreme Value Loss Experiments Conclusion Problem Analysis
Background Proposed Model Extreme Value Loss Experiments Conclusion Problem Analysis
If you have any questions, please contact Daizong Ding Email: 17110240010@fudan.edu.cn Wechat: