PP: A dual-stage attention-based recurrent neural network for time series prediction

Problem: time series prediction

The nonlinear autoregressive exogenous model: The Nonlinear autoregressive exogenous (NARX) model, which predicts the current value of a time series based upon its previous values as well as the current and past values of multiple driving (exogenous) series.

However, few NARX models can capture the long-term temporal dependencies appropriately and select the relevant driving series to make a prediction.

2 issues:

1. capture the long-term temporal dependencies

2. select the relevant driving series to make a prediction

We propose a dual-stage attention-based RNN to address these 2 issues. 

1. first stage: input attention mechanism to extract relevant driving series. 

2. second stage: temporal attention mechanism. 

attention-based encoder-decoder networks for time series prediction/ LSTM/ GRU

One problem with encoder-decoder networks is that their performance will deteriorate rapidly as the length of input sequence increases. 

Contribution: the two-stage attention mechanism. input attention for driving series and temporal attention for all time stamps. 

input attention can select the relevant driving series.

temporal attention capture temporal information. 

Supplementary knowledge:

1. what is driving series?

 

posted @ 2020-02-05 22:22  keeps_you_warm  阅读(377)  评论(0编辑  收藏  举报