skip to main content
10.1145/3394486.3403046acmconferencesArticle/Chapter ViewAbstractPublication PageskddConference Proceedingsconference-collections
research-article
Open access

Preserving Dynamic Attention for Long-Term Spatial-Temporal Prediction

Published: 20 August 2020 Publication History

Abstract

Effective long-term predictions have been increasingly demanded in urban-wise data mining systems. Many practical applications, such as accident prevention and resource pre-allocation, require an extended period for preparation. However, challenges come as long-term prediction is highly error-sensitive, which becomes more critical when predicting urban-wise phenomena with complicated and dynamic spatial-temporal correlation. Specifically, since the amount of valuable correlation is limited, enormous irrelevant features introduce noises that trigger increased prediction errors. Besides, after each time step, the errors can traverse through the correlations and reach the spatial-temporal positions in every future prediction, leading to significant error propagation. To address these issues, we propose a Dynamic Switch-Attention Network (DSAN) with a novel Multi-Space Attention (MSA) mechanism that measures the correlations between inputs and outputs explicitly. To filter out irrelevant noises and alleviate the error propagation, DSAN dynamically extracts valuable information by applying self-attention over the noisy input and bridges each output directly to the purified inputs via implementing a switch-attention mechanism. Through extensive experiments on two spatial-temporal prediction tasks, we demonstrate the superior advantage of DSAN in both short-term and long-term predictions. The source code can be obtained from https://github.com/hxstarklin/DSAN.

Supplementary Material

MP4 File (3394486.3403046.mp4)
Effective long-term predictions have been increasingly demanded in urban-wise data mining systems, as many practical applications require an extended period for preparation. However, challenges come as long-term prediction is highly error-sensitive, which becomes more critical when predicting urban-wise phenomena with complicated and dynamic spatial-temporal correlation. To achieve reliable long-term prediction, we propose a Dynamic Switch-Attention Network (DSAN) with a novel Multi-Space Attention (MSA) mechanism that measures the correlations between inputs and outputs explicitly. To filter out irrelevant noises and alleviate the error propagation, DSAN dynamically extracts valuable information by applying self-attention over the noisy input and bridges each output directly to the purified inputs via implementing a switch-attention mechanism.

References

[1]
D. Bahdanau, K. Cho, and Y. Bengio. 2015. Neural Machine Translation by Jointly Learning to Align and Translate. In ICLR.
[2]
J. Bruna, W. Zaremba, A. Szlam, and Y. Lecun. 2013. Spectral Networks and Locally Connected Networks on Graphs. In ICLR.
[3]
J. Chung, C. Gulcehre, K. Cho, and Y. Bengio. 2014. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling.arXiv preprint arXiv:1412.3555(2014).
[4]
Z. Cui, R. Ke, and Y. Wang. 2016. Deep Bidirectional and Unidirectional LSTM Recurrent Neural Network for Network-wide Traffic Speed Prediction. arXiv preprint arXiv:1801.02143(2016).
[5]
Z. Dai, Z. Yang, Y. Yang, J. Carbonell, Q. Le, and R. Salakhutdinov. 2019. Transformer-XL: Attentive Language Models beyond a Fixed-Length Context. In ACL. 2978--2988.
[6]
J. Devlin, M. Chang, K. Lee, and K. Toutanova. 2018. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arXiv preprint arXiv:1810.04805(2018).
[7]
S. Fang, Q. Zhang, G. Meng, S. Xiang, and C. Pan. 2019. GSTNet: Global Spatial-Temporal Network for Traffic Flow Prediction. In IJCAI. 2286--2293.
[8]
X. Geng, Y. Li, L. Wang, L. Zhang, Q. Yang, J. Ye, and Y. Liu. 2019. Spatiotemporal Multi-Graph Convolution Network for Ride-Hailing Demand Forecasting. In AAAI. 3656--3663.
[9]
K. He, X. Zhang, S. Ren, and J. Sun. 2016. Deep Residual Learning for Image Recognition. In IEEE CVPR. 770--778.
[10]
M. Henaff, J. Bruna, and Y. Lecun. 2015. Deep Convolutional Networks on Graph-Structured Data. arXiv preprint arXiv:1506.05163(2015).
[11]
S. Hochreiter and J. Schmidhuber. 1997. Long Short-Term Memory. Neural Computation 9, 8 (1997), 1735--1780.
[12]
J. Ke, H. Zheng, H. Yang, and X. Chen. 2017. Short-Term Forecasting of Passenger Demand under On-Demand Ride Services: A Spatio-Temporal Deep Learning Approach. arXiv preprint arXiv:1706.06279(2017).
[13]
Y. Kim, C. Denton, L. Hoang, and A. M. Rush. 2017. Structured Attention Networks. In ICLR.
[14]
J. Li, Z. Han, H. Cheng, J. Su, P. Wang, J. Zhang, and L. Pan. 2019. Predicting Path Failure In Time-Evolving Graphs. In ACM SIGKDD. 1279--1289.
[15]
Y. Li, R. Yu, C. Shahabi, and Y. Liu. 2017. Graph Convolutional Recurrent Neural Network: Data-Driven Traffic Forecasting.arXiv preprint arXiv:1707.01926(2017).
[16]
Z. Lin, M. Feng, C. N. dos Santos, M. Yu, B. Xiang, B. Zhou, and Y. Bengio. 2017. A Structured Self-attentive Sentence Embedding. (2017). arXiv:arXiv preprint arXiv:1703.03130
[17]
S. Luan, M. Zhao, X. Chang, and D. Precup. 2019. Break the Ceiling: Stronger Multi-scale Deep Graph Convolutional Networks. In NeurIPS. 10945--10955.
[18]
Z. Lv, J. Xu, K. Zheng, H. Yin, P. Zhao, and X. Zhou. 2018. LC-RNN: A Deep Learning Model for Traffic Speed Prediction. In IJCAI. 3470--3476.
[19]
X. Ma, Z. Tao, Y. Wang, H. Yu, and Y. Wang. 2015. Long Short-Term Memory Neural Network for Traffic Speed Prediction Using Remote Microwave Sensor Data. Transportation Research Part C: Emerging Technologies54 (2015), 187--197.
[20]
A. V. D. Oord, S. Dieleman, H. Zen, K. Simonyan, O. Vinyals, A. Graves, N. Kalchbrenner, A. Senior, and K. Kavukcuoglu. 2016. WaveNet: A Generative Model for Raw Audio. arXiv preprint arXiv:1609.03499(2016).
[21]
A. Radford, J. Wu, R. Child, D. Luan, D. Amodei, and I. Sutskever. 2019. Language Models are Unsupervised Multitask Learners. Open AI(2019).
[22]
X. SHI, Z. Chen, H. Wang, D. Yeung, W. Wong, and W. WOO. 2015. Convolutional LSTM Network: A Machine Learning Approach for Precipitation Nowcasting. In NeurIPS. 802--810.
[23]
X. Shi, Z. Gao, L. Lausen, H. Wang, D. Yeung, W. Wong, and W. WOO. 2017. Deep Learning for Precipitation Nowcasting: A Benchmark and A New Model. In NeurIPS. 5617--5627.
[24]
X. Song, H. Kanasugi, and R. Shibasaki. 2016. Deep Transport: Prediction and Simulation of Human Mobility and Transportation Mode at a Citywide Level. In IJCAI. 2618--2624.
[25]
A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, ?. Kaiser,and I. Polosukhin. 2017. Attention is All you Need. In NeurIPS. 5998--6008.
[26]
Y. Wu and H. Tan. 2016. Short-Term Traffic Flow Forecasting with Spatial-Temporal Correlation in a Hybrid Deep Learning Framework. arXiv preprint arXiv:1612.01022(2016).
[27]
Z. Yang, Z. Dai, Y. Yang, J. Carbonell, R. Salakhutdinov, and Q. V. Le. 2019. XLNet: Generalized Autoregressive Pretraining for Language Understanding. arXiv preprint arXiv:1906.08237(2019).
[28]
H. Yao, X. Tang, H. Wei, G. Zheng, and Y. Yu. 2019. Revisiting Spatial-Temporal Similarity: A Deep Learning Framework for Traffic Prediction. In AAAI. 227--234.
[29]
H. Yao, F. Wu, J. Ke, X. Tang, Y. Jia, S. Lu, P. Gong, J. Ye, and Z. Li. 2018. Deep Multi-View Spatial-Temporal Network for Taxi Demand Prediction. In AAAI.
[30]
B. Yu, H. Yin, and Z. Zhu. 2018. Spatio-Temporal Graph Convolutional Networks:A Deep Learning Framework for Traffic Forecasting. In IJCAI. 3634--3640.
[31]
J. Zhang, Y. Zheng, and D. Qi. 2017. Deep Spatio-Temporal Residual Networks for Citywide Crowd Flows Prediction. In AAAI. 1655--1661.
[32]
J. Zhang, Y. Zheng, D. Qi, R. Li, and X. Yi. 2016. DNN-based Prediction Model for Spatio-temporal Data. In ACM SIGSPATIAL. 92:1--92:4.
[33]
J. Zhang, Y. Zheng, J. Sun, and D. Qi. 2019. Flow Prediction in Spatio-Temporal Networks Based on Multitask Deep Learning. IEEE Trans. on Knowledge and Date Engineering Early Access (2019).
[34]
C. Zheng, X. Fan, C. Wang, and J. Qi. 2020. GMAN: A Graph Multi-Attention Network for Traffic Prediction. In AAAI.
[35]
C. Zheng, X. Fan, C. Wen, L. Chen, C. Wang, and J. Li. 2019. DeepSTD: Mining Spatio-Temporal Disturbances of Multiple Context Factors for Citywide Traffic Flow Prediction.IEEE Trans. on Intelligent Transportation Systems Early Access(2019).

Cited By

View all
  • (2024)STEFT: Spatio-Temporal Embedding Fusion Transformer for Traffic PredictionElectronics10.3390/electronics1319381613:19(3816)Online publication date: 27-Sep-2024
  • (2024)STFM: Accurate Spatio-Temporal Fusion Model for Weather ForecastingAtmosphere10.3390/atmos1510117615:10(1176)Online publication date: 30-Sep-2024
  • (2024)Spatio-Temporal Memory Augmented Multi-Level Attention Network for Traffic PredictionIEEE Transactions on Knowledge and Data Engineering10.1109/TKDE.2023.3322405(1-16)Online publication date: 2024
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
KDD '20: Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining
August 2020
3664 pages
ISBN:9781450379984
DOI:10.1145/3394486
This work is licensed under a Creative Commons Attribution International 4.0 License.

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 20 August 2020

Check for updates

Author Tags

  1. attention mechanism
  2. long-term prediction
  3. mining spatial-temporal information
  4. neural network

Qualifiers

  • Research-article

Funding Sources

  • The Science and Technology Development Fund Macau SAR
  • University of Macau
  • Chinese National Research Fund (NSFC) Key Project

Conference

KDD '20
Sponsor:

Acceptance Rates

Overall Acceptance Rate 1,133 of 8,635 submissions, 13%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)332
  • Downloads (Last 6 weeks)32
Reflects downloads up to 17 Oct 2024

Other Metrics

Citations

Cited By

View all
  • (2024)STEFT: Spatio-Temporal Embedding Fusion Transformer for Traffic PredictionElectronics10.3390/electronics1319381613:19(3816)Online publication date: 27-Sep-2024
  • (2024)STFM: Accurate Spatio-Temporal Fusion Model for Weather ForecastingAtmosphere10.3390/atmos1510117615:10(1176)Online publication date: 30-Sep-2024
  • (2024)Spatio-Temporal Memory Augmented Multi-Level Attention Network for Traffic PredictionIEEE Transactions on Knowledge and Data Engineering10.1109/TKDE.2023.3322405(1-16)Online publication date: 2024
  • (2024)Spatial–Temporal Dynamic Graph Convolutional Network With Interactive Learning for Traffic ForecastingIEEE Transactions on Intelligent Transportation Systems10.1109/TITS.2024.336214525:7(7645-7660)Online publication date: Jul-2024
  • (2024)Contrasting Estimation of Pattern Prototypes for Anomaly Detection in Urban Crowd FlowIEEE Transactions on Intelligent Transportation Systems10.1109/TITS.2024.335514325:8(10231-10245)Online publication date: Aug-2024
  • (2024)Road Network Enhanced Traffic Flow Prediction Service2024 IEEE International Conference on Web Services (ICWS)10.1109/ICWS62655.2024.00156(1300-1305)Online publication date: 7-Jul-2024
  • (2024)A Just-In-Time Framework for Continuous Routing2024 IEEE 40th International Conference on Data Engineering (ICDE)10.1109/ICDE60146.2024.00350(4600-4613)Online publication date: 13-May-2024
  • (2024)Adopting Ensemble Learning and Machine Learning Techniques for Predictive Modeling in Traffic Data Analysis2024 3rd International Conference on Applied Artificial Intelligence and Computing (ICAAIC)10.1109/ICAAIC60222.2024.10575594(528-533)Online publication date: 5-Jun-2024
  • (2024)Enhancing Traffic Data Prediction: Fusion of Autoencoder-Based Spatiotemporal Reconstruction and Forecasting2024 3rd International Conference on Big Data, Information and Computer Network (BDICN)10.1109/BDICN62775.2024.00017(53-57)Online publication date: 12-Jan-2024
  • (2024)A Memory-augmented Conditional Neural Process model for traffic predictionKnowledge-Based Systems10.1016/j.knosys.2024.112578304(112578)Online publication date: Nov-2024
  • Show More Cited By

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Get Access

Login options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media