1. Learning on predictions: Fusing training and autoregressive inference for long-term spatiotemporal forecasts.
- Author
-
Vlachas, P.R. and Koumoutsakos, P.
- Subjects
- *
DATA distribution , *NATURAL language processing , *RECURRENT neural networks , *WEATHER forecasting , *FLUID flow - Abstract
Predictions of complex systems ranging from natural language processing to weather forecasting have benefited from advances in Recurrent Neural Networks (RNNs). RNNs are typically trained using techniques like Backpropagation Through Time (BPTT) to minimize one-step-ahead prediction loss. During testing, RNNs often operate in an auto-regressive mode, with the output of the network fed back into its input. However, this process can eventually result in exposure bias since the network has been trained to process "ground-truth" data rather than its own predictions. This inconsistency causes errors that compound over time, indicating that the distribution of data used for evaluating losses differs from the actual operating conditions encountered by the model during training. Inspired by the solution to this challenge in language processing networks we propose the Scheduled Autoregressive Truncated Backpropagation Through Time (BPTT-SA) algorithm for predicting complex dynamical systems using RNNs. We find that BPTT-SA effectively reduces iterative error propagation in Convolutional and Convolutional Autoencoder RNNs and demonstrates its capabilities in the long-term prediction of high-dimensional fluid flows. [Display omitted] • RNNs trained with teacher forcing (TF) suffer from the exposure bias effect • Exposure bias is the incosistentcy between TF training and autoregressive inference • Scheduled Autoregressive BPTT (BPTT-SA) alleviates the exposure bias effect • We employ BPTT-SA for long-term forecasting of the Navier-Stokes flow past a cylinder • BPTT-SA improves long-term forecasting accuracy for CNN-RNNs, and Conv-RNNs [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF