In a paper published in the journal Scientific Reports, researchers introduced the random search (RS), long short-term memory networks (LSTM) transformer hybrid model for flood forecasting. This model combines RS, LSTM networks, and transformer architecture to address challenges in traditional hydrology models and data-driven approaches.
Tested in the Jingle watershed, it outperformed other models in simulating flood processes, offering more accurate forecasts, especially as lead times extended. The study underscores the model's potential for flood forecasting and the benefits of integrating diverse data-driven techniques for improved modeling.
Background
Previous research has tackled flood prediction challenges using traditional physical models and data-driven approaches like LSTM networks. However, these methods must be improved in handling temporal dependencies and data requirements. Attention mechanisms in models like the transformer show promise but require optimization. LSTM and transformer models hold potential for hydrological forecasting, warranting further exploration.
Hybrid Flood Forecasting Methodology
The methods section outlines the approach taken in this study, starting with the use of the RS algorithm for hyperparameter optimization. RS, introduced in 2012, offers a more efficient alternative to Grid Search by introducing randomness into the search process. For comparison, parameters for the LSTM-Transformer model and LSTM, Transformer, backpropagation (BP), and multilayer perceptron (MLP) models were optimized using RS. The process involved defining the search space, applying the RS algorithm to find optimal parameter combinations, and constructing models based on the best-performing parameter sets.
The original transformer model's limitations in handling multivariate time series data for flood forecasting prompted the development of an improved version. This enhanced model discarded positional encoding and the decoder component, instead incorporating convolutional layers and global average pooling structures to capture local features in time series data better. It retained the multi-head scaled dot-product attention mechanism from the original Transformer, which calculates outputs as weighted sums of values based on similarity with queries. These improvements aimed to adapt the transformer model more effectively to flood forecasting tasks.
Integrating an LSTM layer into the improved transformer model resulted in the RS-LSTM-Transformer hybrid model. This hybrid model featured a single-layer LSTM for feature extraction, multiple encoding layers, and an output layer. The encoding layers included multi-head scaled dot-product attention, residual connections, normalization, and convolutional layers. A dropout layer was added to prevent overfitting, and the output layer incorporated Global Average Pooling for dimensionality reduction and a final dense layer for output generation.
The RS optimization algorithm was employed to construct the optimal RS-LSTM-Transformer model. The case study focused on the Jingle watershed in the middle reaches of the Yellow River, a region prone to flood disasters. Data collection included hourly flow data from the Jingle hydrological station and rainfall data from 14 other stations, covering 98 flood events. Statistical analysis of these events highlighted significant variations in rainfall duration, peak discharge, and rainfall center, emphasizing the complexity of the rainfall-runoff process and the challenges in modeling it effectively.
Flood Forecasting Advancements
The study's comprehensive evaluation of flood forecasting models reveals the robustness and efficacy of the RS-LSTM-Transformer model, particularly in simulating rainfall-runoff dynamics. Across various evaluation metrics, including Nash Sutcliffe efficiency (NSE), root mean square error (RMSE), mean absolute error (MAE), and bias, the RS-LSTM-Transformer consistently outperforms other models, showcasing its superior predictive accuracy. Notably, even as the lead time extends, the RS-LSTM-Transformer maintains relatively high precision, indicating its suitability for long-term forecasting.
Conversely, models such as RS-BP and RS-MLP exhibit a more significant decline in performance metrics with increased lead time, underscoring their limitations in long-term predictions. The results emphasize the RS-LSTM-Transformer's potential as a novel approach to flood forecasting, promising enhanced accuracy and stability in forecasts. Analysis of training efficiency and time consumption among the different models further highlights the advantages of the RS-LSTM-Transformer.
Despite its slightly longer training time than the RS-Transformer, the RS-LSTM-Transformer model demonstrates a more concentrated distribution of training time results. This efficiency can be attributed to incorporating an LSTM layer into the input section, which optimizes data processing and facilitates model convergence. The reduction in complexity and information volume handled by subsequent layers, particularly the Transformer, contributes to improved training efficiency and stability, enhancing the overall performance of the RS-LSTM-Transformer model.
Evaluations were extended to the Guxian watershed in the Luo River to assess the RS-LSTM-Transformer model's versatility and showcase its applicability across different hydrological contexts. Comparisons with the LSTM-Transformer and transformer models in the Guxian watershed reaffirm the superior performance of the RS-LSTM-Transformer.
With highly accurate predictions, even in diverse environmental settings, the RS-LSTM-Transformer underscores its potential as a versatile tool for flood forecasting. Its ability to adapt to varying climates, land uses, and hydrological characteristics positions the RS-LSTM-Transformer as a promising solution for improving flood prediction accuracy in diverse geographical regions.
Conclusion
To sum up, the RS-LSTM-Transformer hybrid model significantly improved rainfall-runoff simulation by integrating LSTM into the transformer framework and optimizing parameters through RS optimization. It consistently outperformed other models across 98 flood instances, demonstrating high accuracy and robustness, especially in the Jingle watershed of the Yellow River. Challenges include data quality and computational demands, suggesting potential avenues for future research to enhance accuracy and stability, such as integrating physics-based models and testing transferability across diverse conditions.