Code & Models
š„ Probing the Limit of Hydrologic Predictability with the Transformer
- First rigorous benchmark comparing Transformers against LSTM on the CAMELS streamflow data set.
- Shows vanilla Transformers lag on high-flow metrics, while the modified variant matches or slightly exceeds LSTM in KGE and other scoresāhinting we are near the data-set predictability limit.
- Take-home: non-recurrent designs are promising for large-scale, knowledge-stored hydrologic models.
Full abstract
For a number of years since their introduction to hydrology, recurrent neural networks like long short-term memory (LSTM) networks have proven remarkably difficult to surpass in terms of daily hydrograph metrics on community-shared benchmarks. Outside of hydrology, Transformers have now become the model of choice for sequential prediction tasks, making it a curious architecture to investigate for application to hydrology. Here, we first show that a vanilla (basic) Transformer architecture is not competitive against LSTM on the widely benchmarked CAMELS streamflow dataset, and lagged especially prominently for the high-flow metrics, perhaps due to the lack of memory mechanisms. However, a recurrence-free variant of the Transformer model obtained mixed comparisons with LSTM, producing very slightly higher Kling-Gupta efficiency coefficients (KGE), along with other metrics. The lack of advantages for the vanilla Transformer network is linked to the nature of hydrologic processes. Additionally, similar to LSTM, the Transformer can also merge multiple meteorological forcing datasets to improve model performance. Therefore, the modified Transformer represents a rare competitive architecture to LSTM in rigorous benchmarks. Valuable lessons were learned: (1) the basic Transformer architecture is not suitable for hydrologic modeling; (2) the recurrence-free modification is beneficial, so future work should continue to test such modifications; and (3) the performance of state-of-the-art models may be close to the prediction limits of the dataset. As a non-recurrent model, the Transformer may bear scale advantages for learning from bigger datasets and storing knowledge. This work lays the groundwork for future explorations into pretraining models, serving as a foundational benchmark that underscores the potential benefits in hydrology.š°ļø A Multiscale Deep Learning Model for Soil Moisture
- Unified multiscale scheme jointly learns from satellite and in-situ data to predict daily 9 km 5 cm-depth soil moisture.
- Spatial cross-validation over CONUS sites: median corr = 0.901, RMSE = 0.034 m³/m³āoutperforming SMAP 9 km, in-situ-only DL, and land-surface models.
- Generic blueprint for other geoscience problems that span multiple spatial scales.
Full abstract
Deep learning (DL) models trained on hydrologic observations can perform extraordinarily well, but they can inherit deficiencies of the training data, such as limited coverage of in situ data or low resolution/accuracy of satellite data. Here we propose a novel multiscale DL scheme learning simultaneously from satellite and in situ data to predict 9 km daily soil moisture (5 cm depth). Based on spatial cross-validation over sites in the conterminous United States, the multiscale scheme obtained a median correlation of 0.901 and root-mean-square error of 0.034 m3/m3. It outperformed the Soil Moisture Active Passive satellite mission's 9 km product, DL models trained on in situ data alone, and land surface models. Our 9 km product showed better accuracy than previous 1 km satellite downscaling products, highlighting limited impacts of improving resolution. Not only is our product useful for planning against floods, droughts, and pests, our scheme is generically applicable to geoscientific domains with data on multiple scales, breaking the confines of individual data sets.Data Sets
š Global Soil Moisture from a Multitask Model (GSM3)
- Global multitask LSTM product; median corr = 0.792 in random hold-out and continental hold-out tests.
- Performs surprisingly well in Africa & Australia even when those continents are excluded from training; integrated into an operational advisory platform serving 13 M African farmers.
Field | Value |
---|---|
Short name | GSM3 |
Long name | Global Soil Moisture Dataset From a Multitask Model |
Version | 1.0 |
Format | GeoTIFF |
Spatial coverage | Global |
Temporal coverage | 2015-01-01 ā 2020-12-31 |
Resolution | 9 km (spatial) ⢠daily (temporal) |
CRS | EPSG:6933 ā WGS 84 / NSIDC EASE-Grid 2.0 Global |
File size | ā 10.3 MB per tile |