TY - JOUR
T1 - HydroBench
T2 - Jupyter supported reproducible hydrological model benchmarking and diagnostic tool
AU - Moges, Edom
AU - Ruddell, Benjamin L.
AU - Zhang, Liang
AU - Driscoll, Jessica M.
AU - Norton, Parker
AU - Perez, Fernando
AU - Larsen, Laurel G.
N1 - Publisher Copyright:
Copyright © 2022 Moges, Ruddell, Zhang, Driscoll, Norton, Perez and Larsen.
PY - 2022/9/30
Y1 - 2022/9/30
N2 - Evaluating whether hydrological models are right for the right reasons demands reproducible model benchmarking and diagnostics that evaluate not just statistical predictive model performance but also internal processes. Such model benchmarking and diagnostic efforts will benefit from standardized methods and ready-to-use toolkits. Using the Jupyter platform, this work presents HydroBench, a model-agnostic benchmarking tool consisting of three sets of metrics: 1) common statistical predictive measures, 2) hydrological signature-based process metrics, including a new time-linked flow duration curve and 3) information-theoretic diagnostics that measure the flow of information among model variables. As a test case, HydroBench was applied to compare two model products (calibrated and uncalibrated) of the National Hydrologic Model - Precipitation Runoff Modeling System (NHM-PRMS) at the Cedar River watershed, WA, United States. Although the uncalibrated model has the highest predictive performance, particularly for high flows, the signature-based diagnostics showed that the model overestimates low flows and poorly represents the recession processes. Elucidating why low flows may have been overestimated, the information-theoretic diagnostics indicated a higher flow of information from precipitation to snowmelt to streamflow in the uncalibrated model compared to the calibrated model, where information flowed more directly from precipitation to streamflow. This test case demonstrated the capability of HydroBench in process diagnostics and model predictive and functional performance evaluations, along with their tradeoffs. Having such a model benchmarking tool not only provides modelers with a comprehensive model evaluation system but also provides an open-source tool that can further be developed by the hydrological community.
AB - Evaluating whether hydrological models are right for the right reasons demands reproducible model benchmarking and diagnostics that evaluate not just statistical predictive model performance but also internal processes. Such model benchmarking and diagnostic efforts will benefit from standardized methods and ready-to-use toolkits. Using the Jupyter platform, this work presents HydroBench, a model-agnostic benchmarking tool consisting of three sets of metrics: 1) common statistical predictive measures, 2) hydrological signature-based process metrics, including a new time-linked flow duration curve and 3) information-theoretic diagnostics that measure the flow of information among model variables. As a test case, HydroBench was applied to compare two model products (calibrated and uncalibrated) of the National Hydrologic Model - Precipitation Runoff Modeling System (NHM-PRMS) at the Cedar River watershed, WA, United States. Although the uncalibrated model has the highest predictive performance, particularly for high flows, the signature-based diagnostics showed that the model overestimates low flows and poorly represents the recession processes. Elucidating why low flows may have been overestimated, the information-theoretic diagnostics indicated a higher flow of information from precipitation to snowmelt to streamflow in the uncalibrated model compared to the calibrated model, where information flowed more directly from precipitation to streamflow. This test case demonstrated the capability of HydroBench in process diagnostics and model predictive and functional performance evaluations, along with their tradeoffs. Having such a model benchmarking tool not only provides modelers with a comprehensive model evaluation system but also provides an open-source tool that can further be developed by the hydrological community.
KW - Hydrological Modeling
KW - Kling-Gupta
KW - Model Benchmarking
KW - Model Diagnostics
KW - Model Evaluation
KW - Nash Sutcliffe
KW - Reproducibility
KW - Uncertainty Analysis
UR - http://www.scopus.com/inward/record.url?scp=85140205246&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85140205246&partnerID=8YFLogxK
U2 - 10.3389/feart.2022.884766
DO - 10.3389/feart.2022.884766
M3 - Article
AN - SCOPUS:85140205246
SN - 2296-6463
VL - 10
JO - Frontiers in Earth Science
JF - Frontiers in Earth Science
M1 - 884766
ER -