Next Article in Journal
Submarine Groundwater Discharge Differentially Modifies Photosynthesis, Growth, and Morphology for Two Contrasting Species of Gracilaria (Rhodophyta)
Previous Article in Journal
An Empirical Mode-Spatial Model for Environmental Data Imputation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Impact of Hydrological Modellers’ Decisions and Attitude on the Performance of a Calibrated Conceptual Catchment Model: Results from a ‘Modelling Contest’

by
Helge Bormann
1,*,
Mariana Madruga De Brito
2,
Despoina Charchousi
3,
Dimitris Chatzistratis
4,
Amrei David
5,
Paula Farina Grosser
6,
Jenny Kebschull
1,
Alexandros Konis
7,
Paschalis Koutalakis
8,
Alkistis Korali
8,
Naomi Krauzig
9,
Jessica Meier
10,
Varvara Meliadou
11,
Markus Meinhardt
12,
Kieran Munnelly
2,
Christiane Stephan
2,
Leon Frederik De Vos
13,
Jörg Dietrich
14 and
Ourania Tzoraki
8
1
Department of Research and Knowledge Transfer, Jade University of Applied Sciences, Ofener Straße 16/19, 26121 Oldenburg, Germany
2
Department of Geography, University of Bonn, Meckenheimer Allee 166, 53115 Bonn, Germany
3
NTU Athens, School of Rural and Surveying Engineering, Zografou Campus 9, Iroon Polytechniou str, 15780 Zografou, Greece
4
Department of Physical Geography, Faculty of Geosciences, Utrecht University, 3512 JE Utrecht, The Netherlands
5
TU Darmstadt, Franziska-Braun-Str. 7, 64287 Darmstadt, Germany
6
BTU Cottbus, Postfach 101344, 03013 Cottbus, Germany
7
TU Dresden, Mommsenstraße 13, 01602 Dresden, Germany
8
University of the Aegean, University Hill, 81100 Mytilene, Lesvos, Greece
9
Parthenope University of Naples, Via Ammiraglio Ferdinando Acton, 38, 80133 Napoli, NA, Italy
10
University of Potsdam, Am Neuen Palais 10, 14469 Potsdam, Germany
11
Democritus University of Thrace, University Campus, 69100 Komotini, Greece
12
University of Jena, Löbdergraben 32, 07743 Jena, Germany
13
TU München, Arcisstraße 21, 80333 München, Germany
14
University of Hannover, Appelstr. 9a, 30167 Hannover, Germany
*
Author to whom correspondence should be addressed.
Hydrology 2018, 5(4), 64; https://doi.org/10.3390/hydrology5040064
Submission received: 20 October 2018 / Revised: 14 November 2018 / Accepted: 15 November 2018 / Published: 19 November 2018

Abstract

:
In this study, 17 hydrologists with different experience in hydrological modelling applied the same conceptual catchment model (HBV) to a Greek catchment, using identical data and model code. Calibration was performed manually. Subsequently, the modellers were asked for their experience, their calibration strategy, and whether they enjoyed the exercise. The exercise revealed that there is considerable modellers’ uncertainty even among the experienced modellers. It seemed to be equally important whether the modellers followed a good calibration strategy, and whether they enjoyed modelling. The exercise confirmed previous studies about the benefit of model ensembles: Different combinations of the simulation results (median, mean) outperformed the individual model simulations, while filtering the simulations even improved the quality of the model ensembles. Modellers’ experience, decisions, and attitude, therefore, have an impact on the hydrological model application and should be considered as part of hydrological modelling uncertainty.

1. Introduction

Reliable flood prediction and water management requires robust hydrological model applications. While a high accuracy of the prediction is usually desired, the underlying uncertainty is a major issue of predicting hydrological variables. Uncertainty analysis is, therefore, a key technique to assess the robustness of model simulations and to communicate the findings to decision makers.
Different sources of uncertainty are usually discussed and considered in the literature. Uncertainty in modelling [1] is usually attributed to:
(1)
Structural and technical uncertainty caused by the underlying model philosophy, the representation of the hydrological processes, its degree of physics [2,3], and the computational implementation;
(2)
parameter uncertainty [4,5]; and
(3)
data uncertainty [6,7].
The international hydrological community has focused on these uncertainty sources for the past decades, however, researchers have only rarely considered the modeller her-/himself explicitly as a behavioural aspect of model uncertainty in environmental modelling [8]. Impacts of the modellers were mostly subsumed as part of the parameter uncertainty. However, the modeller takes decisions during the model selection and application process and has an impact on the model results.
What kind of decisions does a modeller take during model application? She/he chooses a model, selects adequate (or at least available) data, parameterizes the model, and selects objective functions and methods for calibration (e.g., manual vs. automatic optimization of parameters, structured sensitivity analysis prior to calibration). Modellers may have an incomplete understanding of the model, may not be fully motivated, or may have a lack of time to reach the best possible solution. Only few studies have tried to analyse such an impact of the modeller. The influence of modellers’ decisions on model results in risk assessment was investigated by [9]. A study by [10,11,12] focused on the influence modellers have on a priori predictions of catchment hydrological processes by comparing a priori model applications to a small, artificial catchment in Northeast Germany. However, they struggled with the instance that modellers applied different models on the same catchment and data set. Thus, only the combination of modellers’ decisions and the model used could be analysed. Typically, instead of searching for the perfect model for a distinct model application, most modellers prefer using their own and accordingly their standard model for model application varies in different catchments and for different purposes. This is due to trust and experience, and may include slight model adjustments depending on the application. This happened also in another study [10,11], and restricted them to analyse the combination of different models and modellers.
Model ensembles are often constructed to frame the uncertainty of a simulation. Early applications in numerical weather forecasting were done by perturbing the initial conditions of the model [13]. Other sources of uncertainty, or even combined sources of uncertainty, were used to generate ensembles. It was found that multi-model ensembles outperformed single-model ensembles [14]. Multi-model ensembles outperformed single models [15,16]. Ensembles were also successfully applied in flood forecasting [17]. However, analyses of single-model ensembles did not explicitly consider the impact of the modellers themselves. However, there is evidence of a collective intelligence of groups of humans, which can outperform the average intelligence and even the highest intelligence of single humans [18].
To overcome the shortcomings described, and to quantify the impact of the modeller on model application, we designed a modelling-contest-experiment as part of a DAAD (German Academic Exchange Service) summer school on floods and flood risk management. We asked different modellers with diverse hydrological modelling experience and attitudes to apply the same catchment model to one Greek catchment based on identical data and model codes. We assume that experience, attitudes, and the individual calibration strategy have an impact on the model results achieved by different modellers. Based on this experimental design, the power of model ensembles with regard to the impact of modellers’ decisions was also investigated. Due to collective intelligence, we assume modellers’ ensembles to be superior to single modellers’ results. Based on the conceptual design of the study (Section 2), we contribute to the following research questions:
(1)
Do different modellers achieve different model results by applying the same model to the same catchment? (Section 3.1)
(2)
Can the variability in the model results be explained by the experience, attitudes, and calibration strategy of the different modellers? (Section 3.3)
(3)
Do building modellers’ ensembles improve the model performance compared to the results achieved by individual modellers? (Section 3.4)

2. Materials and Methods

2.1. HBV Model

The HBV model (Hydrologiska Byråns Vattenbalansavdelning) is a spatially semi-distributed and conceptual hydrological catchment model [19,20]. In this study the HBV-light version of HBV is used [21]. HBV considers the division of various components, such as snow, soil, upper and lower reservoir, and the river, and estimates the water mass balance in each of them. A basin can be delineated in numerous sub-basins in relation to the user preferences, experience, and basin characteristics. For each sub-basin, HBV calculates the catchment discharge, usually on a daily time step, which is estimated based on time series of precipitation and air temperature as well as potential evaporation estimates. Time series of crop-water-need (Kc), the volume of water needed by the various crops to grow optimally, are provided by the user for each sub-basin. Groundwater recharge and actual evaporation are simulated as functions of the actual soil water storage. Snow accumulation and snowmelt are computed by a degree-day method. The routing routine is based on a triangular weighting function to simulate the routing of the runoff to the catchment outlet. Detailed model descriptions and the central equations can be found, for example, in [19,20,21].
A set of parameters (Table 1) must be calibrated by the user related to the snow routine, soil routine, upper and lower zone response routine, and the transformation routing routine, as well as the initial values in each reservoir. In previous studies, HBV type models were successfully applied to semi-arid catchments [22,23]. It can therefore be assumed that HBV can represent the dominant processes of Mediterranean catchments.

2.2. Tsiknias River Catchment

The Tsiknias river catchment is located in the north-central part of Lesvos Island (Greece, Figure 1). The river drains an area of about 90 km2. The elevation of the catchment ranges from 0 to 968 m. The topography is characterized partly as lowland and as middle-mountainous. The semi-arid climate of the area is typically Mediterranean, characterized by hot and dry summers and mild rainy winters with a high relative humidity. Two meteorological stations operate in the basin: (a) Stipsi (396 m altitude) with an annual mean rain value 870 mm, and (b) Agia Paraskevi (95 m altitude) with an annual mean rain value 664 mm. Average annual potential evapotranspiration is estimated around 1050 mm/a. Since 2014, a telemetric station (Prini) operates on the main channel, providing water level data at 15 min intervals. The main part of the watershed is covered by cultivated land (mainly olive groves), grassland, and brushwood. In a few source areas, pine and oak woods have developed. In most parts of the catchment, soils are developed from volcanic rocks. Soil permeability is low in most of the area, low-medium in the highland areas, and medium in the north part of the area near the estuary. Since the Tsiknias River frequently experiences flash floods, a robust model application is an essential requirement for regional flood prediction and management.
For this study, three years of daily data of the river gauge (Prini; July 2014 to March 2017) was used as well as daily data of one climatic station close to the river gauge. Data was quality-checked before model application. While usually more than three years of data should be used for hydrological catchment model application to represent hydrological variability, in the case of the Tsiknias catchment for this time period, the high quality of the data could be guaranteed. Starting the simulation period after the dry summer period, a warm-up period was not required assuming low antecedent moisture conditions.

2.3. Design of the Modelling Contest

After a general introduction into the HBV model theory, including an explanation of the process equations and the parameterisation of the model, all 17 participants of the modelling contest had one working day available for HBV model calibration to the Tsiknias River. MSc students, PhD students, and Postdocs with diverse modelling experience were participating in the experiment. All modellers were supplied with identical hydro-climatological data and model codes. They were asked not to communicate on their individual model applications during the model contest. At the end of the day, modellers were asked to deliver their manually calibrated parameter sets and their simulation results, and to complete a questionnaire on their hydrological modelling experience and attitudes (see Section 2.4). The best hydrographs according to the Nash-Sutcliffe model efficiency [24] were collected from all modellers as well as their questionnaires. Both sources were used for further evaluation of the results.

2.4. Design of the Questionnaire

The questionnaire on the modellers’ expertise and attitudes consisted of seven questions on modelling experience (based on [11]): If any, how many years, how many different models, experience with HBV (if yes, how many years), experience in semi-arid climates, experience in different climates), two questions on the calibration strategy (trial and error or sensitivity analysis or any other; choice of statistical quality measures for calibration), and a final statement whether the modellers enjoyed the exercise or not.
To analyse the results of the survey in a semi-quantitative way, the following classifications were assumed concerning modelling experience and attitudes:
(1)
Modelling experience was classified into three different classes: No experience, little experience (1 model, max. 1 year experience), some experience (either more than one model, or more than one year experience, or both); and
(2)
The degree of enjoyment was directly taken from the survey (yes/no); however, some participants replied that they enjoyed the contest in general, but they mentioned some criticism (“yes, but...”). They were ranked as a third class between “enjoyed” and “did not enjoy” for the evaluation (“mostly”).

3. Results and Discussion

3.1. HBV Model Application to Tsiknias River Catchment

All participating modellers calibrated the HBV model manually to the Tsiknias catchment for a three year period. Figure 2 gives an impression on the variability of the best simulated hydrograph of each modeller, selected by maximum Nash-Sutcliffe efficiency (NSE). The variability among the simulated hydrographs is quite high, and both underestimation and overestimation of the observations during the rainy season were delivered by the modellers. Figure 3 emphasizes this variability in simulated hydrographs for three exemplary runoff events in early 2015.
Analysing the flow duration curves derived from observations and simulations revealed that none of the modellers succeeded in simulating the low flows during the dry season (Figure 4). While the modellers simulated “no flow” for 40–80% of the year, at least marginal flows were observed for the entire simulation period (discharge of less than 20 l/s for 31% of the year). The variability in simulated discharges was also high with regard to the simulated long term water balances (see Table 2).
While modellers used individual objective functions for calibration, a standard set of criteria was used for the evaluation of the goodness-of-fit of the model results. Statistical analysis of the goodness-of-fit criteria used for evaluation of the simulation results revealed that the simulation quality of the hydrographs delivered by the different modellers differed a lot.
Analysing the calibrated parameters sets did not yield distinct systematics. Depending on their individual calibration strategies, different modellers tried to optimize different parameter sub-sets, in exceptional cases also resulting in implausible parameters (see Table 3 for ranges of calibrated parameters). Since different parameter combinations resulted in similar goodness-of-fit values, parameter equifinalty is obvious for the selected parameter range.
The comparison against the simulation criteria defined by [25] yielded that eight out of 17 modellers achieved satisfactory results related to the Nash-Sutcliffe efficiency ([24]; NSE ≥ 0.5) while nine out of 17 modellers achieved good or satisfactory results related to percent bias (PB ≤ 25%).
These numbers showed that about half of the modellers achieved “satisfactory” simulation results (applying criteria that were designed for monthly time steps to daily simulations) while the remaining half did not (for details see Table 2).
The goodness-of-fit of the simulated discharges was relatively low compared to other HBV applications. We assume that this was at least partly due to the design of the modelling contest, being based on the limited time of the contest (one working day) and working in an unknown environment. Beyond that, we assume that hydrological simulation of a semi-arid catchment, being characterized by an intermittent flow regime, is a specific challenge.

3.2. Experience, Decisions, and Attitudes of the Modellers

The evaluation of the questionnaire revealed that the hydrological modelling experience of the participants differed a lot.
The following answers were provided on the modelling experience: While four out of 17 participants had no hydrological modelling experience at all, nine out of 17 participants had experience with one or two models, and four out of 17 participants had experience with three or more models. Almost half of the participants, eight out of 17, had less than one year modelling experience. Five out of 17 participants had one to two years modelling experience, and four out of 17 participants had three or more years modelling experience. Only one of the 17 participants had modelling experience with HBV. Beyond the modelling experience, knowledge on the regionally important characteristics and processes were asked for as well. Only five of 17 participants worked already in semi-arid regions while 12 out of 17 participants had never worked on the hydrology of semi-arid regions.
With regard to model calibration, most of the modellers calibrated their model by trial and error (10 out of 17) while only six out of 17 applied a systematic sensitivity analysis in order to reduce the number of calibration parameters selected for manual calibration. Most of the modellers used the NSE for calibration (14 out of 17) while only a few modellers considered the water balance (seven out of 17) or the root mean squared error (RMSE) (four out of 17). The use of multiple statistical quality measures did not guarantee achieving satisfactory modelling results. Eight of 13 modellers achieved satisfactory NSE according to [25]. However, those who did not consider more than one index, did not achieve satisfactory NSE values at all. Three of them achieved the smallest NSE values of all modellers. The analysis of the relation of modellers’ experience to the calibration strategy and the choice of the quality measures used for calibration did not show any significant pattern.
Finally, 12 out of 17 participants enjoyed the modelling contest, while two out of 17 participants did not, and a further three out of 17 participants raised arguments pro and contra.

3.3. Evaluation of Model Results against Modellers Experience, Decisions, and Attitudes

The modelling results and the results of the survey were matched and evaluated against the NSE, which was selected for calibration by most of the modellers. Due to the limited number of modellers, the calibration strategy, attitudes, and experience of the modellers was grouped in few classes (Section 2.4) and individually compared to the simulation results. Comparisons between the classified groups are presented as follows. Due to the relatively small number of modellers, statistical significance of differences between different groups was not tested.
Relating the modelling experience against the goodness-of-fit revealed that an increasing experience helped to achieve a good modelling result while also, without any modelling experience, it was possible to achieve satisfactory modelling results according to [25]. However, the best NSE values, ranging from 0.53 to 0.56, were achieved by experienced modellers (Figure 5, left). Fifty-seven percent of the experienced modellers achieved better results than any of the less experienced modellers, and 71% of the experienced modellers achieved satisfactory simulation results (NSE ≥ 0.5) while the proportion decreased to 30% of less experienced modellers.
A similar picture was shown by analysing the experience on hydrological processes in intermittent catchments. While 60% of the experienced participants achieved satisfactory simulation results, the proportion of the unexperienced ones was significantly lower with 41%. Additionally, the variability within the simulation results of experienced modellers was lower compared to the unexperienced ones (Figure 5, right).
The comparison of the calibration strategies of the different modellers against the achieved NSE revealed large differences between the different options. From those participants who decided to apply a systematic sensitivity analysis, only one modeller did not achieve satisfactory results, while all others did (86%). From those modellers who did the calibration based on trial and error, only two (20%) achieved satisfactory results (Figure 6, left).
Similar results were obtained from the question whether modellers enjoyed the contest or not. Although the variability in NSE values of those who enjoyed the contest was larger compared to those did not like it or criticized it, the best six model applications overall were obtained from those modellers who did enjoy the model application and the contest (Figure 6, right). Therefore, enjoyment seems to be an important factor for being successful for hydrological modellers.
Analysing multifunctional combinations could provide deeper insight, but the number of participants was much too small to come to a reliable quantitative estimation. Towards a multifunctional analysis, we conclude that the questionnaire should have been designed in a different way, asking for standardized answers concerning experience and attitudes. Additionally, such an experiment should be realized with a larger group of modellers to better represent the variability within the groups of modellers.

3.4. Improvement of the Model Results through Calculating Model(ler) Ensembles

Finally, the results of the modelling contest were used for falsifying the hypothesis “multi-modeller-ensembles cannot yield better results of conceptual catchment models than a single modeller can obtain due to the uncertainty of the modellers’ decisions”. For each day, the simulated discharges were used to calculate median- and mean-value-ensembles. Three different strategies were tested to select appropriate ensemble members:
(1)
Considering all modellers: For this first multi-modeller-ensemble, the best simulated hydrograph (that with the highest NSE) simulated by each participating modeller was considered, independent of quality, strategy, or experience.
(2)
Considering only those modellers who achieved satisfactory results (a posteriori): For this second multi-modeller-ensemble, only the best simulated hydrograph (that with the highest NSE) simulated by those participating modellers was considered who achieved satisfactory modelling results according to [25] (NSE > 0.5).
(3)
Considering modellers according to their modelling experience (a priori): For this third strategy, the group of modellers was divided in two sub-groups, “participants with modelling experience” and “participants with little or without modelling experience”. For those two sub-groups, ensembles were calculated based on the best simulated hydrograph (that with the highest NSE) simulated by the members of the sub-groups.
Comparing those model ensembles against all individual simulations, depending on the root mean squared error (RMSE) and the Nash-Suttcliffe efficiency (NSE), shows that:
(4)
Independently of the quality of the individual model applications, those two ensembles based on all modellers’ results have a simulation quality comparable to the best individual modeller (Figure 7).
(5)
The two ensembles based on the simulations of those modellers’ results who achieved “satisfactory” results even outperformed the best individual simulation and revealed the most robust simulation results (Figure 7).
(6)
The two ensembles based on the simulated hydrographs of the experienced modellers also outperform all individual simulations (Figure 8) and even all other modellers ensembles investigated in this study (Figure 7). Additionally, both ensembles based on the simulated hydrographs of the less and unexperienced modellers achieved satisfactory modelling results according to [25] (NSE > 0.5).
Hydrographs of selected multi-modellers ensembles are presented in Figure 9. In accordance with [14,15,16], the combination of individual simulations, therefore, increases the simulation quality, also in the case of different modellers applying the same catchment model. Thus, for the case of multi-modellers-ensembles in hydrological modelling, collective intelligence seems to be superior to individual intelligence.

4. Conclusions

The main findings from this modelling contest are:
(1)
One key requirement of successful application of a conceptual catchment model is that a modeller has a good calibration strategy. A systematic sensitivity analysis helps a lot to identify the most sensitive model parameters and to make better decisions in the (manual) calibration process.
(2)
Available experience in model application and knowledge on regional processes of course helps to achieve reasonable model results. However, it does not necessarily guarantee a high model performance.
(3)
To enjoy what one is doing—hydrological model application in this case—is a supporting factor. The analysis of the contest results confirmed this also for hydrological model application.
(4)
“Modeller ensembles” can be used to show the uncertainty caused by behavioural aspects. Combining model results of different modellers helps to make predictions more robust, also with regard to the modellers’ decisions during the model application process. The good performance of a priori ensembles based on modellers’ experience emphasizes the power of collective intelligence also for the case of hydrological model applications.
This study shows, therefore, that modellers’ decisions and modellers’ attitudes have a significant impact on the success of a hydrological model application. Apart from experience and regional process knowledge, strategic issues and enjoyment play an important role for a successful model application. Additionally, collective intelligence matters also in the case of hydrological modelling.

Author Contributions

H.B. and O.T. acquired the funding from DAAD; the study was designed by H.B., J.D. and O.T.; model simulations were delivered by H.B., M.M.d.B., D.C., D.C., A.D., P.F.G., J.K., A.K., P.K., A.K., N.K., J.M., V.M., M.M., K.M., C.S., and L.F.d.V.; data analysis and evaluation was done by H.B.; the original draft was prepared by H.B.; J.K., J.D. and O.T. assisted in reviewing and editing the draft.

Funding

This research was funded by DAAD (German Academic Exchange Service), grant “Floods and Flood Risk Management” as part of the program “Hochschuldialog mit Südeuropa”.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Walker, W.E.; Harremöes, P.; Rotmans, J.; van der Sluijs, J.P.; van Asselt, M.B.A.; Janssen, P.; Krayer von Krauss, M.P. Defining Uncertainty: A Conceptual Basis for Uncertainty Management in Model-based Decision Support. Integr. Assess. 2003, 4, 5–17. [Google Scholar] [CrossRef]
  2. Breuer, L.; Huisman, J.A.; Willems, P.; Bormann, H.; Bronstert, A.; Croke, B.F.W.; Frede, H.-G.; Gräff, T.; Hubrechts, L.; Jakeman, A.J.; et al. Assessing the impact of land use change on hydrology by ensemble modeling (LUCHEM) I: Model intercomparison of current land use. Adv. Water Resour. 2009, 32, 129–146. [Google Scholar] [CrossRef]
  3. Beven, K. On the concept of model structural error. Water Sci. Technol. 2005, 52, 167–175. [Google Scholar] [CrossRef] [PubMed]
  4. Beven, K.; Freer, J. Equifinality, data assimilation, and uncertainty estimation in mechanistic modelling of complex environmental systems using the GLUE methodology. J. Hydrol. 2001, 249, 11–29. [Google Scholar] [CrossRef]
  5. Kuczera, G.; Parent, E. Monte Carlo assessment of parameter uncertainty in conceptual catchment models: The Metropolis algorithm. J. Hydrol. 1998, 211, 69–85. [Google Scholar] [CrossRef]
  6. Bormann, H. Sensitivity of a regionally applied soil vegetation atmosphere scheme to input data resolution and data classification. J. Hydrol. 2008, 351, 154–169. [Google Scholar] [CrossRef]
  7. Kavetski, D.; Kuczera, G.; Franks, S.W. Bayesian analysis of input uncertainty in hydrological modeling: 1. Theory. Water Resour. Res. 2006, 42, W03407. [Google Scholar] [CrossRef]
  8. Hämäläinen, R.P. Behavioural issues in environmental modelling—The missing perspective. Environ. Model. Softw. 2015, 73, 244–253. [Google Scholar] [CrossRef]
  9. Linkov, I.; Burmistrov, D. Model Uncertainty and Choices Made by Modelers: Lessons Learned from the International Atomic Energy Agency Model Intercomparisons. Risk Anal. 2003, 23, 1297–1308. [Google Scholar] [CrossRef]
  10. Holländer, H.M.; Blume, T.; Bormann, H.; Buytaert, W.; Chirico, G.B.; Exbrayat, J.-F.; Gustafsson, D.; Hölzel, H.; Kraft, P.; Stamm, C.; et al. Comparative predictions of discharge from an artificial catchment (Chicken Creek) using sparse data. Hydrol. Earth Syst. Sci. 2009, 13, 2069–2094. [Google Scholar] [CrossRef]
  11. Holländer, H.M.; Bormann, H.; Blume, T.; Buytaert, W.; Chirico, G.B.; Exbrayat, J.-F.; Gustafsson, D.; Hölzel, H.; Krauße, T.; Kraft, P.; et al. Impact of modellers’ decisions on hydrological a priori predictions. Hydrol. Earth Syst. Sci. 2014, 18, 2065–2085. [Google Scholar] [CrossRef]
  12. Bormann, H.; Holländer, H.M.; Blume, T.; Buytaert, W.; Chirico, G.B.; Exbrayat, J.-F.; Gustafsson, D.; Hölzel, H.; Kraft, P.; Krauße, T.; et al. Comparative discharge prediction from a small artificial catchment without model calibration: Representation of initial hydrological catchment development. Die Bodenkultur 2011, 62, 23–29. [Google Scholar]
  13. Molteni, F.; Buizza, R.; Palmer, T.N.; Petroliagis, T. The ECMWF Ensemble Prediction System. Methodology and validation. Q. J. R. Meteorol. Soc. 1996, 122, 73–119. [Google Scholar] [CrossRef]
  14. Ziehmann, C. Comparison of a single-model EPS with a multi-model ensemble consisting of a few operational models. Tellus 2000, 52A, 280–299. [Google Scholar] [CrossRef]
  15. Viney, N.R.; Bormann, H.; Breuer, L.; Bronstert, A.; Croke, B.F.W.; Frede, H.; Gräff, T.; Hubrechts, L.; Huisman, J.A.; Jakeman, A.J.; et al. Assessing the impact of land use change on hydrology by ensemble modelling (LUCHEM) II: Ensemble combinations and predictions. Adv. Water Resour. 2009, 32, 147–158. [Google Scholar] [CrossRef]
  16. Doblas-Reyes, F.J.; Hagedorn, R.; Palmer, T.N. The rationale behind the success of multi-model ensembles in seasonal forecasting—II Calibration and combination. Tellus 2005, 57A, 234–252. [Google Scholar] [CrossRef]
  17. Cloke, H.; Pappenberger, F. Ensemble flood forecasting: A review. J. Hydrol. 2009, 375, 613–626. [Google Scholar] [CrossRef]
  18. Woolley, A.W.; Chabris, C.F.; Pentland, A.; Hashmi, N.; Malone, T.W. Evidence for a collective intelligence factor in the performance of human groups. Science 2010, 330, 686–688. [Google Scholar] [CrossRef] [PubMed]
  19. Bergström, S. The HBV Model: Its Structure and Applications; Swedish Meteorological and Hydrological Institute (SMHI): Norrköpping, Sweden, 1992. [Google Scholar]
  20. Bergström, S. The HBV model (Chapter 13). In Computer Models of Watershed Hydrology; Singh, V.P., Ed.; Water Resources Publications: Highlands Ranch, CO, USA, 1995; pp. 443–476. [Google Scholar]
  21. Seibert, J.; Vis, M.J.P. Teaching hydrological modeling with a user-friendly catchment-runoff-model software package. Hydrol. Earth Syst. Sci. 2012, 16, 3315–3325. [Google Scholar] [CrossRef] [Green Version]
  22. Hessling, M. Hydrological modelling and a pair basin study of Mediterranean catchments. Phys. Chem. Earth (B) 1999, 24, 59–63. [Google Scholar] [CrossRef]
  23. Bormann, H.; Diekkrüger, B. A conceptual hydrological model for Benin (West Africa): Validation, uncertainty assessment and assessment of applicability for environmental change analyses. Phys. Chem. Earth 2004, 29, 759–768. [Google Scholar] [CrossRef]
  24. Nash, J.E.; Sutcliffe, J.V. River flow forecasting through conceptual models part I—A discussion of principles. J. Hydrol. 1970, 10, 282–290. [Google Scholar] [CrossRef]
  25. Moriasi, D.N.; Arnold, J.G.; Van Liew, M.W.; Bingner, R.L.; Harmel, R.D.; Veith, T.L. Model evaluation guidelines for systematic quantification of accuracy in watershed simulations. Trans. ASABE 2007, 50, 885–900. [Google Scholar] [CrossRef]
Figure 1. The Tsiknias River catchment (Left; upper Right: Greece; lower right: Island of Lesvos).
Figure 1. The Tsiknias River catchment (Left; upper Right: Greece; lower right: Island of Lesvos).
Hydrology 05 00064 g001
Figure 2. Variability of the discharges simulated by the different modellers against observations for an exemplary wet season (2014/2015) of the simulation period.
Figure 2. Variability of the discharges simulated by the different modellers against observations for an exemplary wet season (2014/2015) of the simulation period.
Hydrology 05 00064 g002
Figure 3. Variability of the discharges simulated by the different modellers against observations for three selected events in the rainy season of 2015.
Figure 3. Variability of the discharges simulated by the different modellers against observations for three selected events in the rainy season of 2015.
Hydrology 05 00064 g003
Figure 4. Flow duration curve based on observed (black) and simulated discharge for the whole simulation period.
Figure 4. Flow duration curve based on observed (black) and simulated discharge for the whole simulation period.
Hydrology 05 00064 g004
Figure 5. Dependence of model results (Nash-Sutcliffe model efficiency) on the modellers’ experience in relation to model application (Left) and the knowledge of the specific characteristics of intermittent catchments (Right). For graphical reasons, the results of one modeller are not shown who achieved an NSE value of −45.9 (little modelling experience, no experience with intermittent rivers).
Figure 5. Dependence of model results (Nash-Sutcliffe model efficiency) on the modellers’ experience in relation to model application (Left) and the knowledge of the specific characteristics of intermittent catchments (Right). For graphical reasons, the results of one modeller are not shown who achieved an NSE value of −45.9 (little modelling experience, no experience with intermittent rivers).
Hydrology 05 00064 g005
Figure 6. Dependence of model results (Nash-Sutcliffe model efficiency) on the calibration strategy (Left) and on whether the modellers’ enjoyed the modelling contest or not (Right). For graphical reasons, the results of one modeller are not shown who achieved an NSE value of −45.9 (trial and error calibration, mostly enjoyed the contest).
Figure 6. Dependence of model results (Nash-Sutcliffe model efficiency) on the calibration strategy (Left) and on whether the modellers’ enjoyed the modelling contest or not (Right). For graphical reasons, the results of one modeller are not shown who achieved an NSE value of −45.9 (trial and error calibration, mostly enjoyed the contest).
Hydrology 05 00064 g006
Figure 7. Comparison of individual model applications (blue diamonds) against model(er) ensembles (median, mean) based on all models (red squares) and on all models achieving satisfactory model results (yellow triangles) according to [25]. For graphical reasons, the results of one modeller are not shown who achieved an NSE value of −45.9.
Figure 7. Comparison of individual model applications (blue diamonds) against model(er) ensembles (median, mean) based on all models (red squares) and on all models achieving satisfactory model results (yellow triangles) according to [25]. For graphical reasons, the results of one modeller are not shown who achieved an NSE value of −45.9.
Hydrology 05 00064 g007
Figure 8. Comparison of individual model applications (blue diamonds) against model(er) ensembles (median, mean) based on the results of the experienced modellers (red squares) compared to the less or even unexperienced modellers (yellow triangles). For graphical reasons, the results of one modeller are not shown who achieved an NSE value of −45.9.
Figure 8. Comparison of individual model applications (blue diamonds) against model(er) ensembles (median, mean) based on the results of the experienced modellers (red squares) compared to the less or even unexperienced modellers (yellow triangles). For graphical reasons, the results of one modeller are not shown who achieved an NSE value of −45.9.
Hydrology 05 00064 g008
Figure 9. Variability of the discharges simulated by the investigated modellers’ ensembles against observations for three exemplary months of the rainy season of 2014/2015.
Figure 9. Variability of the discharges simulated by the investigated modellers’ ensembles against observations for three exemplary months of the rainy season of 2014/2015.
Hydrology 05 00064 g009
Table 1. Main calibration parameters of the HBV model.
Table 1. Main calibration parameters of the HBV model.
Model ParametersRelevanceIndicative Value
xWeighting factor related to the streams dimensions0.25
kStorage coefficient related to the stream dimensions0.55
HL1/ZParameter related to the elevation30.00–63.00
K0 0.45–10.50
K1Recession rate of the various reservoirs parameters10.70–20.70
K2 30.5
PERCThe maximum percolation rate from the upper to the lower groundwater box1.00
MAXBASRouting parameter of the triangular weighting function5.00–10.00
FCMaximum soil moisture storage800–1000
BETASoil parameter2.00–10.00
Table 2. Statistical analysis of the goodness-of-fit criteria of the best simulations of each modeller.
Table 2. Statistical analysis of the goodness-of-fit criteria of the best simulations of each modeller.
Goodness of Fit IndexBest Simulation75%-Percentile50%-Percentile25%-PercentileWorst Simulation
Nash-Stutcliffe efficiency (NSE)0.560.530.470.35−45.9
Coefficient of determination0.560.550.510.420.082
Percent bias (PB)3.5%14.6%24.9%31.2%526%
Root mean squared error (RMSE)0.67 mm/d0.69 mm/d0.73 mm/d0.81 mm/d6.85 mm/d
The absolute bias (PB) of the simulated discharge ranged from 3.5% (of the best simulation) up to 526% (of the worst simulation).
Table 3. Parameter ranges of calibrated parameters of the HBV model.
Table 3. Parameter ranges of calibrated parameters of the HBV model.
Model ParametersMinimum ValueMedian ValueMaximum Value
HL1/Z0.00040,03765
K00.00950.218
K10.0020.23.6
K20.0010.11
PERC0.0010.023
MAXBAS4798
FC1502501140
BETA1210

Share and Cite

MDPI and ACS Style

Bormann, H.; De Brito, M.M.; Charchousi, D.; Chatzistratis, D.; David, A.; Grosser, P.F.; Kebschull, J.; Konis, A.; Koutalakis, P.; Korali, A.; et al. Impact of Hydrological Modellers’ Decisions and Attitude on the Performance of a Calibrated Conceptual Catchment Model: Results from a ‘Modelling Contest’. Hydrology 2018, 5, 64. https://doi.org/10.3390/hydrology5040064

AMA Style

Bormann H, De Brito MM, Charchousi D, Chatzistratis D, David A, Grosser PF, Kebschull J, Konis A, Koutalakis P, Korali A, et al. Impact of Hydrological Modellers’ Decisions and Attitude on the Performance of a Calibrated Conceptual Catchment Model: Results from a ‘Modelling Contest’. Hydrology. 2018; 5(4):64. https://doi.org/10.3390/hydrology5040064

Chicago/Turabian Style

Bormann, Helge, Mariana Madruga De Brito, Despoina Charchousi, Dimitris Chatzistratis, Amrei David, Paula Farina Grosser, Jenny Kebschull, Alexandros Konis, Paschalis Koutalakis, Alkistis Korali, and et al. 2018. "Impact of Hydrological Modellers’ Decisions and Attitude on the Performance of a Calibrated Conceptual Catchment Model: Results from a ‘Modelling Contest’" Hydrology 5, no. 4: 64. https://doi.org/10.3390/hydrology5040064

APA Style

Bormann, H., De Brito, M. M., Charchousi, D., Chatzistratis, D., David, A., Grosser, P. F., Kebschull, J., Konis, A., Koutalakis, P., Korali, A., Krauzig, N., Meier, J., Meliadou, V., Meinhardt, M., Munnelly, K., Stephan, C., De Vos, L. F., Dietrich, J., & Tzoraki, O. (2018). Impact of Hydrological Modellers’ Decisions and Attitude on the Performance of a Calibrated Conceptual Catchment Model: Results from a ‘Modelling Contest’. Hydrology, 5(4), 64. https://doi.org/10.3390/hydrology5040064

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop