Source code for hydpy.core.hydpytools

# -*- coding: utf-8 -*-
"""This module implements the main features for managing *HydPy* projects.

.. _`NetCDF Climate and Forecast (CF) Metadata Conventions`: http://cfconventions.org/Data/cf-conventions/cf-conventions-1.7/cf-conventions.html  # pylint: disable=line-too-long
"""
# import...
# ...from standard library
from __future__ import annotations
import collections
import contextlib
import itertools
import warnings

# ...from site-packages
import networkx

# ...from HydPy
import hydpy
from hydpy.core import devicetools
from hydpy.core import exceptiontools
from hydpy.core import filetools
from hydpy.core import modeltools
from hydpy.core import objecttools
from hydpy.core import printtools
from hydpy.core import propertytools
from hydpy.core import selectiontools
from hydpy.core import sequencetools
from hydpy.core import timetools
from hydpy.core.typingtools import *

if TYPE_CHECKING:
    from hydpy.core import auxfiletools


[docs] class HydPy: """The main class for managing *HydPy* projects. In typical *HydPy* projects, one prepares a single instance of class |HydPy|. This instance, which we name "hp" throughout this documentation instead of "hydpy" to avoid a naming collision with the `hydpy` site package, provides many convenient methods to perform tasks like reading time series data or starting simulation runs. Additionally, it serves as a root to access most details of a *HydPy* project, allowing for more granular control over the framework features. We elaborate on these short explanations by using the :ref:`HydPy-H-Lahn` example project. Calling function |prepare_full_example_1| copies the complete example project :ref:`HydPy-H-Lahn` into the `iotesting` directory of the *HydPy* site package (alternatively, you can copy the `HydPy-H-Lahn` example project, which can be found in subpackage `data`, into a working directory of your choice): >>> from hydpy.core.testtools import prepare_full_example_1 >>> prepare_full_example_1() At first, the |HydPy| instance needs to know the name of the relevant project, which is identical to the name of the project's root directory. Pass `HydPy-H-Lahn` to the constructor of class |HydPy|: >>> from hydpy import HydPy >>> hp = HydPy("HydPy-H-Lahn") So far, our |HydPy| instance does not know any project configurations except its name. Most of this information would be available via properties |HydPy.nodes| and |HydPy.elements|, but we get the following error responses if we try to access them: >>> hp.nodes Traceback (most recent call last): ... hydpy.core.exceptiontools.AttributeNotReady: The actual HydPy instance does not \ handle any nodes at the moment. >>> hp.elements Traceback (most recent call last): ... hydpy.core.exceptiontools.AttributeNotReady: The actual HydPy instance does not \ handle any elements at the moment. One could continue rather quickly by calling the method |HydPy.prepare_everything|, which would make our |HydPy| instance ready for its first simulation run in one go. However, we prefer to continue step by step by calling the more specific preparation methods, which offers more flexibility. First, the |HydPy| instance needs to know the relevant |Node| and |Element| objects. Method |HydPy.prepare_network| reads this information from so-called "network files". Then, the |Node| and |Element| objects connect automatically and thereby define the topology or the network structure of the project (see the documentation on class |NetworkManager| and module |devicetools| for more detailed explanations): >>> from hydpy import TestIO >>> with TestIO(): ... hp.prepare_network() (Using the "with" statement in combination with class |TestIO| makes sure we are reading the network files from a subdirectory of the `iotesting` directory. Here and in the following, you must omit such "with blocks" in case you copied the :ref:`HydPy-H-Lahn` example project into your current working directory.) Now, our |HydPy| instance offers access to all |Node| objects defined within the :ref:`HydPy-H-Lahn` example project, which are grouped by a |Nodes| object: >>> hp.nodes Nodes("dill_assl", "lahn_kalk", "lahn_leun", "lahn_marb") Taking the node `dill_assl` as an example, we can dive into the details and, for example, search for those elements to which node `dill_assl` is connected (it receives water from element `land_dill_assl` and passes it to element `stream_dill_assl_lahn_leun`), or inspect its simulated discharge value handled by a |Sim| sequence object (so far, zero): >>> hp.nodes.dill_assl.entries Elements("land_dill_assl") >>> hp.nodes.dill_assl.exits Elements("stream_dill_assl_lahn_leun") >>> hp.nodes.dill_assl.sequences.sim sim(0.0) All |Node| objects are ready to be used. The same is only partly true for the |Element| objects, which are also accessible (via a |Elements| instance) and properly connected to the |Node| objects but do not handle workable |Model| objects, which is required to perform any simulation run: >>> hp.elements Elements("land_dill_assl", "land_lahn_kalk", "land_lahn_leun", "land_lahn_marb", "stream_dill_assl_lahn_leun", "stream_lahn_leun_lahn_kalk", "stream_lahn_marb_lahn_leun") >>> hp.elements.stream_dill_assl_lahn_leun Element("stream_dill_assl_lahn_leun", inlets="dill_assl", outlets="lahn_leun", keywords="river") >>> hp.elements.land_dill_assl.model Traceback (most recent call last): ... hydpy.core.exceptiontools.AttributeNotReady: The model object of element \ `land_dill_assl` has been requested but not been prepared so far. Hence, we need to call method |HydPy.prepare_models|, which instructs all |Element| objects to read the relevant parameter control files and prepare their |Model| objects. Note that the individual |Element| object does not know the relevant model type beforehand; both the information on the model type and the parameter settings is encoded in individual control files, making it easy to exchange individual models later (the documentation on method |Elements.prepare_models| of class |Elements| is a good starting point for a deeper understanding on configuring *HydPy* projects via control files): >>> with TestIO(): ... hp.prepare_models() Traceback (most recent call last): ... hydpy.core.exceptiontools.AttributeNotReady: While trying to initialise the model \ object of element `land_dill_assl`, the following error occurred: The initialisation \ period has not been defined via attribute `timegrids` of module `pub` yet but might be \ required to prepare the model properly. Oops, something went wrong. We forgot to define the simulation period, which might be relevant for some time-dependent configurations. We discuss some examples of such configurations below but now use this little accident to discuss the typical pattern of *HydPy* error messages. First, we usually try to add some additional "spatial" information (in this case: the name of the related |Element| object). Second, we try to explain in which program context an error occurs. This context is already available in much more detail in the so-called "stack trace" (the middle part of the printed error response we do not show). Stack trace descriptions are great for programmers but hard to read for others, which is why we often add "While trying to..." explanations to our error messages. In our example, one can see that the error occurred while trying to initialise the |Model| object of element `land_dill_assl`, which is quite evident in our example but could be less evident in more complex *HydPy* applications. The last sentence of the error message tells us that we need to define the attribute `timegrids` of module `pub`. `pub` stands for "public", meaning module `pub` handles all (or at least most of) the globally available configuration data. One example is that module `pub` handles a |Timegrids| instance defining both the initialisation and the simulation period, which can be done by the following assignment (see the documentation on class |Timegrid| and on class |Timegrids| for further information): >>> from hydpy import pub >>> pub.timegrids = "1996-01-01", "1996-01-05", "1d" Now, method |HydPy.prepare_models| does not complain anymore and adds an instance of the |hland_96| application model to element `land_dill_assl`, to which we set an additional reference to shorten the following examples: >>> with TestIO(): ... hp.prepare_models() >>> model = hp.elements.land_dill_assl.model >>> model.name 'hland_96' All control parameter values, defined in the corresponding control file, are correctly set. As an example, we show the values of the control parameter |hland_control.IcMax|, which in this case defines different values for hydrological response units of type |hland_constants.FIELD| (1.0 mm) and of type |hland_constants.FOREST| (1.5 mm): >>> model.parameters.control.icmax icmax(field=1.0, forest=1.5) The appearance (or "string representation") of all parameters that have a unit with a time reference (we call these parameters "time-dependent") like |hland_control.PercMax| depends on the current setting of option |Options.parameterstep|, which is one day by default (see the documentation on class |Parameter| for more information on dealing with time-dependent parameters subclasses): >>> model.parameters.control.percmax percmax(1.39636) >>> pub.options.parameterstep("1h") Period("1d") The values of the derived parameters, which need to be calculated before starting a simulation run based on the control parameters and eventually based on some other settings (e.g. the initialisation period), are also ready. Here we show the value of the derived parameter |hland_derived.Z|, representing the average (reference) subbasin elevation: >>> model.parameters.derived.z z(4.205345) We define all class names in "CamelCase" letters (which is a Python convention) and, whenever practical, name the related objects identically but in lowercase letters. We hope that eases finding the relevant parts of the online documentation when in trouble with a particular object. Three examples we already encountered are the |Timegrids| instance `timegrids` of module `pub`, the |Nodes| instance `nodes` of class `HydPy`, and the |hland_derived.Z| instance `z` of application model |hland_96|: >>> from hydpy import classname >>> classname(pub.timegrids) 'Timegrids' >>> classname(hp.nodes) 'Nodes' >>> classname(model.parameters.derived.z) 'Z' As shown above, all |Parameter| objects of the model of element `land_dill_assl` are ready to be used. However, all sequences (which handle the time variable properties) contain |numpy| |numpy.nan| values, which we use to indicate missing data. We show this for the 0-dimensional input sequence |hland_inputs.T|, the 1-dimensional factor sequence |hland_factors.TC|, the 1-dimensional state sequence |hland_states.SM|, and the 0-dimensional flux sequence |hland_fluxes.QT|: >>> model.sequences.inputs.t t(nan) >>> model.sequences.factors.tc tc(nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan) >>> model.sequences.states.sm sm(nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan) >>> model.sequences.fluxes.qt qt(nan) There are some other sequence types (see the documentation on module |sequencetools| for more details) but |InputSequence|, |FactorSequence| |FluxSequence|, and |StateSequence| are the most common ones (besides the |NodeSequence| subtypes |Obs| and especially |Sim|). |StateSequence| objects describe many aspects of the current state of a model (or, e.g., of a catchment). Each simulation run requires proper initial states, which we call initial conditions in the following (also covering memory aspects represented by |LogSequence| objects). We load all necessary initial conditions by calling the method |HydPy.load_conditions| (see the documentation on method |HydPy.load_conditions| for further details): >>> with TestIO(): ... hp.load_conditions() Now, the states of our model are also ready to be used. However, one should note that state sequence |hland_states.SM| knows only the current soil moisture states for the twelve hydrological response units of element `land_dill_assl` (more specifically, we loaded the soil moisture values related to the start date of the initialisation period, which is January 1 at zero o'clock). By default and for reasons of memory storage efficiency, sequences generally handle the currently relevant values only instead of complete time series: >>> model.sequences.inputs.t t(nan) >>> model.sequences.factors.tc tc(nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan) >>> model.sequences.states.sm sm(185.13164, 181.18755, 199.80432, 196.55888, 212.04018, 209.48859, 222.12115, 220.12671, 230.30756, 228.70779, 236.91943, 235.64427) >>> model.sequences.fluxes.qt qt(nan) For states like |hland_states.SM|, we need to know the values at the beginning of the simulation period only. All following values are calculated subsequentially during the simulation run. However, this is different for input sequences like |hland_inputs.T|. Time variable properties like the air temperature are external forcings. Hence, they must be available over the whole simulation period apriori. Such complete time series can be made available via property |IOSequence.series| of class |IOSequence|, which has not happened for any sequence so far: >>> model.sequences.inputs.t.series Traceback (most recent call last): ... hydpy.core.exceptiontools.AttributeNotReady: Sequence `t` of element \ `land_dill_assl` is not requested to make any time series data available. Before loading time series data, we need to reserve the required memory storage. We do this for all sequences at once (not only the |ModelSequence| objects but also the |NodeSequence| objects as the |Sim| instance handled by node `dill_assl`) by calling the method |HydPy.prepare_allseries|: >>> hp.prepare_allseries() Now, property |IOSequence.series| returns an |InfoArray| object, which is a slight modification of the widely applied |numpy| |numpy.ndarray|. The first axis (or the only axis) corresponds to the number of days of the initialisation period (a *HydPy* convention). For the 1-dimensional sequences |hland_factors.TC| and |hland_states.SM|, the second axis corresponds to the number of hydrological response units (a |hland| convention): >>> model.sequences.inputs.t.series InfoArray([nan, nan, nan, nan]) >>> model.sequences.factors.tc.series InfoArray([[nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan], [nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan], [nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan], [nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan]]) >>> model.sequences.states.sm.series InfoArray([[nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan], [nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan], [nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan], [nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan, nan]]) >>> model.sequences.fluxes.qt.series InfoArray([nan, nan, nan, nan]) >>> hp.nodes.dill_assl.sequences.sim.series InfoArray([nan, nan, nan, nan]) So far, each time series array is empty. The :ref:`HydPy-H-Lahn` example project provides time series files for the input sequences, which is the minimum requirement for starting a simulation run. We use method |HydPy.load_inputseries| to load this data: >>> with TestIO(): ... hp.load_inputseries() >>> from hydpy import round_ >>> round_(model.sequences.inputs.t.series) 0.0, -0.5, -2.4, -6.8 Finally, we can perform the simulation run by calling the method |HydPy.simulate|: >>> hp.simulate() The time series arrays of all sequences now contain calculated values --- except those of input sequence |hland_inputs.T|, of course (for the sequences |hland_factors.TC| and |hland_states.SM|, we show the time series of the first hydrological response unit only): >>> round_(model.sequences.inputs.t.series) 0.0, -0.5, -2.4, -6.8 >>> round_(model.sequences.factors.tc.series[:, 0]) 1.323207, 0.823207, -1.076793, -5.476793 >>> round_(model.sequences.states.sm.series[:, 0]) 184.958475, 184.763638, 184.610776, 184.553224 >>> round_(model.sequences.fluxes.qt.series) 11.757526, 8.865079, 7.101815, 5.994195 >>> round_(hp.nodes.dill_assl.sequences.sim.series) 11.757526, 8.865079, 7.101815, 5.994195 By comparison, you see that the lastly calculated (or read) time series value is the actual one for each |Sequence_| object. This mechanism allows, for example, to write the final states of soil moisture sequence |hland_states.SM| and use them as initial conditions later, even if its complete time series were not available: >>> model.sequences.inputs.t t(-6.8) >>> model.sequences.states.sm sm(184.553224, 180.623124, 199.181137, 195.947542, 212.04018, 209.48859, 222.12115, 220.12671, 230.30756, 228.70779, 236.91943, 235.64427) >>> model.sequences.fluxes.qt qt(5.994195) >>> hp.nodes.dill_assl.sequences.sim sim(5.994195) In many applications, the simulated time series is the result we are interested in. Hence, we close our explanations with some detailed examples on this topic that also cover the potential problem of limited rapid access storage availability. The *HydPy* framework does not overwrite already existing time series by default files. However, you can change this and related settings via the |SequenceManager| object available in module |pub| (module |pub| also handles |ControlManager| and |ConditionManager| objects for settings related to reading and writing control files and condition files). We change the default behaviour by setting the |SequenceManager.overwrite| attribute to |True|: >>> pub.sequencemanager.overwrite = True Now, we can (over)write all possible time series: >>> with TestIO(): ... hp.save_inputseries() ... hp.save_factorseries() ... hp.save_fluxseries() ... hp.save_stateseries() ... hp.save_simseries() ... hp.save_obsseries() Alternatively, apply |HydPy.save_modelseries| to write the series of all the |InputSequence|, |FactorSequence|, |FluxSequence|, and |StateSequence| objects and |HydPy.save_nodeseries| to write the series of all |Sim| and |Obs| objects in one step: >>> with TestIO(): ... hp.save_modelseries() ... hp.save_nodeseries() Even shorter, just apply the method |HydPy.save_allseries|: >>> with TestIO(): ... hp.save_allseries() Next, we show how the reading of time series works. We first set the time series values of all considered sequences to zero for this purpose: >>> model.sequences.inputs.t.series = 0.0 >>> model.sequences.states.sm.series = 0.0 >>> model.sequences.inputs.t.series = 0.0 >>> hp.nodes.dill_assl.sequences.sim.series = 0.0 Now, we can reload the time series of all relevant sequences. However, doing so would result in a warning due to incomplete data (for example, of the observation data handled by the |Obs| sequence objects). To circumvent this problem, we turn off the |Options.checkseries| option, which is one of the public options handled by the instance of class |Options| available as another attribute of module |pub|. We again use "with blocks", making sure the option (and the current working directory) changes only temporarily while loading the time series: >>> with TestIO(), pub.options.checkseries(False): ... hp.load_inputseries() ... hp.load_factorseries() ... hp.load_fluxseries() ... hp.load_stateseries() ... hp.load_simseries() ... hp.load_obsseries() >>> with TestIO(), pub.options.checkseries(False): ... hp.load_modelseries() ... hp.load_nodeseries() >>> with TestIO(), pub.options.checkseries(False): ... hp.load_allseries() The read time series data equals the previously written one: >>> round_(model.sequences.inputs.t.series) 0.0, -0.5, -2.4, -6.8 >>> round_(model.sequences.factors.tc.series[:, 0]) 1.323207, 0.823207, -1.076793, -5.476793 >>> round_(model.sequences.states.sm.series[:, 0]) 184.958475, 184.763638, 184.610776, 184.553224 >>> round_(model.sequences.fluxes.qt.series) 11.757526, 8.865079, 7.101815, 5.994195 >>> round_(hp.nodes.dill_assl.sequences.sim.series) 11.757526, 8.865079, 7.101815, 5.994195 We mentioned the possibility for more granular control of *HydPy* by using the different objects handled by the |HydPy| object instead of using its convenience methods. Here is an elaborate example showing how to (re)load the states of an arbitrary simulation time step, which might be relevant for more complex workflows implementing data assimilation techniques: >>> model.sequences.states.load_data(1) >>> model.sequences.states.sm sm(184.763638, 180.829058, 199.40823, 196.170947, 212.04018, 209.48859, 222.12115, 220.12671, 230.30756, 228.70779, 236.91943, 235.64427) Using the node sequence |Sim| as an example, we also show the inverse functionality of changing time series values: >>> hp.nodes.dill_assl.sequences.sim = 0.0 >>> hp.nodes.dill_assl.sequences.save_data(2) >>> round_(hp.nodes.dill_assl.sequences.sim.series) 11.757526, 8.865079, 0.0, 5.994195 >>> hp.nodes.dill_assl.sequences.load_data(1) >>> hp.nodes.dill_assl.sequences.sim sim(8.865079) In the examples above, we keep all data in rapid access memory, which can be problematic when handling long time series in huge *HydPy* projects. When in trouble, first try to prepare only those time series that are strictly required (very often, it is sufficient to call |HydPy.prepare_inputseries|, |HydPy.load_inputseries|, and |HydPy.prepare_simseries| only). If this does not work in your project, you can read input data from and write output data to NetCDF files during simulation. These follow the `NetCDF Climate and Forecast (CF) Metadata Conventions`_. To benefit from this feature, assign |False| to the `allocate_ram` argument of the individual "prepare series" methods (which disables handling the time series in RAM) and assign |True| to the respective "jit" arguments (which prepares the "just-in-time" file access). The methods |HydPy.prepare_factorseries|, |HydPy.prepare_fluxseries|, and |HydPy.prepare_stateseries| deal with "output sequences" for which read data would be overwritten during the simulation and thus only support the `write_jit` argument. The |HydPy.prepare_inputseries| method, on the other hand, supports both the `read_jit` and the `write_jit` argument. However, in most cases, only reading makes sense. The argument `write_jit` is thought for when other methods (for example data assimilation approaches) modify the input data, and we need to keep track of these modifications: >>> hp.prepare_inputseries(allocate_ram=False, read_jit=True) >>> hp.prepare_factorseries(allocate_ram=False, write_jit=True) >>> hp.prepare_fluxseries(allocate_ram=False, write_jit=True) >>> hp.prepare_stateseries(allocate_ram=False, write_jit=True) >>> hp.prepare_simseries(allocate_ram=False, write_jit=True) >>> hp.prepare_obsseries(allocate_ram=False, read_jit=True) By doing so, you lose the previously available time series information. We use function |attrready| to check this: >>> from hydpy import attrready >>> attrready(model.sequences.inputs.t, "series") False >>> attrready(model.sequences.factors.tc, "series") False >>> attrready(model.sequences.states.sm, "series") False >>> attrready(model.sequences.fluxes.qt, "series") False >>> attrready(hp.nodes.dill_assl.sequences.sim, "series") False Reloading the initial conditions and starting a new simulation run leads to the same results as the simulation run above: >>> with TestIO(), pub.options.checkseries(False): ... hp.load_conditions() ... hp.simulate() This time, reading input data from files happened during simulation. Likewise, the calculated output data is not directly available in RAM but in different NetCDF files. To check all results are identical to those shown above, we must load them into RAM. Therefore, we first need to prepare the |IOSequence.series| objects again: >>> hp.prepare_allseries() By default, *HydPy* handles time series data in simple text files ("asc" files): >>> pub.sequencemanager.filetype 'asc' One way to prepare to load the results from the available NetCDF files instead is to set the |SequenceManager.filetype| attribute of the public |SequenceManager| object to "nc": >>> pub.sequencemanager.filetype = "nc" We can load the previously written results into RAM (see the documentation on module |netcdftools| for further information) and inspect the results: >>> with TestIO(): ... hp.load_modelseries() ... hp.load_simseries() >>> round_(model.sequences.inputs.t.series) 0.0, -0.5, -2.4, -6.8 >>> round_(model.sequences.factors.tc.series[:, 0]) 1.323207, 0.823207, -1.076793, -5.476793 >>> round_(model.sequences.states.sm.series[:, 0]) 184.958475, 184.763638, 184.610776, 184.553224 >>> round_(model.sequences.fluxes.qt.series) 11.757526, 8.865079, 7.101815, 5.994195 >>> round_(hp.nodes.dill_assl.sequences.sim.series) 11.757526, 8.865079, 7.101815, 5.994195 You can handle time series in RAM and allow just-in-time NetCDF file access at the same time. Before showing how this works, we first disable both functionalities for all sequences and delete all previously written NetCDF files: >>> hp.prepare_allseries(allocate_ram=False) >>> attrready(model.sequences.inputs.t, "series") False >>> attrready(model.sequences.factors.tc, "series") False >>> attrready(model.sequences.states.sm, "series") False >>> attrready(model.sequences.fluxes.qt, "series") False >>> attrready(hp.nodes.dill_assl.sequences.sim, "series") False >>> import os >>> with TestIO(): ... for filename in os.listdir(f"HydPy-H-Lahn/series/default"): ... if "input" not in filename: ... os.remove(f"HydPy-H-Lahn/series/default/{filename}") We again call method |HydPy.prepare_allseries|, but now, by assigning |True| to the arguments `allocate_ram` and `jit`: >>> hp.prepare_allseries(allocate_ram=True, jit=True) After another simulation run, all input data (read during simulation) and output data (calculated during simulation) are directly available: >>> with TestIO(), pub.options.checkseries(False): ... hp.load_conditions() ... hp.simulate() >>> round_(model.sequences.inputs.t.series) 0.0, -0.5, -2.4, -6.8 >>> round_(model.sequences.factors.tc.series[:, 0]) 1.323207, 0.823207, -1.076793, -5.476793 >>> round_(model.sequences.states.sm.series[:, 0]) 184.958475, 184.763638, 184.610776, 184.553224 >>> round_(model.sequences.fluxes.qt.series) 11.757526, 8.865079, 7.101815, 5.994195 >>> round_(hp.nodes.dill_assl.sequences.sim.series) 11.757526, 8.865079, 7.101815, 5.994195 After subsequent deallocation and allocation for refreshing RAM, reading the previously written NetCDF files makes the same data available: >>> hp.prepare_allseries(allocate_ram=False) >>> hp.prepare_allseries(allocate_ram=True) >>> with TestIO(): ... hp.load_modelseries() ... hp.load_simseries() >>> round_(model.sequences.inputs.t.series) 0.0, -0.5, -2.4, -6.8 >>> round_(model.sequences.factors.tc.series[:, 0]) 1.323207, 0.823207, -1.076793, -5.476793 >>> round_(model.sequences.states.sm.series[:, 0]) 184.958475, 184.763638, 184.610776, 184.553224 >>> round_(model.sequences.fluxes.qt.series) 11.757526, 8.865079, 7.101815, 5.994195 >>> round_(hp.nodes.dill_assl.sequences.sim.series) 11.757526, 8.865079, 7.101815, 5.994195 All filenames of meteorological input time series provided by the :ref:`HydPy-H-Lahn` example follow a "model-specific" pattern. Each filename contains not only the name of the corresponding |InputSequence| subclass (e.g., "t") in lowercase letters but also the sequence's group (e.g., "input") and the responsible model's name (e.g., "hland_96"): >>> old_model = hp.elements.land_dill_assl.model >>> old_model.sequences.inputs.t.filename 'hland_96_input_t.nc' For file types that store the time series of single sequence instances, file names must also contain information about the location. Therefore, the relevant element's name prefixes the pure model-specific pattern: >>> pub.sequencemanager.filetype = "asc" >>> old_model.sequences.inputs.t.filename 'land_dill_assl_hland_96_input_t.asc' This naming pattern is exact but not always convenient. For example, assume we want to simulate the `dill_assl` subcatchment with application model |hland_96p| instead of |hland_96|: >>> from hydpy import prepare_model >>> hp.elements.land_dill_assl.model = new_model = prepare_model("hland_96p") |hland_96p| requires the same meteorological input as |hland_96|, for example, the air temperature: >>> new_model.prepare_inputseries() >>> round_(new_model.sequences.inputs.t.series) nan, nan, nan, nan Reading these data from the available files does not work because of the described filename convention: >>> with TestIO(): ... hp.load_inputseries() # doctest: +ELLIPSIS Traceback (most recent call last): ... FileNotFoundError: While trying to load the time series data of sequence `p` of \ element `land_dill_assl`, the following error occurred: [Errno 2] No such file or \ directory: '...land_dill_assl_hland_96p_input_p.asc' To circumvent this problem, one can use the standard "HydPy" convention instead of the "model-specific" convention when reading or writing input sequences. >>> pub.sequencemanager.convention = "HydPy" Then, the standard names defined by the |StandardInputNames| enum replace the model-specific filename parts. For example, the input sequence |hland_inputs.T| selects |StandardInputNames.AIR_TEMPERATURE| as its standard name: >>> with pub.sequencemanager.filetype("nc"): ... old_model.sequences.inputs.t.filename 'air_temperature.nc' >>> old_model.sequences.inputs.t.filename 'land_dill_assl_air_temperature.asc' If we use the |hland_96| instance to store the meteorological time series based on the "HydPy" convention, the |hland_96p| instance can load them without further effort: >>> hp.elements.land_dill_assl.model = old_model >>> with TestIO(): ... hp.save_inputseries() ... hp.elements.land_dill_assl.model = new_model ... hp.load_inputseries() >>> round_(old_model.sequences.inputs.t.series) 0.0, -0.5, -2.4, -6.8 >>> round_(new_model.sequences.inputs.t.series) 0.0, -0.5, -2.4, -6.8 The standard "HydPy" naming convention is compatible with reading data "just-in-time" from NetCDF files, even if main models and submodels come with different input sequences that require the same data. Before explaining this, we restore the original |hland_96| submodel and write all input time series to NetCDF files following the standard naming convention: >>> with TestIO(): ... hp.elements.land_dill_assl.prepare_model() ... hp.elements.land_dill_assl.model.load_conditions() ... hp.elements.land_dill_assl.prepare_inputseries() ... hp.elements.land_dill_assl.load_inputseries() ... with pub.sequencemanager.filetype("nc"): ... hp.save_inputseries() Now, instead of letting the |evap_pet_hbv96| sub-submodel query the air temperature from the |hland_96| main model, we add a |meteo_temp_io| sub-sub-submodel that comes with an independent air temperature sequence >>> hland = hp.elements.land_dill_assl.model >>> with hland.aetmodel.petmodel.add_tempmodel_v2("meteo_temp_io"): ... temperatureaddend(0.0) >>> hp.prepare_inputseries(allocate_ram=True, read_jit=True) >>> round_(hland.aetmodel.petmodel.tempmodel.sequences.inputs.temperature.series) nan, nan, nan, nan Reading data "just-in-time" makes the same data of the "air_temperature.nc" file available to both sequences (and leads to the same simulation results): >>> hland.sequences.inputs.t.series = -777.0 >>> with TestIO(): ... hp.prepare_fluxseries() ... hp.simulate() >>> round_(hland.sequences.inputs.t.series) 0.0, -0.5, -2.4, -6.8 >>> round_(hland.aetmodel.petmodel.tempmodel.sequences.inputs.temperature.series) 0.0, -0.5, -2.4, -6.8 >>> round_(hland.sequences.fluxes.qt.series) 11.757526, 8.865079, 7.101815, 5.994195 """ _deviceorder: Optional[tuple[Union[devicetools.Node, devicetools.Element], ...]] _nodes: Optional[devicetools.Nodes] _elements: Optional[devicetools.Elements] _collectives: Optional[devicetools.Elements] def __init__(self, projectname: Optional[str] = None) -> None: self._nodes = None self._elements = None self._collectives = None self._deviceorder = None if projectname is not None: hydpy.pub.projectname = projectname hydpy.pub.networkmanager = filetools.NetworkManager() hydpy.pub.controlmanager = filetools.ControlManager() hydpy.pub.sequencemanager = filetools.SequenceManager() hydpy.pub.conditionmanager = filetools.ConditionManager() nodes = propertytools.Property[devicetools.NodesConstrArg, devicetools.Nodes]() @nodes.getter def _get_nodes(self) -> devicetools.Nodes: """The currently handled |Node| objects. You are allowed to get, set and delete the currently handled nodes: >>> from hydpy.core.testtools import prepare_full_example_2 >>> hp, pub, TestIO = prepare_full_example_2() >>> hp.nodes Nodes("dill_assl", "lahn_kalk", "lahn_leun", "lahn_marb") >>> del hp.nodes >>> hp.nodes Traceback (most recent call last): ... hydpy.core.exceptiontools.AttributeNotReady: The actual HydPy instance does \ not handle any nodes at the moment. >>> hp.nodes = "dill_assl", "lahn_marb" >>> hp.nodes Nodes("dill_assl", "lahn_marb") However, note that doing so might result in erroneous networks and that you, even in case of correctness, must most likely call method |HydPy.update_devices| before performing the next simulation run. """ nodes = self._nodes if nodes is None: raise exceptiontools.AttributeNotReady( "The actual HydPy instance does not handle any nodes at the moment." ) return nodes @nodes.setter def _set_nodes(self, values: devicetools.NodesConstrArg) -> None: self._nodes = devicetools.Nodes(values).copy() @nodes.deleter def _del_nodes(self) -> None: self._nodes = None elements = propertytools.Property[ devicetools.ElementsConstrArg, devicetools.Elements ]() @elements.getter def _get_elements(self) -> devicetools.Elements: """The currently handled |Element| objects. You are allowed to get, set and delete the currently handled elements: >>> from hydpy.core.testtools import prepare_full_example_2 >>> hp, pub, TestIO = prepare_full_example_2() >>> hp.elements Elements("land_dill_assl", "land_lahn_kalk", "land_lahn_leun", "land_lahn_marb", "stream_dill_assl_lahn_leun", "stream_lahn_leun_lahn_kalk", "stream_lahn_marb_lahn_leun") >>> del hp.elements >>> hp.elements Traceback (most recent call last): ... hydpy.core.exceptiontools.AttributeNotReady: The actual HydPy instance does \ not handle any elements at the moment. >>> hp.elements = "land_dill_assl", "land_lahn_marb" >>> hp.elements Elements("land_dill_assl", "land_lahn_marb") However, note that doing so might result in erroneous networks and that you, even in case of correctness, must most likely call method |HydPy.update_devices| before performing the next simulation run. """ elements = self._elements if elements is None: raise exceptiontools.AttributeNotReady( "The actual HydPy instance does not handle any elements at the moment." ) return elements @elements.setter def _set_elements(self, values: devicetools.ElementsConstrArg) -> None: self._elements = devicetools.Elements(values).copy() @elements.deleter def _del_elements(self) -> None: self._elements = None @property def collectives(self) -> devicetools.Elements: """The elements relevant for simulation. Usually, |HydPy.collectives| returns the "original" elements also available via property |HydPy.elements|. However, if some of these elements belong to a |Element.collective|, |HydPy.collectives| does not return them but overarching elements suitable for simulation. |HydPy.collectives| raises the following error if the overarching elements are unavailable: >>> from hydpy import HydPy >>> HydPy().collectives Traceback (most recent call last): ... RuntimeError: The currently available elements have not been analysed and \ eventually combined regarding their collectives. Please call method `update_devices` \ first. """ if (collectives := self._collectives) is None: raise RuntimeError( "The currently available elements have not been analysed and " "eventually combined regarding their collectives. Please call method " "`update_devices` first." ) return collectives
[docs] def prepare_everything(self) -> None: """Convenience method to make the actual |HydPy| instance runnable. Method |HydPy.prepare_everything| is the fastest approach to get a runnable |HydPy| object. You only need to import class |Hydpy|, initialise it with the project name, define the simulation period via the |Timegrids| object of module |pub|, and call method |HydPy.prepare_everything| (in this documentation, we first need to prepare the example project via function |prepare_full_example_1| and change the current working directory via class |TestIO|): >>> from hydpy.core.testtools import prepare_full_example_1 >>> prepare_full_example_1() >>> from hydpy import HydPy, print_vector, pub, TestIO >>> with TestIO(): ... hp = HydPy("HydPy-H-Lahn") ... pub.timegrids = "1996-01-01", "1996-01-05", "1d" ... hp.prepare_everything() Now, you can start a simulation run and inspect the calculated time series of all relevant sequences. We take the discharge values of the flux sequence |hland_fluxes.QT| of |Element| object `land_dill_assl` and of the node sequence |Sim| of |Node| object `dill_assl` as examples, which provide the same information: >>> hp.simulate() >>> print_vector(hp.elements.land_dill_assl.model.sequences.fluxes.qt.series) 11.757526, 8.865079, 7.101815, 5.994195 >>> print_vector(hp.nodes.dill_assl.sequences.sim.series) 11.757526, 8.865079, 7.101815, 5.994195 The observed discharge values are also directly available via the node sequence |Obs|: >>> print_vector(hp.nodes.dill_assl.sequences.obs.series) 4.84, 5.19, 4.22, 3.65 """ self.prepare_network() self.prepare_models() self.load_conditions() self.prepare_nodeseries() with hydpy.pub.options.warnmissingobsfile(False): self.load_obsseries() self.prepare_modelseries() self.load_inputseries()
[docs] @printtools.print_progress def prepare_network(self) -> None: """Load all network files as |Selections| (stored in module |pub|) and assign the "complete" selection to the |HydPy| object. .. testsetup:: >>> from hydpy import pub >>> del pub.selections First, we call function |prepare_full_example_1| to prepare the :ref:`HydPy-H-Lahn` example project, including its network files `headwaters.py`, `nonheadwaters.py`, and `streams.py`: >>> from hydpy.core.testtools import prepare_full_example_1 >>> prepare_full_example_1() Directly after initialising class |HydPy|, neither the resulting object nor module |pub| contain any information stemming from the network files: >>> from hydpy import HydPy, pub, TestIO >>> hp = HydPy("HydPy-H-Lahn") >>> pub.selections Traceback (most recent call last): ... hydpy.core.exceptiontools.AttributeNotReady: Attribute selections of module \ `pub` is not defined at the moment. By calling the method |HydPy.prepare_network|, one loads all three network files into separate |Selection| objects, all handled by the |Selections| object of module |pub|. Additionally, there is a |Selection| object named `complete`, covering all |Node| and |Element| objects of the other |Selection| objects: >>> with TestIO(): ... hp.prepare_network() >>> pub.selections Selections("complete", "headwaters", "nonheadwaters", "streams") >>> pub.selections.headwaters <= pub.selections.complete True >>> pub.selections.nonheadwaters <= pub.selections.complete True >>> pub.selections.streams <= pub.selections.complete True Initially, the |HydPy| object is aware of the complete set of |Node| and |Element| objects: >>> hp.nodes == pub.selections.complete.nodes True >>> hp.elements == pub.selections.complete.elements True See the documentation on method |HydPy.update_devices| on how to "activate| another selection in the safest manner. """ hydpy.pub.selections = selectiontools.Selections() hydpy.pub.selections += hydpy.pub.networkmanager.load_files() self.update_devices(selection=hydpy.pub.selections.complete, silent=True)
[docs] def prepare_models(self) -> None: """Read all control files related to the current |Element| objects, initialise the defined models, and prepare their parameter values. .. testsetup:: >>> from hydpy import pub >>> del pub.options.parameterstep First, we call function |prepare_full_example_1| to prepare the :ref:`HydPy-H-Lahn` example project: >>> from hydpy.core.testtools import prepare_full_example_1 >>> prepare_full_example_1() Now, we can initialise a |HydPy| instance accordingly and call its methods |HydPy.prepare_network| and |HydPy.prepare_models|: >>> from hydpy import HydPy, pub, round_, TestIO >>> with TestIO(): ... pub.timegrids = "1996-01-01", "1996-01-05", "1d" ... hp = HydPy("HydPy-H-Lahn") ... hp.prepare_network() ... hp.prepare_models() As a result, each |Element| object handles a model of the type and with the parameter values defined in the relevant control file: >>> hp.elements.land_dill_assl.model.name 'hland_96' >>> hp.elements.land_dill_assl.model.parameters.control.area area(692.3) >>> hp.elements.stream_lahn_marb_lahn_leun.model.name 'musk_classic' >>> hp.elements.stream_lahn_marb_lahn_leun.model.parameters.control.nmbsegments nmbsegments(lag=0.583) The :ref:`HydPy-H-Lahn` example project comes with one auxiliary file, named `land.py`. This file defines general parameter values, valid for all single parameter objects of the different model instances referencing this file via the `auxfile` keyword argument. The following examples use the `land_dill_assl` element to show that the affected parameters are also correctly prepared: >>> control = hp.elements.land_dill_assl.model.parameters.control >>> control.alpha alpha(1.0) >>> control.pcorr pcorr(1.0) >>> control.resparea resparea(True) >>> control.icmax icmax(field=1.0, forest=1.5) We show that the individual |hland_control.IcMax| values for two different elements are different to demonstrate that parameter values defined within a master control file (|hland_control.ZoneType|) can affect the actual values of parameters defined in auxiliary control files: >>> from hydpy import round_ >>> round_(control.icmax.values) 1.0, 1.5, 1.0, 1.5, 1.0, 1.5, 1.0, 1.5, 1.0, 1.5, 1.0, 1.5 >>> round_( ... hp.elements.land_lahn_leun.model.parameters.control.icmax.values) 1.0, 1.5, 1.0, 1.5, 1.0, 1.5, 1.0, 1.5, 1.0, 1.5 Missing parameter information in auxiliary files results in errors like the following: >>> filepath = "HydPy-H-Lahn/control/default/land.py" >>> with TestIO(): ... with open(filepath) as infile: ... text = infile.read().replace("alpha(1.0)", "") ... with open(filepath, "w") as outfile: ... outfile.write(text) ... hp.prepare_models() # doctest: +ELLIPSIS Traceback (most recent call last): ... RuntimeError: While trying to initialise the model object of element \ `land_dill_assl`, the following error occurred: While trying to load the control file \ `...land_dill_assl.py`, the following error occurred: While trying to extract \ information for parameter `alpha` from file `land`, the following error occurred: The \ selected auxiliary file does not define value(s) for parameter `alpha`. Completely wrong control files result in the following error: >>> with TestIO(): ... with open("HydPy-H-Lahn/control/default/land_dill_assl.py", "w"): ... pass ... hp.prepare_models() # doctest: +ELLIPSIS Traceback (most recent call last): ... RuntimeError: While trying to initialise the model object of element \ `land_dill_assl`, the following error occurred: Model parameters cannot be loaded from \ control file `...land_dill_assl.py`. Please refer to the HydPy documentation on how \ to prepare control files properly. """ self.elements.prepare_models()
[docs] def init_models(self) -> None: """Deprecated! Use method |HydPy.prepare_models| instead. >>> from hydpy import HydPy >>> from unittest import mock >>> from hydpy.core.testtools import warn_later >>> with warn_later(), mock.patch.object(HydPy, "prepare_models") as mocked: ... hp = HydPy("test") ... hp.init_models() HydPyDeprecationWarning: Method `init_models` of class `HydPy` is \ deprecated. Use method `prepare_models` instead. >>> mocked.call_args_list [call()] """ self.prepare_models() warnings.warn( "Method `init_models` of class `HydPy` is deprecated. Use method " "`prepare_models` instead.", exceptiontools.HydPyDeprecationWarning, )
[docs] def save_controls( self, parameterstep: Optional[timetools.PeriodConstrArg] = None, simulationstep: Optional[timetools.PeriodConstrArg] = None, auxfiler: Optional[auxfiletools.Auxfiler] = None, ) -> None: """Write the control files of all current |Element| objects. .. testsetup:: >>> from hydpy import pub >>> del pub.options.parameterstep We use the :ref:`HydPy-H-Lahn` example project to demonstrate how to write a complete set of parameter control files. For convenience, we let function |prepare_full_example_2| prepare a fully functional |HydPy| object, handling seven |Element| objects controlling four |hland_96| and three |musk_classic| application models: >>> from hydpy.core.testtools import prepare_full_example_2 >>> hp, pub, TestIO = prepare_full_example_2() At first, there is only one control subfolder named "default", containing the seven master control files used in the step above: >>> import os >>> with TestIO(): ... os.listdir("HydPy-H-Lahn/control") ['default'] Next, we use the |ControlManager| to create a new directory and write analogue control files into it: >>> with TestIO(): ... pub.controlmanager.currentdir = "newdir" ... hp.save_controls() ... sorted(os.listdir("HydPy-H-Lahn/control")) ['default', 'newdir'] First, we focus our examples on the (shorter) control files of the application model |musk_classic|. These control files define the values of the parameters |musk_control.NmbSegments| and |musk_control.Coefficients| via the keyword arguments `lag` and `damp`. For the river channel connecting the outlets of subcatchment `lahn_marb` and `lahn_leun`, the `lag` value is 0.583 days, and the `damp` value is zero: >>> model = hp.elements.stream_lahn_marb_lahn_leun.model >>> model.parameters.control nmbsegments(lag=0.583) coefficients(damp=0.0) Its control file's name equals the element's name: >>> dir_ = "HydPy-H-Lahn/control/newdir/" >>> with TestIO(): ... with open(dir_ + "stream_lahn_marb_lahn_leun.py") as controlfile: ... print(controlfile.read()) # -*- coding: utf-8 -*- <BLANKLINE> from hydpy.models.musk_classic import * <BLANKLINE> simulationstep("1d") parameterstep("1d") <BLANKLINE> nmbsegments(lag=0.583) coefficients(damp=0.0) <BLANKLINE> The time step information stems from the |Timegrid| object available via |pub|: >>> pub.timegrids.stepsize Period("1d") Use the |Auxfiler| class to avoid redefining the same parameter values in multiple control files. We prepare an |Auxfiler| object that handles the model's two parameters discussed above: >>> from hydpy import Auxfiler >>> auxfiler = Auxfiler("musk_classic") >>> auxfiler.musk_classic.add_parameter( ... model.parameters.control.nmbsegments, filename="stream") >>> auxfiler.musk_classic.add_parameter( ... model.parameters.control.coefficients, filename="stream") When passing the |Auxfiler| object to the method |HydPy.save_controls|, the control file of element `stream_lahn_marb_lahn_leun` does not define the values of both parameters on its own but references the auxiliary file `stream.py` instead: >>> with TestIO(): ... pub.controlmanager.currentdir = "newdir" ... hp.save_controls(auxfiler=auxfiler) ... with open(dir_ + "stream_lahn_marb_lahn_leun.py") as controlfile: ... print(controlfile.read()) # -*- coding: utf-8 -*- <BLANKLINE> from hydpy.models.musk_classic import * <BLANKLINE> simulationstep("1d") parameterstep("1d") <BLANKLINE> nmbsegments(auxfile="stream") coefficients(auxfile="stream") <BLANKLINE> `stream.py` contains the actual value definitions: >>> with TestIO(): ... with open(dir_ + "stream.py") as controlfile: ... print(controlfile.read()) # -*- coding: utf-8 -*- <BLANKLINE> from hydpy.models.musk_classic import * <BLANKLINE> simulationstep("1d") parameterstep("1d") <BLANKLINE> nmbsegments(lag=0.583) coefficients(damp=0.0) <BLANKLINE> The |musk_classic| model of element `stream_lahn_leun_lahn_kalk` defines the same value for parameter |musk_control.Coefficients| but a different one for parameter |musk_control.NmbSegments|. Hence, only |musk_control.Coefficients| can reference the control file `stream.py`: >>> with TestIO(): ... with open(dir_ + "stream_lahn_leun_lahn_kalk.py") as controlfile: ... print(controlfile.read()) # -*- coding: utf-8 -*- <BLANKLINE> from hydpy.models.musk_classic import * <BLANKLINE> simulationstep("1d") parameterstep("1d") <BLANKLINE> nmbsegments(lag=0.417) coefficients(auxfile="stream") <BLANKLINE> Another option is to pass alternative step size information. The `simulationstep` information, which is no integral part of control files but helpful in testing them, has no impact on the written data. However, passing an alternative `parameterstep` information changes the written values of time-dependent parameters both in the primary and the auxiliary control files: >>> with TestIO(): ... pub.controlmanager.currentdir = "newdir" ... hp.save_controls( ... auxfiler=auxfiler, parameterstep="2d", simulationstep="1h") ... with open(dir_ + "stream_lahn_marb_lahn_leun.py") as controlfile: ... print(controlfile.read()) # -*- coding: utf-8 -*- <BLANKLINE> from hydpy.models.musk_classic import * <BLANKLINE> simulationstep("1h") parameterstep("2d") <BLANKLINE> nmbsegments(auxfile="stream") coefficients(auxfile="stream") <BLANKLINE> >>> with TestIO(): ... with open(dir_ + "stream.py") as controlfile: ... print(controlfile.read()) # -*- coding: utf-8 -*- <BLANKLINE> from hydpy.models.musk_classic import * <BLANKLINE> simulationstep("1h") parameterstep("2d") <BLANKLINE> nmbsegments(lag=0.2915) coefficients(damp=0.0) <BLANKLINE> >>> with TestIO(): ... with open(dir_ + "stream_lahn_leun_lahn_kalk.py") as controlfile: ... print(controlfile.read()) # -*- coding: utf-8 -*- <BLANKLINE> from hydpy.models.musk_classic import * <BLANKLINE> simulationstep("1h") parameterstep("2d") <BLANKLINE> nmbsegments(lag=0.2085) coefficients(auxfile="stream") <BLANKLINE> In the :ref:`HydPy-H-Lahn` example project, all |hland_96| instances use an |evap_pet_hbv96| submodel for calculating potential evapotranspiration. The discussed writing mechanisms include such submodels automatically. The written files rely on the preferred "with" block syntax for adding submodels and defining their parameter values: >>> with TestIO(): ... with open(dir_ + "land_dill_assl.py") as controlfile: ... print(controlfile.read()) # doctest: +ELLIPSIS # -*- coding: utf-8 -*- <BLANKLINE> from hydpy.models.hland_96 import * from hydpy.models import evap_aet_hbv96 from hydpy.models import evap_pet_hbv96 from hydpy.models import rconc_uh <BLANKLINE> simulationstep("1h") parameterstep("2d") <BLANKLINE> area(692.3) ... gamma(0.0) with model.add_aetmodel_v1(evap_aet_hbv96): temperaturethresholdice(nan) soilmoisturelimit(0.9) excessreduction(0.0) with model.add_petmodel_v1(evap_pet_hbv96): evapotranspirationfactor(1.0) altitudefactor(0.0) precipitationfactor(0.01) airtemperaturefactor(0.1) with model.add_rconcmodel_v1(rconc_uh): uh("triangle", tb=0.18364) <BLANKLINE> When delegating parameter value definitions to auxiliary files, it makes no difference if these parameters are members of a main model or a submodel: >>> auxfiler = Auxfiler("evap_pet_hbv96") >>> for element in hp.elements.search_keywords("catchment"): ... control = element.model.aetmodel.petmodel.parameters.control ... control.airtemperaturefactor(field=0.2, forest=0.1) >>> auxfiler.evap_pet_hbv96.add_parameter( ... control.airtemperaturefactor, filename="evap", ... keywordarguments=control.airtemperaturefactor.keywordarguments) >>> with TestIO(): ... hp.save_controls( ... auxfiler=auxfiler, parameterstep="2d", simulationstep="1h") ... with open(dir_ + "evap.py") as controlfile: ... print(controlfile.read()) # doctest: +ELLIPSIS # -*- coding: utf-8 -*- <BLANKLINE> from hydpy.models.evap_pet_hbv96 import * <BLANKLINE> simulationstep("1h") parameterstep("2d") <BLANKLINE> airtemperaturefactor(field=0.2, forest=0.1) <BLANKLINE> >>> with TestIO(): ... with open(dir_ + "land_dill_assl.py") as controlfile: ... print(controlfile.read()) # doctest: +ELLIPSIS # -*- coding: utf-8 -*- ... gamma(0.0) with model.add_aetmodel_v1(evap_aet_hbv96): temperaturethresholdice(nan) soilmoisturelimit(0.9) excessreduction(0.0) with model.add_petmodel_v1(evap_pet_hbv96): evapotranspirationfactor(1.0) altitudefactor(0.0) precipitationfactor(0.01) airtemperaturefactor(auxfile="evap") with model.add_rconcmodel_v1(rconc_uh): uh("triangle", tb=0.18364) <BLANKLINE> >>> with TestIO(): ... hp.prepare_models() >>> parameters = hp.elements.land_dill_assl.model.aetmodel.petmodel.parameters >>> control = parameters.control >>> control.airtemperaturefactor airtemperaturefactor(field=0.2, forest=0.1) The :ref:`HydPy-H-Lahn` example project relies only upon "scalar" submodels (handled by |SubmodelProperty| instances) and not on "vectorial" submodels (handled by |SubmodelsProperty| instances). Therefore, we now prepare an instance of model |sw1d_channel| and add, for a start, two |sw1d_storage| models to it, assign it to a new |Element| object, and show how method |HydPy.save_controls| writes the further information into a control file: >>> from hydpy import prepare_model >>> channel = prepare_model("sw1d_channel") >>> channel.parameters.control.nmbsegments(2) >>> with channel.add_storagemodel_v1("sw1d_storage", position=0) as storage: ... length(1.0) ... with storage.add_crosssection_v2("wq_trapeze"): ... nmbtrapezes(1) ... bottomlevels(5.0) ... bottomwidths(10.0) ... sideslopes(0.0) >>> with channel.add_storagemodel_v1("sw1d_storage", position=1) as storage: ... length(2.0) ... with storage.add_crosssection_v2("wq_trapeze"): ... nmbtrapezes(1) ... bottomlevels(5.0) ... bottomwidths(10.0) ... sideslopes(0.0) >>> from hydpy import Element >>> my_channel = Element("my_channel") >>> my_channel.model = channel >>> hp.update_devices(elements=my_channel) >>> with TestIO(): ... hp.save_controls() ... with open(dir_ + "my_channel.py") as controlfile: ... print(controlfile.read()) # doctest: +ELLIPSIS # -*- coding: utf-8 -*- <BLANKLINE> from hydpy.models.sw1d_channel import * from hydpy.models import sw1d_storage from hydpy.models import wq_trapeze <BLANKLINE> simulationstep("1d") parameterstep("1d") <BLANKLINE> nmbsegments(2) with model.add_storagemodel_v1(sw1d_storage, position=0): length(1.0) ... with model.add_storagemodel_v1(sw1d_storage, position=1): length(2.0) ... <BLANKLINE> We now add one of three possible routing models in the second position to demonstrate the writing mechanism not only for completely missing (like in the last example) but also for incomplete submodel vectors: >>> with channel.add_routingmodel_v2("sw1d_lias", position=1): ... lengthupstream(1.0) ... lengthdownstream(2.0) ... stricklercoefficient(30.0) ... timestepfactor(0.7) ... diffusionfactor(0.2) ... with storage.add_crosssection_v2("wq_trapeze"): ... nmbtrapezes(1) ... bottomlevels(5.0) ... bottomwidths(10.0) ... sideslopes(0.0) >>> with TestIO(): ... hp.save_controls() ... with open(dir_ + "my_channel.py") as controlfile: ... print(controlfile.read()) # doctest: +ELLIPSIS # -*- coding: utf-8 -*- <BLANKLINE> from hydpy.models.sw1d_channel import * from hydpy.models import sw1d_lias from hydpy.models import sw1d_storage from hydpy.models import wq_trapeze <BLANKLINE> simulationstep("1d") parameterstep("1d") <BLANKLINE> nmbsegments(2) with model.add_routingmodel_v2(sw1d_lias, position=1): lengthupstream(1.0) ... with model.add_storagemodel_v1(sw1d_storage, position=0): length(1.0) ... with model.add_storagemodel_v1(sw1d_storage, position=1): length(2.0) ... <BLANKLINE> Finally, we demonstrate the writing mechanism for the case of different submodel types within the same submodel vector: >>> with channel.add_routingmodel_v1("sw1d_q_in", position=0): ... lengthdownstream(1.0) >>> with channel.add_routingmodel_v3("sw1d_weir_out", position=2): ... lengthupstream(2.0) >>> with TestIO(): ... hp.save_controls() ... with open(dir_ + "my_channel.py") as controlfile: ... print(controlfile.read()) # doctest: +ELLIPSIS # -*- coding: utf-8 -*- <BLANKLINE> from hydpy.models.sw1d_channel import * from hydpy.models import sw1d_lias from hydpy.models import sw1d_q_in from hydpy.models import sw1d_storage from hydpy.models import sw1d_weir_out from hydpy.models import wq_trapeze <BLANKLINE> simulationstep("1d") parameterstep("1d") <BLANKLINE> nmbsegments(2) with model.add_routingmodel_v1(sw1d_q_in, position=0): lengthdownstream(1.0) ... with model.add_routingmodel_v2(sw1d_lias, position=1): lengthupstream(1.0) lengthdownstream(2.0) ... with model.add_routingmodel_v3(sw1d_weir_out, position=2): lengthupstream(2.0) ... with model.add_storagemodel_v1(sw1d_storage, position=0): length(1.0) ... with model.add_storagemodel_v1(sw1d_storage, position=1): length(2.0) ... <BLANKLINE> """ self.elements.save_controls( parameterstep=parameterstep, simulationstep=simulationstep, auxfiler=auxfiler, )
[docs] def load_conditions(self) -> None: """Load all currently relevant initial conditions. .. testsetup:: >>> from hydpy import pub >>> del pub.options.parameterstep The following examples demonstrate both the functionality of method |HydPy.load_conditions| and |HydPy.save_conditions| based on the :ref:`HydPy-H-Lahn` project, which we prepare via function |prepare_full_example_2|: >>> from hydpy.core.testtools import prepare_full_example_2 >>> hp, pub, TestIO = prepare_full_example_2() Our |HydPy| instance `hp` is ready for the first simulation run, meaning the required initial conditions are available already. First, we start a simulation run covering the whole initialisation period and inspect the resulting soil moisture values of |Element| `land_dill_assl`, handled by a sequence object of type |hland_states.SM|: >>> hp.simulate() >>> sm = hp.elements.land_dill_assl.model.sequences.states.sm >>> sm sm(184.553224, 180.623124, 199.181137, 195.947542, 212.04018, 209.48859, 222.12115, 220.12671, 230.30756, 228.70779, 236.91943, 235.64427) By default, method |HydPy.load_conditions| always (re)loads the initial conditions from the directory with its name matching the start date of the simulation period, which we prove by also showing the related content of the respective condition file `land_dill_assl.py`: >>> with TestIO(): ... hp.load_conditions() >>> sm sm(185.13164, 181.18755, 199.80432, 196.55888, 212.04018, 209.48859, 222.12115, 220.12671, 230.30756, 228.70779, 236.91943, 235.64427) >>> path = "HydPy-H-Lahn/conditions/init_1996_01_01_00_00_00/land_dill_assl.py" >>> with TestIO(): ... with open(path, "r") as file_: ... lines = file_.read().split("\\n") ... print(lines[10]) ... print(lines[11]) sm(185.13164, 181.18755, 199.80432, 196.55888, 212.04018, 209.48859, 222.12115, 220.12671, 230.30756, 228.70779, 236.91943, 235.64427) Now, we perform two consecutive runs, covering the first and the second half of the initialisation period, respectively, and write, in both cases, the resulting final conditions to disk: >>> pub.timegrids.sim.lastdate = "1996-01-03" >>> hp.simulate() >>> sm sm(184.763638, 180.829058, 199.40823, 196.170947, 212.04018, 209.48859, 222.12115, 220.12671, 230.30756, 228.70779, 236.91943, 235.64427) >>> with TestIO(): ... hp.save_conditions() >>> pub.timegrids.sim.firstdate = "1996-01-03" >>> pub.timegrids.sim.lastdate = "1996-01-05" >>> hp.simulate() >>> with TestIO(): ... hp.save_conditions() >>> sm sm(184.553224, 180.623124, 199.181137, 195.947542, 212.04018, 209.48859, 222.12115, 220.12671, 230.30756, 228.70779, 236.91943, 235.64427) Analogous to method |HydPy.load_conditions|, method |HydPy.save_conditions| writes the resulting conditions to a directory with its name matching the end date of the simulation period, which we prove by reloading the conditions related to the middle of the initialisation period and showing the relevant file content: >>> with TestIO(): ... hp.load_conditions() >>> sm sm(184.763638, 180.829058, 199.40823, 196.170947, 212.04018, 209.48859, 222.12115, 220.12671, 230.30756, 228.70779, 236.91943, 235.64427) >>> path = "HydPy-H-Lahn/conditions/init_1996_01_03_00_00_00/land_dill_assl.py" >>> with TestIO(): ... with open(path, "r") as file_: ... lines = file_.read().split("\\n") ... print(lines[12]) ... print(lines[13]) sm(184.763638, 180.829058, 199.40823, 196.170947, 212.04018, 209.48859, 222.12115, 220.12671, 230.30756, 228.70779, 236.91943, 235.64427) You can define another directory by assigning a different name to the attribute |FileManager.currentdir| of the actual |ConditionManager| instance: >>> with TestIO(): ... pub.conditionmanager.currentdir = "test" ... hp.save_conditions() >>> path = "HydPy-H-Lahn/conditions/test/land_dill_assl.py" >>> with TestIO(): ... with open(path, "r") as file_: ... lines = file_.read().split("\\n") ... print(lines[12]) ... print(lines[13]) sm(184.763638, 180.829058, 199.40823, 196.170947, 212.04018, 209.48859, 222.12115, 220.12671, 230.30756, 228.70779, 236.91943, 235.64427) This change remains permanent until you undo it manually: >>> sm(0.0) >>> pub.timegrids.sim.firstdate = "1996-01-01" >>> with TestIO(): ... hp.load_conditions() >>> sm sm(184.763638, 180.829058, 199.40823, 196.170947, 212.04018, 209.48859, 222.12115, 220.12671, 230.30756, 228.70779, 236.91943, 235.64427) >>> with TestIO(): ... del pub.conditionmanager.currentdir ... hp.load_conditions() >>> sm sm(185.13164, 181.18755, 199.80432, 196.55888, 212.04018, 209.48859, 222.12115, 220.12671, 230.30756, 228.70779, 236.91943, 235.64427) The :ref:`HydPy-H-Lahn` example project relies only upon "scalar" submodels (handled by |SubmodelProperty| instances) and not on "vectorial" submodels (handled by |SubmodelsProperty| instances). Therefore, we now prepare an instance of model |sw1d_channel| and add, for a start, two |sw1d_storage| models to it, assign it to a new |Element| object, and show how method |HydPy.save_conditions| writes their states into a condition file: >>> from hydpy import Element, prepare_model >>> channel = prepare_model("sw1d_channel") >>> channel.parameters.control.nmbsegments(2) >>> with channel.add_storagemodel_v1("sw1d_storage", position=0): ... states.watervolume = 10.0 >>> with channel.add_storagemodel_v1("sw1d_storage", position=1): ... states.watervolume = 20.0 >>> my_channel = Element("my_channel") >>> my_channel.model = channel >>> hp.update_devices(elements=my_channel) >>> pub.timegrids.sim.lastdate = "1996-01-03" >>> path = "HydPy-H-Lahn/conditions/init_1996_01_03_00_00_00/my_channel.py" >>> with TestIO(): ... hp.save_conditions() ... with open(path) as conditionfile: ... print(conditionfile.read()) # -*- coding: utf-8 -*- <BLANKLINE> from hydpy.models.sw1d_channel import * from hydpy.models import sw1d_storage <BLANKLINE> controlcheck(projectdir=r"HydPy-H-Lahn", controldir="default", stepsize="1d") <BLANKLINE> with model.storagemodels[0].define_conditions(sw1d_storage): watervolume(10.0) with model.storagemodels[1].define_conditions(sw1d_storage): watervolume(20.0) <BLANKLINE> For testing, we set both submodels' water volume to zero: >>> channel.storagemodels[0].sequences.states.watervolume = 0.0 >>> channel.storagemodels[1].sequences.states.watervolume = 0.0 Loading the previously written condition file restores the original values: >>> pub.timegrids.sim.firstdate = "1996-01-03" >>> with TestIO(): ... hp.load_conditions() >>> channel.storagemodels[0].sequences.states.watervolume watervolume(10.0) >>> channel.storagemodels[1].sequences.states.watervolume watervolume(20.0) We now add one of three possible routing models in the second position to demonstrate the writing mechanism not only for completely missing (like in the last example) but also for incomplete submodel vectors: >>> with channel.add_routingmodel_v2("sw1d_lias", position=1, update=False): ... states.discharge(1.1) >>> with TestIO(): ... hp.save_conditions() ... with open(path) as conditionfile: ... print(conditionfile.read()) # -*- coding: utf-8 -*- <BLANKLINE> from hydpy.models.sw1d_channel import * from hydpy.models import sw1d_lias from hydpy.models import sw1d_storage <BLANKLINE> controlcheck(projectdir=r"HydPy-H-Lahn", controldir="default", stepsize="1d") <BLANKLINE> with model.routingmodels[1].define_conditions(sw1d_lias): discharge(1.1) with model.storagemodels[0].define_conditions(sw1d_storage): watervolume(10.0) with model.storagemodels[1].define_conditions(sw1d_storage): watervolume(20.0) <BLANKLINE> >>> channel.storagemodels[0].sequences.states.watervolume = 0.0 >>> channel.storagemodels[1].sequences.states.watervolume = 0.0 >>> channel.routingmodels[1].sequences.states.discharge = 0.0 >>> with TestIO(): ... hp.load_conditions() >>> channel.storagemodels[0].sequences.states.watervolume watervolume(10.0) >>> channel.storagemodels[1].sequences.states.watervolume watervolume(20.0) >>> channel.routingmodels[1].sequences.states.discharge discharge(1.1) Finally, we demonstrate the writing mechanism for the case of different submodel types within the same submodel vector: >>> with channel.add_routingmodel_v1("sw1d_q_in", position=0): ... states.discharge = 1.0 >>> with channel.add_routingmodel_v3("sw1d_weir_out", position=2): ... states.discharge = 1.2 >>> with TestIO(): ... hp.save_conditions() ... with open(path) as conditionfile: ... print(conditionfile.read()) # -*- coding: utf-8 -*- <BLANKLINE> from hydpy.models.sw1d_channel import * from hydpy.models import sw1d_lias from hydpy.models import sw1d_q_in from hydpy.models import sw1d_storage from hydpy.models import sw1d_weir_out <BLANKLINE> controlcheck(projectdir=r"HydPy-H-Lahn", controldir="default", stepsize="1d") <BLANKLINE> with model.routingmodels[0].define_conditions(sw1d_q_in): discharge(1.0) with model.routingmodels[1].define_conditions(sw1d_lias): discharge(1.1) with model.routingmodels[2].define_conditions(sw1d_weir_out): discharge(1.2) with model.storagemodels[0].define_conditions(sw1d_storage): watervolume(10.0) with model.storagemodels[1].define_conditions(sw1d_storage): watervolume(20.0) <BLANKLINE> >>> channel.storagemodels[0].sequences.states.watervolume = 0.0 >>> channel.storagemodels[1].sequences.states.watervolume = 0.0 >>> channel.routingmodels[0].sequences.states.discharge = 0.0 >>> channel.routingmodels[1].sequences.states.discharge = 0.0 >>> channel.routingmodels[2].sequences.states.discharge = 0.0 >>> with TestIO(): ... hp.load_conditions() >>> channel.storagemodels[0].sequences.states.watervolume watervolume(10.0) >>> channel.storagemodels[1].sequences.states.watervolume watervolume(20.0) >>> channel.routingmodels[0].sequences.states.discharge discharge(1.0) >>> channel.routingmodels[1].sequences.states.discharge discharge(1.1) >>> channel.routingmodels[2].sequences.states.discharge discharge(1.2) """ self.elements.load_conditions()
[docs] def save_conditions(self) -> None: """Save all currently relevant final conditions. See the documentation on method |HydPy.load_conditions| for further information. """ self.elements.save_conditions()
[docs] def trim_conditions(self) -> None: """Check all values of the condition sequences (|StateSequence| and |LogSequence| objects) for boundary violations and fix them if necessary. We use the :ref:`HydPy-H-Lahn` example project to explain the functionality of the method |HydPy.trim_conditions|, which gives no response when all conditions are correctly set: >>> from hydpy.core.testtools import prepare_full_example_2 >>> hp, pub, TestIO = prepare_full_example_2() >>> with pub.options.warntrim(True): ... hp.trim_conditions() If you try, for example, to set the snow layer's liquid water content (|hland_states.WC|) to a value larger than the allowed fraction of the snow layer's frozen water content (|hland_states.SP|), you get a direct response based on function |trim|: >>> from hydpy.core.testtools import warn_later >>> with pub.options.warntrim(True), warn_later(): # doctest: +ELLIPSIS ... hp.elements.land_dill_assl.model.sequences.states.wc(1.0) UserWarning: For variable `wc` of element `land_dill_assl` at least one value \ needed to be trimmed. The old and the new value(s) are `1.0, ..., 1.0` and `0.0, \ ..., 0.0`, respectively. However, changing the allowed fraction (|hland_control.WHC|) without adjusting the conditions cannot be detected automatically. Whenever in doubt, call method |HydPy.trim_conditions| explicitly: >>> hp.elements.land_dill_assl.model.sequences.states.sp(10.0) >>> hp.elements.land_dill_assl.model.sequences.states.wc(1.0) >>> hp.elements.land_dill_assl.model.parameters.control.whc(0.0) >>> with pub.options.warntrim(True), warn_later(): ... hp.trim_conditions() # doctest: +ELLIPSIS UserWarning: For variable `wc` of element `land_dill_assl` at least one value \ needed to be trimmed. The old and the new value(s) are `1.0, ..., 1.0` and `0.0, \ ..., 0.0`, respectively. """ self.elements.trim_conditions()
[docs] def reset_conditions(self) -> None: """Reset all currently relevant condition sequences. Method |HydPy.reset_conditions| is the most convenient way to perform simulations repeatedly for the same period, each time starting from the same initial conditions, e.g. for parameter calibration. Each |StateSequence| and |LogSequence| object remembers the last assigned values and can reactivate them for the mentioned purpose. For demonstration, we perform a simulation for the :ref:`HydPy-H-Lahn` example project spanning four days: >>> from hydpy.core.testtools import prepare_full_example_2 >>> hp, pub, TestIO = prepare_full_example_2() >>> hp.simulate() >>> from hydpy import print_vector >>> print_vector(hp.nodes.lahn_kalk.sequences.sim.series) 54.019337, 37.257561, 31.865308, 28.359542 Just repeating the simulation gives different results due to applying the final states of the first simulation run as the initial states of the second run: >>> hp.simulate() >>> print_vector(hp.nodes.lahn_kalk.sequences.sim.series) 26.196165, 25.047469, 24.227865, 23.307609 Calling |HydPy.reset_conditions| first allows repeating the first simulation run exactly multiple times: >>> hp.reset_conditions() >>> hp.simulate() >>> print_vector(hp.nodes.lahn_kalk.sequences.sim.series) 54.019337, 37.257561, 31.865308, 28.359542 >>> hp.reset_conditions() >>> hp.simulate() >>> print_vector(hp.nodes.lahn_kalk.sequences.sim.series) 54.019337, 37.257561, 31.865308, 28.359542 """ self.elements.reset_conditions()
@property def conditions(self) -> Conditions: """A nested dictionary that contains the values of all condition sequences of all currently handled models. The primary purpose of property |HydPy.conditions| is, similar to method |HydPy.reset_conditions|, to allow repeated calculations starting from the same initial conditions. Nevertheless, |HydPy.conditions| is more flexible when handling multiple conditions, which can, for example, help apply ensemble-based assimilation algorithms. For demonstration, we perform simulations for the :ref:`HydPy-H-Lahn` example project spanning the first three months of 1996. We begin with a preparation run beginning on January 1 and ending on February 20: >>> from hydpy.core.testtools import prepare_full_example_1 >>> prepare_full_example_1() >>> from hydpy import HydPy, pub, TestIO, print_vector >>> with TestIO(): ... hp = HydPy("HydPy-H-Lahn") ... pub.timegrids = "1996-01-01", "1996-04-01", "1d" ... hp.prepare_everything() >>> pub.timegrids.sim.lastdate = "1996-02-20" >>> hp.simulate() >>> print_vector(hp.nodes.lahn_kalk.sequences.sim.series[48:52]) 74.80616, 97.932993, nan, nan At the end of the preparation run, a snow layer is covering the Lahn catchment. In the `lahn_marb` subcatchment, this snow layer contains 13.7 mm of frozen water and 1.3 mm of liquid water: >>> lahn1_states = hp.elements.land_lahn_marb.model.sequences.states >>> print_vector([lahn1_states.sp.average_values()]) 13.727031 >>> print_vector([lahn1_states.wc.average_values()]) 1.250427 Now, we save the current conditions and perform the first simulation run from the 20th day of February until the end of March: >>> conditions = hp.conditions >>> hp.nodes.lahn_kalk.sequences.sim.series = 0.0 >>> pub.timegrids.sim.firstdate = "1996-02-20" >>> pub.timegrids.sim.lastdate = "1996-04-01" >>> hp.simulate() >>> first = hp.nodes.lahn_kalk.sequences.sim.series.copy() >>> print_vector(first[48:52]) 0.0, 0.0, 88.226976, 66.667346 To exactly repeat the last simulation run, we assign the memorised conditions to property |HydPy.conditions|: >>> hp.conditions = conditions >>> print_vector([lahn1_states.sp.average_values()]) 13.727031 >>> print_vector([lahn1_states.wc.average_values()]) 1.250427 All discharge values of the second simulation run are identical to the ones of the first simulation run: >>> hp.nodes.lahn_kalk.sequences.sim.series = 0.0 >>> pub.timegrids.sim.firstdate = "1996-02-20" >>> pub.timegrids.sim.lastdate = "1996-04-01" >>> hp.simulate() >>> second = hp.nodes.lahn_kalk.sequences.sim.series.copy() >>> print_vector(second[48:52]) 0.0, 0.0, 88.226976, 66.667346 >>> from numpy import isclose >>> assert all(isclose(first, second, rtol=0.0, atol=1e-12)) We selected the snow period as an example due to potential problems with the limited water-holding capacity of the snow layer, which depends on the ice content of the snow layer (|hland_states.SP|) and the relative water-holding capacity (|hland_control.WHC|). Due to this restriction, problems can occur. To give an example, we set |hland_control.WHC| to zero temporarily, apply the memorised conditions, and finally reset the original values of | hland_control.WHC|: >>> for element in hp.elements.catchment: ... element.whc = element.model.parameters.control.whc.values ... element.model.parameters.control.whc = 0.0 >>> with pub.options.warntrim(False): ... hp.conditions = conditions >>> for element in hp.elements.catchment: ... element.model.parameters.control.whc = element.whc Without any water-holding capacity of the snow layer, its water content is zero despite the actual memorised value of 1.1 mm: >>> print_vector([lahn1_states.sp.average_values()]) 13.727031 >>> print_vector([lahn1_states.wc.average_values()]) 0.0 What happens in such conflicts depends on the implementation of the respective application model. For safety, we suggest setting the option |Options.warntrim| to |True| before resetting conditions. """ return self.elements.conditions @conditions.setter def conditions(self, conditions: Conditions) -> None: self.elements.conditions = conditions @property def networkproperties( self, ) -> dict[ str, Union[int, Union[dict[str, int], dict[devicetools.NodeVariableType, int]]] ]: """Some properties of the network defined by the currently relevant |Node| and |Element| objects. See the documentation on method |HydPy.print_networkproperties| for further information. """ return { "Number of nodes": len(self.nodes), "Number of elements": len(self.elements), "Number of end nodes": len(self.endnodes), "Number of distinct networks": len(self.segregatednetworks), "Applied node variables": self.variables, "Applied model types": self.modeltypes, }
[docs] def print_networkproperties(self) -> None: """Print some properties of the network defined by the currently relevant |Node| and |Element| objects. |HydPy.print_networkproperties| is for convenience to summarise specific network measures like |HydPy.segregatednetworks|. The :ref:`HydPy-H-Lahn` example project defines a small, single network, with all catchments ultimately discharging to node `lahn_kalk`: >>> from hydpy.core.testtools import prepare_full_example_1 >>> prepare_full_example_1() >>> from hydpy import HydPy, pub, TestIO >>> pub.timegrids = "1996-01-01", "1996-01-05", "1d" >>> with TestIO(): ... hp = HydPy("HydPy-H-Lahn") ... hp.prepare_network() ... hp.prepare_models() >>> hp.print_networkproperties() Number of nodes: 4 Number of elements: 7 Number of end nodes: 1 Number of distinct networks: 1 Applied node variables: Q (4) Applied model types: hland_96 (4) and musk_classic (3) """ value: Union[ str, int, Union[dict[str, int], dict[devicetools.NodeVariableType, int]] ] for key, value in self.networkproperties.items(): if isinstance(value, dict): value = objecttools.enumeration( f"{name} ({nmb})" for name, nmb in value.items() ) print(f"{key}: {value}")
@property def endnodes(self) -> devicetools.Nodes: """All currently relevant |Node| objects that define a downstream endpoint of the network. The :ref:`HydPy-H-Lahn` example project defines a small, single network, with all catchments ultimately discharging to node `lahn_kalk`: >>> from hydpy.core.testtools import prepare_full_example_1 >>> prepare_full_example_1() >>> from hydpy import HydPy, TestIO >>> with TestIO(): ... hp = HydPy("HydPy-H-Lahn") ... hp.prepare_network() >>> hp.endnodes Nodes("lahn_kalk") After breaking the connection between node `lahn_marb` and its downstream river channel element `stream_lahn_marb_lahn2`, `lahn_marb` also becomes an end node: >>> hp.nodes.lahn_marb.exits.remove_device("stream_lahn_marb_lahn_leun", ... force=True) >>> hp.elements.stream_lahn_marb_lahn_leun.inlets.remove_device("lahn_marb", ... force=True) >>> hp.endnodes Nodes("lahn_kalk", "lahn_marb") Even with a proper connection to a downstream element, a node counts as an end node as long as these elements are not part of the currently relevant network (meaning, currently handled by the |HydPy| object): >>> del hp.elements.stream_dill_assl_lahn_leun >>> hp.nodes.dill_assl.exits Elements("stream_dill_assl_lahn_leun") >>> hp.endnodes Nodes("dill_assl", "lahn_kalk", "lahn_marb") Connections with "remote" elements are considered irrelevant: >>> stream = hp.elements.stream_lahn_leun_lahn_kalk >>> stream.receivers.add_device(stream.inlets.lahn_leun, force=True) >>> stream.inlets.remove_device("lahn_leun", force=True) >>> hp.endnodes Nodes("dill_assl", "lahn_kalk", "lahn_leun", "lahn_marb") """ endnodes = devicetools.Nodes() for node in self.nodes: for element in node.exits: if (element in self.elements) and (node not in element.receivers): break else: endnodes += node return endnodes @property def segregatednetworks(self) -> selectiontools.Selections: """The number of segregated networks, as defined by the currently relevant |Node| and |Element| objects. Each end node (as defined by property |HydPy.endnodes|) eventually defines a single network, segregated from the networks of other end nodes. Due to the :ref:`HydPy-H-Lahn` example project defining only a single end node, there can be only one segregated network, accordingly: >>> from hydpy.core.testtools import prepare_full_example_1 >>> prepare_full_example_1() >>> from hydpy import HydPy, TestIO >>> with TestIO(): ... hp = HydPy("HydPy-H-Lahn") ... hp.prepare_network() >>> hp.segregatednetworks Selections("lahn_kalk") >>> hp.segregatednetworks.lahn_kalk Selection("lahn_kalk", nodes=("dill_assl", "lahn_kalk", "lahn_leun", "lahn_marb"), elements=("land_dill_assl", "land_lahn_kalk", "land_lahn_leun", "land_lahn_marb", "stream_dill_assl_lahn_leun", "stream_lahn_leun_lahn_kalk", "stream_lahn_marb_lahn_leun")) Revisiting the examples of the documentation on property |HydPy.endnodes|, we get the similar results. Note that the segregated networks are always |Selection| objects that do not overlap each other (meaning, no |Node| or |Element| object occurs more than one time): >>> hp.nodes.lahn_marb.exits.remove_device("stream_lahn_marb_lahn_leun", ... force=True) >>> hp.elements.stream_lahn_marb_lahn_leun.inlets.remove_device("lahn_marb", ... force=True) >>> hp.segregatednetworks Selections("lahn_kalk", "lahn_marb") >>> hp.segregatednetworks.lahn_marb Selection("lahn_marb", nodes="lahn_marb", elements="land_lahn_marb") >>> hp.segregatednetworks.lahn_kalk Selection("lahn_kalk", nodes=("dill_assl", "lahn_kalk", "lahn_leun"), elements=("land_dill_assl", "land_lahn_kalk", "land_lahn_leun", "stream_dill_assl_lahn_leun", "stream_lahn_leun_lahn_kalk", "stream_lahn_marb_lahn_leun")) >>> del hp.elements.stream_dill_assl_lahn_leun >>> hp.nodes.dill_assl.exits Elements("stream_dill_assl_lahn_leun") >>> hp.segregatednetworks Selections("dill_assl", "lahn_kalk", "lahn_marb") >>> hp.segregatednetworks.dill_assl Selection("dill_assl", nodes="dill_assl", elements="land_dill_assl") >>> hp.segregatednetworks.lahn_marb Selection("lahn_marb", nodes="lahn_marb", elements="land_lahn_marb") >>> hp.segregatednetworks.lahn_kalk Selection("lahn_kalk", nodes=("lahn_kalk", "lahn_leun"), elements=("land_lahn_kalk", "land_lahn_leun", "stream_lahn_leun_lahn_kalk", "stream_lahn_marb_lahn_leun")) >>> stream = hp.elements.stream_lahn_leun_lahn_kalk >>> stream.receivers.add_device(stream.inlets.lahn_leun, force=True) >>> stream.inlets.remove_device("lahn_leun", force=True) >>> hp.segregatednetworks Selections("dill_assl", "lahn_kalk", "lahn_leun", "lahn_marb") >>> hp.segregatednetworks.dill_assl Selection("dill_assl", nodes="dill_assl", elements="land_dill_assl") >>> hp.segregatednetworks.lahn_marb Selection("lahn_marb", nodes="lahn_marb", elements="land_lahn_marb") >>> hp.segregatednetworks.lahn_leun Selection("lahn_leun", nodes="lahn_leun", elements=("land_lahn_leun", "stream_lahn_marb_lahn_leun")) >>> hp.segregatednetworks.lahn_kalk Selection("lahn_kalk", nodes="lahn_kalk", elements=("land_lahn_kalk", "stream_lahn_leun_lahn_kalk")) In all the examples above, the number of the end nodes and the number of the segregated networks are identical, which is not the case when two or more networks share the same network. We restore our original network and add two additional end nodes, `nowhere` and `somewhere`, linking the first one with element `stream_lahn_leun_lahn_kalk` and the second one with the additional element `stream_lahn_marb_nowhere`, which we connect to node `lahn_marb`: >>> with TestIO(): ... hp = HydPy("HydPy-H-Lahn") ... hp.prepare_network() >>> from hydpy import Element >>> _ = Element("stream_lahn_leun_lahn_kalk", outlets="nowhere") >>> hp.nodes += "nowhere" >>> hp.elements += Element("stream_lahn_marb_nowhere", ... inlets="lahn_marb", ... outlets="somewhere") >>> hp.nodes += "somewhere" Now, there are three end nodes but only two segregated networks, as node `nowhere` does not reference any upstream devices, which is not also referenced by node `lahn_kalk`. The unique feature of the elements `lahn_kalk` and `stream_lahn_marb_nowhere` is that they drain to either node `lahn_kalk` or `somewhere` but not both, which is why they are the only members of selection `lahn_kalk` and `somewhere`, respectively: >>> hp.endnodes Nodes("lahn_kalk", "nowhere", "somewhere") >>> hp.segregatednetworks Selections("lahn_kalk", "somewhere") >>> hp.segregatednetworks.lahn_kalk Selection("lahn_kalk", nodes="lahn_kalk", elements="land_lahn_kalk") >>> hp.segregatednetworks.somewhere Selection("somewhere", nodes="somewhere", elements="stream_lahn_marb_nowhere") """ sels1, sels2 = selectiontools.Selections(), selectiontools.Selections() complete = selectiontools.Selection("complete", self.nodes, self.elements) for node in self.endnodes: sel = complete.search_upstream(device=node, name=node.name, inclusive=False) sels1.add_selections(sel) sels2.add_selections(sel.copy(node.name)) for sel1, sel2 in itertools.product(sels1, sels2): if sel1.name != sel2.name: sel1 -= sel2 for name in tuple(sels1.names): if not sels1[name].elements: del sels1[name] return sels1 @property def variables(self) -> dict[devicetools.NodeVariableType, int]: """Summary of all |Node.variable| properties of the currently relevant |Node| objects. >>> from hydpy.core.testtools import prepare_full_example_1 >>> prepare_full_example_1() >>> from hydpy import HydPy, TestIO >>> with TestIO(): ... hp = HydPy("HydPy-H-Lahn") ... hp.prepare_network() >>> hp.variables {'Q': 4} >>> from hydpy import FusedVariable, Node >>> from hydpy.aliases import hland_inputs_T >>> hp.nodes += Node("test", variable=FusedVariable("T", hland_inputs_T)) >>> hp.variables {'Q': 4, FusedVariable("T", hland_inputs_T): 1} """ variables: collections.defaultdict[ Union[ str, type[sequencetools.InputSequence], type[sequencetools.InletSequence], type[sequencetools.ReceiverSequence], type[sequencetools.OutputSequence], type[sequencetools.OutletSequence], type[sequencetools.SenderSequence], devicetools.FusedVariable, ], int, ] = collections.defaultdict(lambda: 0) for node in self.nodes: variables[node.variable] += 1 return dict( sorted( variables.items(), key=lambda tuple_: f"{(tuple_[0])}{str(tuple_[1]).rjust(9)}", ) ) @property def modeltypes(self) -> dict[str, int]: """Summary of all |Model| subclasses of the currently relevant |Element| objects. >>> from hydpy.core.testtools import prepare_full_example_1 >>> prepare_full_example_1() >>> from hydpy import HydPy, pub, TestIO >>> with TestIO(): ... hp = HydPy("HydPy-H-Lahn") ... hp.prepare_network() >>> hp.modeltypes {'unprepared': 7} >>> pub.timegrids = "1996-01-01", "1996-01-05", "1d" >>> with TestIO(): ... hp.prepare_models() >>> hp.modeltypes {'hland_96': 4, 'musk_classic': 3} """ modeltypes: collections.defaultdict[str, int] modeltypes = collections.defaultdict(lambda: 0) for element in self.elements: model = exceptiontools.getattr_( element, "model", "unprepared", modeltools.Model, # type: ignore[type-abstract] ) modeltypes[str(model)] += 1 return dict(sorted(modeltypes.items())) @overload def update_devices( self, *, selection: selectiontools.Selection, silent: bool = False ) -> None: ... @overload def update_devices( self, *, nodes: Optional[devicetools.NodesConstrArg] = None, elements: Optional[devicetools.ElementsConstrArg] = None, silent: bool = False, ) -> None: ...
[docs] def update_devices( self, *, selection: Optional[selectiontools.Selection] = None, nodes: Optional[devicetools.NodesConstrArg] = None, elements: Optional[devicetools.ElementsConstrArg] = None, silent: bool = False, ) -> None: """Determine the order in which method |HydPy.simulate| processes the currently relevant |Node| and |Element| objects. Passed |Node| and |Element| objects (for example, contained within a |Selection| object) replace existing ones. As described in the documentation on the method |HydPy.prepare_network|, a |HydPy| object usually starts with the "complete" network of the considered project: >>> from hydpy.core.testtools import prepare_full_example_2 >>> hp, pub, TestIO = prepare_full_example_2() The safest approach to "activate" another selection is to use the method |HydPy.update_devices|. The first option is to pass a complete |Selection| object: >>> pub.selections.headwaters Selection("headwaters", nodes=("dill_assl", "lahn_marb"), elements=("land_dill_assl", "land_lahn_marb")) >>> hp.update_devices(selection=pub.selections.headwaters) >>> hp.nodes Nodes("dill_assl", "lahn_marb") >>> hp.elements Elements("land_dill_assl", "land_lahn_marb") Method |HydPy.update_devices| automatically updates the `deviceorder`, assuring method |HydPy.simulate| processes "upstream" model instances before it processes their "downstream" neighbours: >>> for device in hp.deviceorder: ... print(device) land_dill_assl land_lahn_marb dill_assl lahn_marb Second, you can pass some nodes only, which, by the way, removes the old elements: >>> hp.update_devices(nodes="dill_assl") >>> hp.nodes Nodes("dill_assl") >>> hp.elements Elements() >>> for device in hp.deviceorder: ... print(device) dill_assl Third, you can pass some elements only, which, by the way, removes the old nodes: >>> hp.update_devices(elements=["land_lahn_marb", "land_dill_assl"]) >>> hp.nodes Nodes() >>> hp.elements Elements("land_dill_assl", "land_lahn_marb") >>> for device in hp.deviceorder: ... print(device) land_dill_assl land_lahn_marb Fourth, you can pass nodes and elements at the same time: >>> hp.update_devices(nodes="dill_assl", elements=["land_lahn_marb", ... "land_dill_assl"]) >>> hp.nodes Nodes("dill_assl") >>> hp.elements Elements("land_dill_assl", "land_lahn_marb") >>> for device in hp.deviceorder: ... print(device) land_dill_assl land_lahn_marb dill_assl Fifth, you can pass no argument at all, which only updates the device order: >>> del hp.nodes.dill_assl >>> for device in hp.deviceorder: ... print(device) land_dill_assl land_lahn_marb dill_assl >>> hp.update_devices() >>> for device in hp.deviceorder: ... print(device) land_dill_assl land_lahn_marb Method |HydPy.update_devices| does not allow using the `selections` argument and the `nodes` or `elements` argument simultaneously: >>> hp.update_devices(selection=pub.selections.headwaters, nodes="dill_assl") Traceback (most recent call last): ... ValueError: Method `update_devices` of class `HydPy` does not allow to use \ both the `selection` argument and the `nodes` or the `elements` argument at the same \ time. >>> hp.update_devices(selection=pub.selections.headwaters, ... elements=["land_lahn_marb", "land_dill_assl"]) Traceback (most recent call last): ... ValueError: Method `update_devices` of class `HydPy` does not allow to use \ both the `selection` argument and the `nodes` or the `elements` argument at the same \ time. Elements of the same |Element.collective| require special attention, as they and their models must be combined (at the latest before simulation). Method |HydPy.update_devices| raises the following error if any model is unavailable: >>> from hydpy import Element, Node >>> junction = Node("junction") >>> reach1 = Element("reach1", collective="network", inlets=junction) >>> reach2 = Element("reach2", collective="network", outlets=junction) >>> hp.update_devices(elements=(reach1, reach2), nodes=junction) Traceback (most recent call last): ... hydpy.core.exceptiontools.AttributeNotReady: While trying to unite the \ elements belonging to collective `network`, the following error occurred: The model \ object of element `reach1` has been requested but not been prepared so far. After such an error, at least the passed devices are accessible: >>> hp.nodes Nodes("junction") >>> hp.elements Elements("reach1", "reach2") Use the `silent` argument if you are aware of the incompleteness of the current project configuration and want to avoid capturing the error: >>> hp.update_devices(selection=pub.selections.headwaters) >>> hp.update_devices(elements=(reach1, reach2), nodes=junction, silent=True) >>> hp.nodes Nodes("junction") >>> hp.elements Elements("reach1", "reach2") Of course, silencing the error still does not mean |HydPy.update_devices| can do the complete work, which one might realise later when trying to access one of the unprepared properties: >>> hp.deviceorder Traceback (most recent call last): ... hydpy.core.exceptiontools.AttributeNotReady: While trying to access the \ simulation order of the currently relevant devices, the following error occurred: \ While trying to unite the elements belonging to collective `network`, the following \ error occurred: The model object of element `reach1` has been requested but not been \ prepared so far. """ if (nodes is not None) or (elements is not None): if selection is not None: raise ValueError( "Method `update_devices` of class `HydPy` does not allow to use " "both the `selection` argument and the `nodes` or the `elements` " "argument at the same time." ) del self.nodes if nodes is None: nodes = devicetools.Nodes() self.nodes = nodes del self.elements if elements is None: elements = devicetools.Elements() self.elements = elements if selection is not None: self.nodes = selection.nodes self.elements = selection.elements self._update_collectives_and_deviceorder(silent=silent)
def _update_collectives_and_deviceorder(self, silent: bool): try: self._collectives = self.elements.unite_collectives() except exceptiontools.AttributeNotReady as exc: self._collectives = None self._deviceorder = None if not silent: raise exc else: graph = create_directedgraph(self.nodes, self._collectives) devices = tuple(networkx.topological_sort(graph)) names = set(self.nodes.names) names.update(self._collectives.names) self._deviceorder = tuple(d for d in devices if d.name in names) @property def deviceorder(self) -> tuple[Union[devicetools.Node, devicetools.Element], ...]: """The simulation order of the currently selected devices. |HydPy| needs to know the devices before determining their order: >>> from hydpy import HydPy >>> HydPy().deviceorder Traceback (most recent call last): ... hydpy.core.exceptiontools.AttributeNotReady: While trying to access the \ simulation order of the currently relevant devices, the following error occurred: The \ actual HydPy instance does not handle any elements at the moment. Usually, |HydPy.deviceorder| is ready after calling |HydPy.prepare_network|. |HydPy.update_devices| updates the given devices' order, too. """ try: if self._deviceorder is None: self._update_collectives_and_deviceorder(silent=False) assert (deviceorder := self._deviceorder) is not None return deviceorder except BaseException: objecttools.augment_excmessage( "While trying to access the simulation order of the currently " "relevant devices" ) @property def methodorder(self) -> list[Callable[[int], None]]: """All methods of the currently relevant |Node| and |Element| objects, which are to be processed by method |HydPy.simulate| during a simulation time step, in a working execution order. Property |HydPy.methodorder| should be of interest to framework developers only. """ # due to https://github.com/python/mypy/issues/9718: # pylint: disable=consider-using-in funcs: list[Callable[[int], None]] = [] if exceptiontools.attrready(hydpy.pub, "sequencemanager"): funcs.append(hydpy.pub.sequencemanager.read_netcdfslices) for node in self.nodes: if ( (dm := node.deploymode) == "oldsim" or dm == "obs_oldsim" or dm == "oldsim_bi" or dm == "obs_oldsim_bi" ): funcs.append(node.sequences.fastaccess.load_simdata) elif dm == "newsim" or dm == "obs" or dm == "obs_newsim" or dm == "obs_bi": funcs.append(node.sequences.fastaccess.reset) else: assert_never(dm) funcs.append(node.sequences.fastaccess.load_obsdata) for device in self.deviceorder: if isinstance(device, devicetools.Element): funcs.append(device.model.simulate) elif ( (dm := device.deploymode) == "obs_newsim" or dm == "obs_oldsim" or dm == "obs_oldsim_bi" ): funcs.append(device.sequences.fastaccess.fill_obsdata) elif not ( dm == "newsim" or dm == "oldsim" or dm == "obs" or dm == "oldsim_bi" or dm == "obs_bi" ): assert_never(dm) elements = self.collectives for element in elements: funcs.append(element.model.update_senders) for element in elements: funcs.append(element.model.update_receivers) for element in elements: funcs.append(element.model.save_data) for node in self.nodes: if ( (dm := node.deploymode) == "obs_newsim" or dm == "obs_oldsim" or dm == "obs_oldsim_bi" ): funcs.append(node.sequences.fastaccess.reset_obsdata) elif not ( dm == "newsim" or dm == "oldsim" or dm == "obs" or dm == "oldsim_bi" or dm == "obs_bi" ): assert_never(dm) funcs.append(node.sequences.fastaccess.save_simdata) funcs.append(node.sequences.fastaccess.save_obsdata) if exceptiontools.attrready(hydpy.pub, "sequencemanager"): funcs.append(hydpy.pub.sequencemanager.write_netcdfslices) return funcs
[docs] @printtools.print_progress def simulate(self) -> None: """Perform a simulation run over the actual simulation period defined by the |Timegrids| object stored in module |pub|. We let function |prepare_full_example_2| prepare a runnable |HydPy| object related to the :ref:`HydPy-H-Lahn` example project: >>> from hydpy.core.testtools import prepare_full_example_2 >>> hp, pub, TestIO = prepare_full_example_2() First, we execute a default simulation run covering the whole simulation period and inspect the discharge series simulated at the outlet of the river basin, represented by node `lahn_kalk`: >>> hp.simulate() >>> from hydpy import round_ >>> round_(hp.nodes.lahn_kalk.sequences.sim.series) 54.019337, 37.257561, 31.865308, 28.359542 After resetting the initial conditions via method |HydPy.reset_conditions|, we repeat the simulation run and get the same results: >>> hp.reset_conditions() >>> hp.simulate() >>> round_(hp.nodes.lahn_kalk.sequences.sim.series) 54.019337, 37.257561, 31.865308, 28.359542 Simulation runs do not need to cover the whole initialisation period at once. After setting the |Timegrid.lastdate| property of the `sim` |Timegrid| of the |Timegrids| objects stored within module |pub| to the middle of the initialisation period, method |HydPy.simulate| calculates the first two discharge values only: >>> hp.reset_conditions() >>> hp.nodes.lahn_kalk.sequences.sim.series = 0.0 >>> pub.timegrids.sim.lastdate = "1996-01-03" >>> hp.simulate() >>> round_(hp.nodes.lahn_kalk.sequences.sim.series) 54.019337, 37.257561, 0.0, 0.0 After adjusting both the |Timegrid.firstdate| and |Timegrid.lastdate| of the `sim` |Timegrid| to the second half of the initialisation period, |HydPy.simulate| completes the time series: >>> pub.timegrids.sim.firstdate = "1996-01-03" >>> pub.timegrids.sim.lastdate = "1996-01-05" >>> hp.simulate() >>> round_(hp.nodes.lahn_kalk.sequences.sim.series) 54.019337, 37.257561, 31.865308, 28.359542 In the above examples, each |Model| object (handled by an |Element| object) passes its simulated values via a |Node| object to its downstream |Model| object. There are four ways to deviate from this default behaviour that can be selected for each node individually via the property |Node.deploymode|. We focus on node `lahn_leun` as the upstream neighbour of node `lahn_kalk`. So far, its deploy mode is `newsim`, meaning that the node passes newly calculated simulation values to the downstream element `stream_lahn_leun_lahn_kalk`: >>> hp.nodes.lahn_leun.deploymode 'newsim' Under the second option, `oldsim`, node `lahn_leun` does not pass the discharge values simulated in the next simulation run, but the "old" discharge values already available by the |IOSequence.series| array of the |Sim| sequence. This behaviour can, for example, be useful when calibrating subsequent subareas of a river basin sequentially, beginning with the headwaters and continuing with their downstream neighbours. For the clarity of this example, we decrease all values of the "old" simulated series of node `lahn_leun` by 10 m³/s: >>> round_(hp.nodes.lahn_leun.sequences.sim.series) 42.346475, 27.157472, 22.88099, 20.156836 >>> hp.nodes.lahn_leun.deploymode = "oldsim" >>> hp.nodes.lahn_leun.sequences.sim.series -= 10.0 After performing another simulation run (over the whole initialisation period, again), the modified discharge values of node `lahn_leun` are unchanged. The simulated values of node `lahn_kalk` are, compared to the `newsim` runs, decreased by 10 m³/s (there is no time delay or dampening of the discharge values between both nodes due to the lag time of application model |musk_classic| being shorter than the simulation time step): >>> hp.reset_conditions() >>> pub.timegrids.sim.firstdate = "1996-01-01" >>> pub.timegrids.sim.lastdate = "1996-01-05" >>> hp.simulate() >>> round_(hp.nodes.lahn_leun.sequences.sim.series) 32.346475, 17.157472, 12.88099, 10.156836 >>> round_(hp.nodes.lahn_kalk.sequences.sim.series) 44.019337, 27.257561, 21.865308, 18.359542 The third option is `obs`, where node `lahn_leun` receives and stores the values from its upstream models but passes other, observed values, handled by sequence |Obs|, which we, for simplicity, set to zero for the complete initialisation and simulation period (more often, one would read measured data from files via methods as |HydPy.load_obsseries|): >>> hp.nodes.lahn_leun.deploymode = "obs" >>> hp.nodes.lahn_leun.sequences.obs.series = 0.0 Now, the simulated values of node `lahn_leun` are identical with the ones of the `newsim` example, but the simulated values of node `lahn_kalk` are lower due to receiving the observed instead of the simulated values from upstream: >>> hp.reset_conditions() >>> hp.nodes.lahn_kalk.sequences.sim.series = 0.0 >>> hp.simulate() >>> round_(hp.nodes.lahn_leun.sequences.obs.series) 0.0, 0.0, 0.0, 0.0 >>> round_(hp.nodes.lahn_leun.sequences.sim.series) 42.346475, 27.157472, 22.88099, 20.156836 >>> round_(hp.nodes.lahn_kalk.sequences.sim.series) 11.672862, 10.100089, 8.984317, 8.202706 Unfortunately, observation time series are often incomplete. *HydPy* generally uses |numpy| |numpy.nan| to represent missing values. Passing |numpy.nan| inputs to a model usually results in |numpy.nan| outputs. Hence, after assigning |numpy.nan| to some entries of the observation series of node `lahn_leun`, the simulation series of node `lahn_kalk` also contains |numpy.nan| values: >>> from numpy import nan >>> with pub.options.checkseries(False): ... hp.nodes.lahn_leun.sequences.obs.series= 0.0, nan, 0.0, nan >>> hp.reset_conditions() >>> hp.nodes.lahn_kalk.sequences.sim.series = 0.0 >>> hp.simulate() >>> round_(hp.nodes.lahn_leun.sequences.obs.series) 0.0, nan, 0.0, nan >>> round_(hp.nodes.lahn_leun.sequences.sim.series) 42.346475, 27.157472, 22.88099, 20.156836 >>> round_(hp.nodes.lahn_kalk.sequences.sim.series) 11.672862, nan, 8.984317, nan To avoid calculating |numpy.nan| values, one can select the fourth option, `obs_newsim`. Now, the priority for node `lahn_leun` is to deploy its observed values. However, for each missing observation, it deploys its newly simulated value instead: >>> hp.nodes.lahn_leun.deploymode = "obs_newsim" >>> hp.reset_conditions() >>> hp.simulate() >>> round_(hp.nodes.lahn_leun.sequences.obs.series) 0.0, nan, 0.0, nan >>> round_(hp.nodes.lahn_leun.sequences.sim.series) 42.346475, 27.157472, 22.88099, 20.156836 >>> round_(hp.nodes.lahn_kalk.sequences.sim.series) 11.672862, 37.257561, 8.984317, 28.359542 The fifth option, `obs_oldsim`, serves the same purpose as option `obs_newsim` but uses already available "old" simulation results as substitutes: >>> hp.nodes.lahn_leun.deploymode = "obs_oldsim" >>> hp.reset_conditions() >>> hp.nodes.lahn_leun.sequences.sim.series = ( ... 32.3697, 17.210443, 12.930066, 10.20133) >>> hp.simulate() >>> round_(hp.nodes.lahn_leun.sequences.obs.series) 0.0, nan, 0.0, nan >>> round_(hp.nodes.lahn_leun.sequences.sim.series) 32.3697, 17.210443, 12.930066, 10.20133 >>> round_(hp.nodes.lahn_kalk.sequences.sim.series) 11.672862, 27.310532, 8.984317, 18.404036 The last example shows that resetting option |Node.deploymode| to `newsim` results in the default behaviour of the method |HydPy.simulate| again: >>> hp.nodes.lahn_leun.deploymode = "newsim" >>> hp.reset_conditions() >>> hp.simulate() >>> round_(hp.nodes.lahn_leun.sequences.sim.series) 42.346475, 27.157472, 22.88099, 20.156836 >>> round_(hp.nodes.lahn_kalk.sequences.sim.series) 54.019337, 37.257561, 31.865308, 28.359542 """ idx_start, idx_end = hydpy.pub.timegrids.simindices methodorder = self.methodorder cm: AbstractContextManager[None] = contextlib.nullcontext() if exceptiontools.attrready(hydpy.pub, "sequencemanager"): cm = hydpy.pub.sequencemanager.provide_netcdfjitaccess(self.deviceorder) with cm: for idx in printtools.progressbar(range(idx_start, idx_end)): for func in methodorder: func(idx)
[docs] def doit(self) -> None: """Deprecated! Use method |HydPy.simulate| instead. >>> from hydpy import HydPy >>> from hydpy.core.testtools import warn_later >>> from unittest import mock >>> with warn_later(), mock.patch.object(HydPy, "simulate") as mocked: ... hp = HydPy("test") ... hp.doit() HydPyDeprecationWarning: Method `doit` of class `HydPy` is deprecated. Use \ method `simulate` instead. >>> mocked.call_args_list [call()] """ self.simulate() warnings.warn( "Method `doit` of class `HydPy` is deprecated. Use method " "`simulate` instead.", exceptiontools.HydPyDeprecationWarning, )
[docs] def prepare_allseries(self, allocate_ram: bool = True, jit: bool = False) -> None: """Tell all current |IOSequence| objects how to handle time series data. Assign |True| to the `allocate_ram` argument (default) to activate the |IOSequence.series| property of all sequences so that their time series data can become available in RAM. Assign |True| to the `jit` argument to activate the "just-in-time" reading from NetCDF files for all |InputSequence| and |Obs| objects and to activate the "just-in-time" writing of NetCDF files for all |FactorSequence|, |FluxSequence|, |StateSequence| and |Sim| objects. See the main documentation on class |HydPy| for further information. """ self.prepare_modelseries(allocate_ram=allocate_ram, jit=jit) self.prepare_nodeseries(allocate_ram=allocate_ram, jit=jit)
[docs] def prepare_modelseries(self, allocate_ram: bool = True, jit: bool = False) -> None: """An alternative method for |HydPy.prepare_allseries| specialised for model sequences.""" self.elements.prepare_allseries(allocate_ram=allocate_ram, jit=jit)
[docs] def prepare_inputseries( self, allocate_ram: bool = True, read_jit: bool = False, write_jit: bool = False ) -> None: """An alternative method for |HydPy.prepare_allseries| specialised for model input sequences.""" self.elements.prepare_inputseries( allocate_ram=allocate_ram, read_jit=read_jit, write_jit=write_jit )
[docs] def prepare_factorseries( self, allocate_ram: bool = True, write_jit: bool = False ) -> None: """An alternative method for |HydPy.prepare_allseries| specialised for model factor sequences.""" self.elements.prepare_factorseries( allocate_ram=allocate_ram, write_jit=write_jit )
[docs] def prepare_fluxseries( self, allocate_ram: bool = True, write_jit: bool = False ) -> None: """An alternative method for |HydPy.prepare_allseries| specialised for model flux sequences.""" self.elements.prepare_fluxseries(allocate_ram=allocate_ram, write_jit=write_jit)
[docs] def prepare_stateseries( self, allocate_ram: bool = True, write_jit: bool = False ) -> None: """An alternative method for |HydPy.prepare_allseries| specialised for model state sequences.""" self.elements.prepare_stateseries( allocate_ram=allocate_ram, write_jit=write_jit )
[docs] def prepare_nodeseries(self, allocate_ram: bool = True, jit: bool = False) -> None: """An alternative method for |HydPy.prepare_allseries| specialised for node sequences.""" self.nodes.prepare_allseries(allocate_ram=allocate_ram, jit=jit)
[docs] def prepare_simseries( self, allocate_ram: bool = True, read_jit: bool = False, write_jit: bool = False ) -> None: """An alternative method for |HydPy.prepare_allseries| specialised for simulation sequences of nodes.""" self.nodes.prepare_simseries( allocate_ram=allocate_ram, read_jit=read_jit, write_jit=write_jit )
[docs] def prepare_obsseries( self, allocate_ram: bool = True, read_jit: bool = False, write_jit: bool = False ) -> None: """An alternative method for |HydPy.prepare_allseries| specialised for observation sequences of nodes.""" self.nodes.prepare_obsseries( allocate_ram=allocate_ram, read_jit=read_jit, write_jit=write_jit )
[docs] def save_allseries(self) -> None: """Write the time series data of all current |IOSequence| objects at once to data file(s). See the main documentation on class |HydPy| for further information. """ self.save_modelseries() self.save_nodeseries()
[docs] def save_modelseries(self) -> None: """An alternative method for |HydPy.save_modelseries| specialised for model sequences.""" self.elements.save_allseries()
[docs] def save_inputseries(self) -> None: """An alternative method for |HydPy.save_modelseries| specialised for model input sequences.""" self.elements.save_inputseries()
[docs] def save_factorseries(self) -> None: """An alternative method for |HydPy.save_modelseries| specialised for model factor sequences.""" self.elements.save_factorseries()
[docs] def save_fluxseries(self) -> None: """An alternative method for |HydPy.save_modelseries| specialised for model flux sequences.""" self.elements.save_fluxseries()
[docs] def save_stateseries(self) -> None: """An alternative method for |HydPy.save_modelseries| specialised for model state sequences.""" self.elements.save_stateseries()
[docs] def save_nodeseries(self) -> None: """An alternative method for |HydPy.save_modelseries| specialised for node sequences.""" self.nodes.save_allseries()
[docs] def save_simseries(self) -> None: """An alternative method for |HydPy.save_modelseries| specialised for simulation sequences of nodes.""" self.nodes.save_simseries()
[docs] def save_obsseries(self) -> None: """An alternative method for |HydPy.save_modelseries| specialised for observation sequences of nodes.""" self.nodes.save_obsseries()
[docs] def load_allseries(self) -> None: """Read the time series data of all current |IOSequence| objects at once from data file(s). See the main documentation on class |HydPy| for further information. """ self.load_modelseries() self.load_nodeseries()
[docs] def load_modelseries(self) -> None: """An alternative method for |HydPy.load_modelseries| specialised for model sequences.""" self.elements.load_allseries()
[docs] def load_inputseries(self) -> None: """An alternative method for |HydPy.load_modelseries| specialised for model input sequences.""" self.elements.load_inputseries()
[docs] def load_factorseries(self) -> None: """An alternative method for |HydPy.load_modelseries| specialised for model factor sequences.""" self.elements.load_factorseries()
[docs] def load_fluxseries(self) -> None: """An alternative method for |HydPy.load_modelseries| specialised for model flux sequences.""" self.elements.load_fluxseries()
[docs] def load_stateseries(self) -> None: """An alternative method for |HydPy.load_modelseries| specialised for model state sequences.""" self.elements.load_stateseries()
[docs] def load_nodeseries(self) -> None: """An alternative method for |HydPy.load_modelseries| specialised for node sequences.""" self.nodes.load_allseries()
[docs] def load_simseries(self) -> None: """An alternative method for |HydPy.load_modelseries| specialised for simulation sequences of nodes.""" self.nodes.load_simseries()
[docs] def load_obsseries(self) -> None: """An alternative method for |HydPy.load_modelseries| specialised for observation sequences of nodes.""" self.nodes.load_obsseries()
[docs] def create_directedgraph( nodes: devicetools.Nodes, elements: devicetools.Elements ) -> networkx.DiGraph: """Create a directed graph based on the given devices.""" digraph = networkx.DiGraph() digraph.add_nodes_from(elements) digraph.add_nodes_from(nodes) for element in elements: for node in itertools.chain(element.inlets, element.inputs): digraph.add_edge(node, element) for node in itertools.chain(element.outlets, element.outputs): digraph.add_edge(element, node) return digraph