dispaset.preprocessing package

Submodules

dispaset.preprocessing.data_check module

This files gathers different functions used in the DispaSET to check the input data

__author__ = ‘Sylvain Quoilin (sylvain.quoilin@ec.europa.eu)’

dispaset.preprocessing.data_check.check_AvailabilityFactors(plants, AF)[source]

Function that checks the validity of the provided availability factors and warns if a default value of 100% is used.

dispaset.preprocessing.data_check.check_FlexibleDemand(flex)[source]

Function that checks the validity of the provided flexibility demand time series

dispaset.preprocessing.data_check.check_MinMaxFlows(df_min, df_max)[source]

Function that checks that there is no incompatibility between the minimum and maximum flows

dispaset.preprocessing.data_check.check_chp(config, plants)[source]

Function that checks the CHP plant characteristics

dispaset.preprocessing.data_check.check_clustering(plants, plants_merged)[source]

Function that checks that the installed capacities are still equal after the clustering process

Parameters:
  • plants – Non-clustered list of units
  • plants_merged – clustered list of units
dispaset.preprocessing.data_check.check_df(df, StartDate=None, StopDate=None, name='')[source]

Function that check the time series provided as inputs

dispaset.preprocessing.data_check.check_heat_demand(plants, data)[source]

Function that checks the validity of the heat demand profiles

Parameters:plants – List of CHP plants
dispaset.preprocessing.data_check.check_p2h(config, plants)[source]

Function that checks the p2h unit characteristics

dispaset.preprocessing.data_check.check_simulation_environment(SimulationPath, store_type='pickle', firstline=7)[source]

Function to test the validity of disapset inputs :param SimulationPath: Path to the simulation folder :param store_type: choose between: “list”, “excel”, “pickle” :param firstline: Number of the first line in the data (only if type==’excel’)

dispaset.preprocessing.data_check.check_sto(config, plants, raw_data=True)[source]

Function that checks the storage plant characteristics

dispaset.preprocessing.data_check.check_temperatures(plants, Temperatures)[source]

Function that checks the presence and validity of the temperatures profiles for units with temperature-dependent characteristics

Parameters:plants – List of all units
dispaset.preprocessing.data_check.check_units(config, plants)[source]

Function that checks the power plant characteristics

dispaset.preprocessing.data_check.isStorage(tech)[source]

Function that returns true the technology is a storage technology

dispaset.preprocessing.data_check.isVRE(tech)[source]

Function that returns true the technology is a variable renewable energy technology

dispaset.preprocessing.data_handler module

dispaset.preprocessing.data_handler.NodeBasedTable(varname, config, default=None)[source]

This function loads the tabular data stored in csv files relative to each zone of the simulation.

Parameters:
  • varname – Variable name (as defined in config)
  • idx – Pandas datetime index to be used for the output
  • zones – List with the zone codes to be considered
  • fallback – List with the order of data source.
  • default – Default value to be applied if no data is found
Returns:

Dataframe with the time series for each unit

dispaset.preprocessing.data_handler.UnitBasedTable(plants, varname, config, fallbacks=['Unit'], default=None, RestrictWarning=None)[source]

This function loads the tabular data stored in csv files and assigns the proper values to each unit of the plants dataframe. If the unit-specific value is not found in the data, the script can fallback on more generic data (e.g. fuel-based, technology-based, zone-based) or to the default value. The order in which the data should be loaded is specified in the fallback list. For example, [‘Unit’,’Technology’] means that the script will first try to find a perfect match for the unit name in the data table. If not found, a column with the unit technology as header is search. If not found, the default value is assigned.

Parameters:
  • plants – Dataframe with the units for which data is required
  • varname – Variable name (as defined in config)
  • idx – Pandas datetime index to be used for the output
  • zones – List with the zone codes to be considered
  • fallback – List with the order of data source.
  • default – Default value to be applied if no data is found
  • RestrictWarning – Only display the warnings if the unit belongs to the list of technologies provided in this parameter
Returns:

Dataframe with the time series for each unit

dispaset.preprocessing.data_handler.define_parameter(sets_in, sets, value=0)[source]

Function to define a DispaSET parameter and fill it with a constant value

Parameters:
  • sets_in – List with the labels of the sets corresponding to the parameter
  • sets – dictionary containing the definition of all the sets (must comprise those referenced in sets_in)
  • value – Default value to attribute to the parameter
dispaset.preprocessing.data_handler.export_yaml_config(ExcelFile, YAMLFile)[source]

Function that loads the DispaSET excel config file and dumps it as a yaml file.

Parameters:
  • ExcelFile – Path to the Excel config file
  • YAMLFile – Path to the YAML config file to be written
dispaset.preprocessing.data_handler.load_config(ConfigFile, AbsPath=True)[source]

Wrapper function around load_config_excel and load_config_yaml

dispaset.preprocessing.data_handler.load_config_excel(ConfigFile, AbsPath=True)[source]

Function that loads the DispaSET excel config file and returns a dictionary with the values

Parameters:
  • ConfigFile – String with (relative) path to the DispaSET excel configuration file
  • AbsPath – If true, relative paths are automatically changed into absolute paths (recommended)
dispaset.preprocessing.data_handler.load_config_yaml(filename, AbsPath=True)[source]

Loads YAML file to dictionary

dispaset.preprocessing.data_handler.load_time_series(config, path, header='infer')[source]

Function that loads time series data, checks the compatibility of the indexes and guesses when no exact match between the required index and the data is present

dispaset.preprocessing.data_handler.merge_series(plants, data, mapping, method='WeightedAverage', tablename='')[source]

Function that merges the times series corresponding to the merged units (e.g. outages, inflows, etc.)

Parameters:
  • plants – Pandas dataframe with the information relative to the original units
  • data – Pandas dataframe with the time series and the original unit names as column header
  • mapping – Mapping between the merged units and the original units. Output of the clustering function
  • method – Select the merging method (‘WeightedAverage’/’Sum’)
  • tablename – Name of the table being processed (e.g. ‘Outages’), used in the warnings
Return merged:

Pandas dataframe with the merged time series when necessary

dispaset.preprocessing.data_handler.read_truefalse(sheet, rowstart, colstart, rowstop, colstop, colapart=1)[source]

Function that reads a two column format with a list of strings in the first columns and a list of true false in the second column The list of strings associated with a True value is returned

dispaset.preprocessing.preprocessing module

This is the main file of the DispaSET pre-processing tool. It comprises a single function that generated the DispaSET simulation environment.

@author: S. Quoilin

dispaset.preprocessing.preprocessing.build_simulation(config, mts_plot=None)[source]

Dispa-SET function that builds different simulation environments based on the hydro scheduling option in the config file Hydro scheduling options:

  • Off - Hydro scheduling turned off, normal call of BuildSimulation function
  • Zonal - Zonal variation of hydro scheduling, if zones are not individually specified in a list (e.a. zones = [‘AT’,’DE’]) hydro scheduling is imposed on all active zones from the Config file
  • Regional - Regional variation of hydro scheduling, if zones from a specific region are not individually specified in a list (e.a. zones = [‘AT’,’DE’]), hydro scheduling is imposed on all active zones from the Config file simultaneously
Config:Read config file
Zones_mts:List of zones where new reservoir levels should be calculated eg. [‘AT’,’BE’,…’UK’]
Mts_plot:If ms_plot = True indicative plot with temporary computed reservoir levels is displayed
dispaset.preprocessing.preprocessing.get_temp_sim_results(config, gams_dir=None)[source]

This function reads the simulation environment folder once it has been solved and loads the input variables together with the results.

Parameters:
  • path – Relative path to the simulation environment folder (current path by default)
  • cache – If true, caches the simulation results in a pickle file for faster loading the next time
  • temp_path – Temporary path to store the cache file
Returns inputs,results:
 

Two dictionaries with all the outputs

dispaset.preprocessing.preprocessing.mid_term_scheduling(config, zones, profiles=None)[source]

This function reads the DispaSET config file, searches for active zones, loads data for each zone individually and solves model using UCM_h_simple.gms

Config:Read config file

dispaset.preprocessing.utils module

dispaset.preprocessing.utils.EfficiencyTimeSeries(config, plants, Temperatures)[source]

Function that calculates an efficiency time series for each unit In case of generation unit, the efficiency is constant in time (for now) In case of of p2h units, the efficicncy is defined as the COP, which can be temperature-dependent or not. If it is temperature-dependent, the formula is: \(COP = COP_{nom} + coef_a (T-T_{nom}) + coef_b (T-T_{nom})^2\)

Parameters:
  • plants – Pandas dataframe with the original list of units
  • Temperatures – Dataframe with the temperature for all relevant units
Returns:

Dataframe with a time series of the efficiency for each unit

dispaset.preprocessing.utils.adjust_capacity(inputs, tech_fuel, scaling=1, value=None, singleunit=False, write_gdx=False, dest_path='')[source]

Function used to modify the installed capacities in the Dispa-SET generated input data The function update the Inputs.p file in the simulation directory at each call

Parameters:
  • inputs – Input data dictionary OR path to the simulation directory containing Inputs.p
  • tech_fuel – tuple with the technology and fuel type for which the capacity should be modified
  • scaling – Scaling factor to be applied to the installed capacity
  • value – Absolute value of the desired capacity (! Applied only if scaling != 1 !)
  • singleunit – Set to true if the technology should remain lumped in a single unit
  • write_gdx – boolean defining if Inputs.gdx should be also overwritten with the new data
  • dest_path – Simulation environment path to write the new input data. If unspecified, no data is written!
Returns:

New SimData dictionary

dispaset.preprocessing.utils.adjust_storage(inputs, tech_fuel, scaling=1, value=None, write_gdx=False, dest_path='')[source]

Function used to modify the storage capacities in the Dispa-SET generated input data The function update the Inputs.p file in the simulation directory at each call

Parameters:
  • inputs – Input data dictionary OR path to the simulation directory containing Inputs.p
  • tech_fuel – tuple with the technology and fuel type for which the capacity should be modified
  • scaling – Scaling factor to be applied to the installed capacity
  • value – Absolute value of the desired capacity (! Applied only if scaling != 1 !)
  • write_gdx – boolean defining if Inputs.gdx should be also overwritten with the new data
  • dest_path – Simulation environment path to write the new input data. If unspecified, no data is written!
Returns:

New SimData dictionary

dispaset.preprocessing.utils.clustering(plants, method='Standard', Nslices=20, PartLoadMax=0.1, Pmax=30)[source]

Merge excessively disaggregated power Units.

Parameters:
  • plants – Pandas dataframe with each power plant and their characteristics (following the DispaSET format)
  • method – Select clustering method (‘Standard’/’LP’/None)
  • Nslices – Number of slices used to fingerprint each power plant characteristics. slices in the power plant data to categorize them (fewer slices involves that the plants will be aggregated more easily)
  • PartLoadMax – Maximum part-load capability for the unit to be clustered
  • Pmax – Maximum power for the unit to be clustered
Returns:

A list with the merged plants and the mapping between the original and merged units

dispaset.preprocessing.utils.incidence_matrix(sets, set_used, parameters, param_used)[source]

This function generates the incidence matrix of the lines within the nodes A particular case is considered for the node “Rest Of the World”, which is no explicitely defined in DispaSET

dispaset.preprocessing.utils.interconnections(Simulation_list, NTC_inter, Historical_flows)[source]

Function that checks for the possible interconnections of the zones included in the simulation. If the interconnections occurs between two of the zones defined by the user to perform the simulation with, it extracts the NTC between those two zones. If the interconnection occurs between one of the zones selected by the user and one country outside the simulation, it extracts the physical flows; it does so for each pair (country inside-country outside) and sums them together creating the interconnection of this country with the RoW.

Parameters:
  • Simulation_list – List of simulated zones
  • NTC – Day-ahead net transfer capacities (pd dataframe)
  • Historical_flows – Historical flows (pd dataframe)
dispaset.preprocessing.utils.select_units(units, config)[source]

Function returning a new list of units by removing the ones that have unknown technology, zero capacity, or unknown zone

Parameters:
  • units – Pandas dataframe with the original list of units
  • config – Dispa-SET config dictionnary
Returns:

New list of units

Module contents