-
Notifications
You must be signed in to change notification settings - Fork 354
Streamline workflow #1838
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: master
Are you sure you want to change the base?
Streamline workflow #1838
Conversation
- Add test_streamlined.smk with 4-step workflow (base → clustered → composed → solved) - Support all three foresight modes: overnight, myopic, perfect - Add robust horizon handling for single values and lists - Update test configs with proper temporal and clustering sections - Add comprehensive test configurations for all foresight modes - Add implementation plan and documentation Features: - Config-driven approach (no wildcard expansion) - Dynamic input resolution based on foresight and horizons - Sequential dependencies for myopic/perfect foresight - Unified rule structure across all modes - Robust parameter handling with defaults Tested workflows: - Overnight: 6 jobs, single horizon optimization - Myopic: 10 jobs, sequential horizon optimization with brownfield - Perfect: 8 jobs, multi-period simultaneous optimization Refs: GitHub Discussion #1529
- Add Status Update section documenting completed proof of concept - Update timeline to reflect completed Phase 0 (POC) and current focus - Document validated implementation patterns from testing experience - Add robust parameter handling and dynamic input resolution examples - Update success criteria to distinguish completed vs remaining goals - Reorganize open questions to focus on remaining implementation details - Document created test files and artifacts Key achievements validated: - 4-step workflow architecture working across all foresight modes - Config-driven approach with temporal section replacing scenario wildcards - Dynamic dependencies with proper sequential chaining - Robust error handling for missing config sections - Flexible horizon handling for single values and lists Next phase: Implement actual scripts with real PyPSA network operations
- Create scripts/compose_network.py that combines all network building steps
in one big main section without additional function definitions
- Implement network concatenation logic for perfect foresight mode
- Add brownfield constraints for myopic optimization
- Create production rules/compose.smk with dynamic dependencies
- Update Snakefile to include compose rules when streamlined_workflow is enabled
- Add temporal configuration section to default config
- Create example configuration file demonstrating the new workflow
The new workflow follows the 4-step structure:
base → clustered → composed_{horizon} → solved_{horizon}
All configuration is now driven by config sections rather than wildcards,
simplifying the workflow and making it more maintainable.
🤖 Generated with [Claude Code](https://claude.ai/code)
Co-Authored-By: Claude <[email protected]>
CRITICAL FIXES:
- Fix snapshot handling bug: prevent overwriting multi-period snapshots in perfect foresight
- Fix function signature error in load_and_aggregate_powerplants call
- Remove resource path conflict by using existing path provider system
- Simplify overly complex perfect foresight concatenation logic
- Fix parameter access patterns throughout the script
- Improve error handling for missing config sections
RESOURCE PATH UPDATES:
- Update all resource references to use existing path patterns with {clusters} wildcards
- Remove custom resources() function that conflicted with existing system
- Align file paths with existing workflow conventions
ROBUSTNESS IMPROVEMENTS:
- Add proper error handling for missing inputs
- Simplify carbon budget handling to avoid complex logic failures
- Use safer parameter access patterns with fallbacks
These fixes address the most serious issues that would have prevented
the workflow from functioning correctly.
🤖 Generated with [Claude Code](https://claude.ai/code)
Co-Authored-By: Claude <[email protected]>
- Fix critical wildcard mismatch: resource paths used {clusters} but rules didn't have this wildcard
- Get cluster count from config and substitute into resource paths
- Add missing cluster_network rule to generate clustered.nc input
- Fix all resource path references to use config-derived cluster values
- Ensure proper dependency chain: base → clustered → composed → solved
This resolves the fundamental mismatch between wildcard-based resource
paths and the new config-driven workflow approach.
🤖 Generated with [Claude Code](https://claude.ai/code)
Co-Authored-By: Claude <[email protected]>
STANDARDIZATION IMPROVEMENTS: - Replace all direct config access with config_provider function calls - Ensure consistent use of resources() function from existing path provider - Follow PyPSA-EUR conventions for parameter handling in Snakemake rules - Use proper wildcard handling patterns from existing codebase KEY CHANGES: - get_compose_inputs: Use config_provider for all config value retrieval - cluster_network rule: Convert all params to use config_provider - compose_network rule: Standardize all parameter definitions - solve_network rule: Use config_provider for solver configuration - Collection rules: Use config_provider for run name retrieval - Validation rule: Standardize parameter access patterns This ensures the streamlined workflow follows the same patterns and conventions as the existing PyPSA-EUR workflow implementation. 🤖 Generated with [Claude Code](https://claude.ai/code) Co-Authored-By: Claude <[email protected]>
|
Thanks for starting this. Had only a short look so far. I guess it is too early for reviewing. How do you think we can best support?! I have opinions about many of the "critical open questions". |
|
You are totally right, it is too early for reviewing. I guess at this stage, it is best to think about whether I can proceed with this approach or if there is potential red flags, which speak against this strategy in the first place. But perhaps I spend more time on refining this and make a proper report for you guys what are this wider implications. |
I must admit that the section |
…y network creation
…ed workflow - Rename set_line_nom_max to cap_transmission_capacity with clearer parameter names - s_nom_max_set → line_max (AC line capacity limit) - p_nom_max_set → link_max (DC link capacity limit) - s_nom_max_ext → line_max_extension (AC line extension limit) - p_nom_max_ext → link_max_extension (DC link extension limit) - Port cap_transmission_capacity to compose_network.py for network composition - Port enforce_autarky to solve_network.py for pre-solve constraints - Add cap_transmission_capacity config section to config.default.yaml - Remove .get() calls in favor of strict dictionary access for better error messages - All functionality now config-based instead of wildcard-based 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <[email protected]>
- Add adjust_renewable_capacity_limits() function to compose_network.py for brownfield scenarios in myopic foresight mode - Function subtracts existing renewable capacities from previous horizons from p_nom_max values in current horizon - Uses renewable_carriers list from config (not hardcoded) - Remove redundant add_land_use_constraint() from solve_network.py for myopic mode (functionality now in compose step) - Keep add_land_use_constraint() for perfect foresight (different logic) - Add comprehensive test suite with 8 test cases covering: * Basic capacity subtraction * Multiple existing generators * Cases where existing exceeds potential * Negative value clipping * All renewable carriers * Edge cases (no existing, non-renewables) * Custom carrier lists - Update compose.smk to pass renewable_carriers as param 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <[email protected]>
Remove InputFiles dataclass and 60-line manual mapping code. Use snakemake.input directly in all functions. Reduces code by 94 lines while maintaining full functionality across all workflows.
The old expression ('transmission_expansion') is no longer in the config file and therefore fails to overwrite the 'transmission_limit' if it's used in the 'opts'
* Add option to set minimum unit dispatch for electrolysis * electrolysis p_min_pu documentation
Bumps the github-actions group with 2 updates: [actions/upload-artifact](https://github.com/actions/upload-artifact) and [actions/download-artifact](https://github.com/actions/download-artifact). Updates `actions/upload-artifact` from 4 to 5 - [Release notes](https://github.com/actions/upload-artifact/releases) - [Commits](actions/upload-artifact@v4...v5) Updates `actions/download-artifact` from 5 to 6 - [Release notes](https://github.com/actions/download-artifact/releases) - [Commits](actions/download-artifact@v5...v6) --- updated-dependencies: - dependency-name: actions/upload-artifact dependency-version: '5' dependency-type: direct:production update-type: version-update:semver-major dependency-group: github-actions - dependency-name: actions/download-artifact dependency-version: '6' dependency-type: direct:production update-type: version-update:semver-major dependency-group: github-actions ... Signed-off-by: dependabot[bot] <[email protected]> Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
# Conflicts: # .gitignore # Makefile # Snakefile # config/test/config.myopic.yaml # config/test/config.overnight.yaml # config/test/config.perfect.yaml # doc/release_notes.rst # envs/linux-64.lock.yaml # envs/osx-64.lock.yaml # envs/osx-arm64.lock.yaml # envs/win-64.lock.yaml # rules/build_electricity.smk # rules/build_sector.smk # rules/solve_myopic.smk # rules/solve_perfect.smk # scripts/_helpers.py # scripts/add_electricity.py # scripts/make_summary_perfect.py # scripts/prepare_sector_network.py
Remove unnecessary wildcards (opts, clusters, sector_opts) from mock_snakemake. Standardize planning_horizons to horizon parameter with numeric values. Apply strict indexing for required params, keep .get() for optional config.
…ctor - Replace separate resolution_elec and resolution_sector settings with single clustering.temporal.resolution key - Update compose_network.py and _helpers.py to use unified key - Update all test configs and documentation
- Remove electricity.co2limit_enable, co2limit, co2base config options - Remove Co2L opts wildcard parsing from _helpers.py - Migrate test configs to use co2_budget with upper/lower bounds - Update release notes with comprehensive PR #1838 breaking changes - Update documentation (electricity.csv, opts.csv)
- Delete opts.csv and sector-opts.csv (wildcards now config-driven) - Remove Co2L and cb* wildcard parsing from sector_opts in _helpers.py - Update release notes to document cb* wildcard removal
Streamlined and restructured workflow
This PR refactors the workflow in order to streamline and unify optimization approaches and configuration settings. here are some highlights of the PR, otherwise no one will have the motivation to look at this large change. I suggest before going into the code, read this here first, even though it is long...
Highlights
base.nc->clustered.nc->composed_{horizon}.nc->solved_{horizon}.nccsvs/nodal_costs.csv,csvs/capacities.csv...)composerule is the entry point for green-field, brown-field, perfect foresight model for both sector-coupled and electricity-only.and now a more descriptive summary
Overview
base → simplified → clustered → composed → solved) and standardises filenames so scenario information now lives in configuration instead of wildcards.*_s_{clusters}or..._{planning_horizons}artefacts have been renamed; the new naming scheme makes horizons explicit via{horizon}and removes cluster suffixes.Workflow Changes
networks/base.nc→networks/simplified.nc→networks/clustered.nc→networks/composed_{horizon}.nc→RESULTS/networks/solved_{horizon}.nc. Old intermediate targets such asnetworks/base_s_{clusters}_{opts}_{sector_opts}_{planning_horizons}.ncno longer exist.rules/compose.smkencapsulates what used to be multipleprepare_*,add_*, and brownfield rules; it assembles everything needed for a givenhorizonand handles the myopic/perfect brownfield inputs.rules/solve.smkcontains onesolve_networkrule; electricity-only and sector-coupled cases are distinguished by config, not by separate rule files (solve_*smks were deleted).rules/collect.smknow checks four milestones: clustered networks, composed networks, solved networks, and plotting;{horizon}replaces{planning_horizons}wildcards in file names.navigate_config,get_full_config, andget_config(all inrules/common.smk) cache fully merged configs per wildcard set. Any new rules should obtain parameters throughconfig_provider(...)orget_config(w)to stay scenario compatible.RESULTS/networks/solved_{prev}.nc, while perfect foresight reuses the priornetworks/composed_{prev}.nc. Overnight runs must supply a single planning horizon.rules/postprocess.smknow readsRESULTS/networks/solved_{horizon}.ncfor all foresight modes, so every map (power_network_{horizon}.pdf,h2_network_{horizon}.pdf,{carrier}_balance_map_{horizon}.pdf, etc.) and CSV summary is generated with the same naming scheme regardless of foresight setting.Stage Notes
rules/build_electricity.smkwritesregions_onshore_base.geojson/regions_offshore_base.geojsontogether withnetworks/base.nc. Administrative shapes and OSM/TYNDP inputs remain unchanged.simplify_networknow emitsnetworks/simplified.nc,regions_onshore_simplified.geojson,regions_offshore_simplified.geojson, andbusmap_simplified.csv.build_electricity_demand_baseconsumes those files and produceselectricity_demand_simplified.nc.process_cost_datareads per-horizoncosts_{horizon}.csv.cluster_networktakesnetworks/simplified.ncplusbusmap_simplified.csvand emitsnetworks/clustered.nc,regions_onshore.geojson,regions_offshore.geojson,busmap.csv, andlinemap.csv. Cluster counts are configured, not embedded in filenames._s_{clusters}suffix; ie. population layouts collapse topop_layout.csvandpop_layout_simplified.csv; Gas locations becomegas_input_locations.geojson/gas_input_locations_simplified.csv; Powerplant list becomespowerplants.csvetc.compose_network(one rule) flattens all previousadd_existing_baseyear,add_brownfield,prepare_perfect_*, etc. It automatically wires the previous-horizon inputs depending on foresight and provides the final pre-solve network (networks/composed_{horizon}.nc).solve_networkreadsnetworks/composed_{horizon}.nc, writesRESULTS/networks/solved_{horizon}.nc, and stores solver/memory/python logs underRESULTS/logs/solve_network/. Custom extra functionality and solver settings are driven purely by config.maps/base_network.pdfandmaps/clustered_network.pdf; solved outputs areRESULTS/maps/power_network_{horizon}.pdf,.../h2_network_{horizon}.pdf,.../ch4_network_{horizon}.pdf, and.../{carrier}_balance_map_{horizon}.pdf.make_summaryandplot_summaryoperate onRESULTS/networks/solved_{horizon}.ncregardless of foresight.Config Changes
planning_horizonsdirectly under the root of your config (config/config.default.yaml:34-35). The oldscenario.planning_horizonsentry is ignored, and the workflow expectsconfig["planning_horizons"]to exist even when scenarios are disabled. Configuration bundles such asconfig/test/config.scenarios.yaml:17-20already follow this format.scenarioblock removed – The legacyscenario:section (withclusters,opts,sector_opts, etc.) is no longer part ofconfig.default.yaml. Scenario sweeps should now be described viarun.scenariosplus the dedicated scenario YAML file; individual dimensions (e.g. clusters) are configured directly under their respective sections. If you keep ascenarioblock in a local config it will simply be ignored.co2_budgetnow specifies anemissions_scope, avaluesinterpretation flag, and nestedupper/lowerdictionaries with their ownenableswitches (config/config.default.yaml:90-113). Update custom configs accordingly if you previously listed plain year-value pairs.linesusess_nom_max/s_nom_max_extension, andlinksusesp_nom_max/p_nom_max_extension(config/config.default.yaml:320-355). Rename any overrides that still refer tomax_extensionso the new limits are applied.electricity.exclude_carriers(config/config.default.yaml:115-162) lets you strip carriers during clustering; custom busmap logic should now read that list instead of hard-coding exclusions.existing_capacities.enabled(config/config.default.yaml:479-486). Set it totruewhen you wantcompose_networkto import historical assets; otherwise the new workflow assumes a greenfield build.Script Changes
scripts/compose_network.pycombines function calls from multiple preparatory scripts (add_electricity.py,add_existing_baseyear.py,add_brownfield.py,prepare_network.py,prepare_sector_network.py,prepare_perfect_foresight.py). At a later stage a clearer packaging structure should be used here.compose_network.py.scripts/solve_network.pyis the only solver script called from Snakemake;solve_operations_network.py(operations-only runs) is no longer referenced by the workflow.scripts/make_summary.py– the script now loads all horizons viapypsa.NetworkCollectionfor overnight, myopic, and perfect runs, so the dedicated helpersscripts/make_summary_perfect.pyandscripts/make_global_summary.pywere deleted. Any custom tooling should invokemake_summary.pyand read the unified CSV outputs.File Name Mapping
networks/base_s.ncnetworks/simplified.ncregions_onshore_simplified.geojson,regions_offshore_simplified.geojson, andbusmap_simplified.csv.regions_onshore_base_s_{clusters}.geojson/regions_offshore_base_s_{clusters}.geojsonregions_onshore.geojson/regions_offshore.geojsoncluster_network; no{clusters}wildcard in filename.busmap_base_s_{clusters}.csv/linemap_base_s_{clusters}.csvbusmap.csv/linemap.csvpowerplants_s_{clusters}.csvpowerplants_s.csvnetworks/clustered.nc.electricity_demand_base_s.ncelectricity_demand_simplified.ncavailability_matrix_{clusters}_{technology}.ncavailability_matrix_{technology}.ncprofile_{clusters}_{technology}.ncprofile_{technology}.ncregions_by_class_*follow the same pattern.pop_layout_base_s_{clusters}.csv/pop_layout_base_s.csvpop_layout.csv/pop_layout_simplified.csvgas_input_locations_s_{clusters}.geojsongas_input_locations.geojsoncosts_{planning_horizons}.csv/_processed.csvcosts_{horizon}.csv/_processed.csvcollect.smkexpands over{horizon}.networks/base_s_{clusters}_{opts}_{sector_opts}_{planning_horizons}.nc(+_brownfield*)networks/composed_{horizon}.ncRESULTS/networks/base_s_{clusters}_{opts}_{sector_opts}_{planning_horizons}.ncRESULTS/networks/solved_{horizon}.ncmaps/power-network.pdf/maps/power-network-s-{clusters}.pdfmaps/base_network.pdf/maps/clustered_network.pdfRESULTS/maps/base_s_{clusters}_{opts}_{sector_opts}-costs-all_{planning_horizons}.pdfRESULTS/maps/power_network_{horizon}.pdfMigration Checklist
*_base_s*or*_base_s_{clusters}_*targets with the new filenames; consumenetworks/composed_{horizon}.nc/RESULTS/networks/solved_{horizon}.ncinstead of the legacybase_s_*artefacts.composerule incompose.smkand update to the new wildcard convention (only keep{horizon}wildcard)compose_network.pyreference the old scripts to clearly indicate where to add the changes.{horizon}-based filenames and the new CSV outputs frommake_summary.RESULTS/networks/solved_{prev}.nc; perfect foresight helpers should import the previousnetworks/composed_{prev}.nc. Do not craft filenames manually.Checklist
envs/environment.yaml.config/config.default.yaml.doc/configtables/*.csv.doc/data_sources.rst.doc/release_notes.rstis added.