-
-
Notifications
You must be signed in to change notification settings - Fork 56
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Scope out forecast evaluation method
- Loading branch information
Showing
1 changed file
with
34 additions
and
0 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,34 @@ | ||
""" | ||
To evaluate the performance of the solar forecast, a predefined testset is used. | ||
A file has been added to this branch (make-testset) which defines a set of random timestamps and sites ids. | ||
This contains 50 sites each with 50 timestamps to make 2500 samples in total. | ||
""" | ||
|
||
def run_eval(testset_path): | ||
# load testset from csv | ||
testset = df.read_csv(testset_path) | ||
|
||
# Extract generation data and metadata for specific sites and timestamps for the testset from Hugging Face. | ||
|
||
|
||
# Split data into PV inputs and ground truth. | ||
|
||
|
||
# Collect NWP data from Hugging Face, ICON. | ||
|
||
|
||
# Run forecast with PV and NWP inputs. | ||
|
||
|
||
# Combine the forecast results with the ground truth (ts, id, horizon (in hours), pred, truth, diff) | ||
|
||
|
||
# Save file | ||
|
||
|
||
# Calculate and print metrics: MAE | ||
|
||
|
||
# Visulisations |