Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
* Updated model constants (primarily loss years/model years) for annual update. Also tentatively updated input and output folder dates on s3 from 2021 to 2022. Finally, refactored "gain" folder to "removals" folder (don't know if refactor worked entirely correctly but it did results in "gain" being changed to "removals" in comments and variables throughout the model). * Updated burn area scripts for 2022 model run. * Ran model successfully on 00N_000E locally in Docker, with a combination of constants for v1.2.1 and v1.2.2 in order to make it run to completion (but not necessarily produce reasonable results). Now going to run on a spot machine. * Ran model successfully on 00N_000E locally in Docker, with a combination of constants for v1.2.1 and v1.2.2 in order to make it run to completion (but not necessarily produce reasonable results). Now going to run on a spot machine. * Updating MODIS burned area for 2021 tree cover loss. Only running steps 1-3 for now because need TCL update for step 4. * Updating MODIS burned area for 2021 tree cover loss. Only running steps 1-3 for now because need TCL update for step 4. * Working on burn year update for 2021. * Working on burn year update for 2021. On step 3 now. * Did steps 1-3 of burn year update. Waiting on 2021 TCL for step 4 of burn year update. * Experimenting with tracking memory during model run. I can't figure out how to have a memory profiler that runs concurrently with the model processing (tracemalloc doesn't seem to do the thing I want), so I'm just checking memory at specific points and hopefully that's representative of peak memory usage. * Experimenting with tracking memory during model run. Using my new psutil-based memory tracker in uu.end_of_fx_summary seems to capture peak memory usage. I checked that in a local test and in a r5d.4xlarge instance running 3 tiles. Both using mp_model_extent stage. Haven't tried memory tracking using other stages yet. * Added memory tracking to all model steps used in run_full_model.py. I've only tested it in mp_model_extent.py but it should work similarly elsewhere. Used htop memory graph for validation for psutil memory value. * Changed postgres steps for plantation preprocessing to work in r5d instance. Dockerfile now sets up the correct environment and permissions but the actual database creation and postgis extension addition has to be done manually in the container. I can't figure out how to get the container to launch with postgres running, the database created, or postgis exension added. Anyhow, this is still workable. I tested out adding plantation data to the postgres database and it seemed fine, although I didn't then try running the plantation script to rasterize that. So there could easily be issues with the postgres database; I just don't know. * test commit. changed plantation preparation. * Trying full model run with 2021 TCL. Theoretically will run all the way through, but using 2020 drivers and burned area. Will update burned area through 2021 later on. * Trying full model run with 2021 TCL. Theoretically will run all the way through, but using 2020 drivers and burned area. Will update burned area through 2021 later on. * Changing memory warning threshold to be higher. * Changing processor count for model extent. * Issue with full model run-- not finding s3 folder correctly. Trying with two test tiles first. * Experimenting with full model run again. * Experimenting with full model run again. * 96% memory warning threshold led to unnecessary program termination. Increasing to 99%. That's because the percentage reported by psutil.virtual_memory() doesn't exactly equal what's shown in htop. * Reducing processor count for deadwood/litter creation. * Finishing burn year update for 2021 now that TCL 2021 is available. * Changed directory and names for updated drivers. * Changed directory and names for updated drivers. * Trying to rewindow the tcd, gain, etc. tiles again. * Trying to run aggregation step again. * Trying to run aggregation step again. * Issue with 20N_030W in emissions aggregation step. Never rewindowed and can't tell why. * Issue with 10N_150E net flux rewindowing. Randomly didn't happen. Trying net flux tile aggregation again. * Having rewindowed area tiles from aggregation step on the spot machine made the spot machine think it already had the area tiles, even though it had the rewindowed version. So adding the deletion of all rewindowed tiles after aggregate maps are made. * Wrote a script that downloads the tiles using a list of directories and patterns for a list of tile_ids. It then builds overviews for them for easy viewing in ArcMap. This is just to simplify the process of downloading all the intermediate and final outputs of the model for QCing (don't have to download them individually using Cloudberry). * Running full model, hopefully from start to finish. Using 2021 TCL with drivers and burn year updated. * Running full model, from gross emissions aggregation onwards. Emissions aggregation skipped 20N_030W again, so the model stopped. This happened with the same tile before. Still don't know what the deal is. * Running full model, from net flux aggregation onwards. Net flux aggregation skipped another tile again (I think it was 10N_150E but I actually didn't record it-- oops). This happened with net flux before. Still don't know what the deal is. * Revised download_tile_set.py to create overviews (.ovr) for tile set that are smaller (DEFLATE compression). * Rerunning drivers preparation and emissions onwards because of corrected drivers map. * Rerunning drivers preparation and emissions onwards because of corrected drivers map. * Rerunning emissions aggregation onwards because of corrected drivers map. * Rerunning net flux aggregation onwards because of corrected drivers map. * Creating soil C emissions for 2021 update (v1.2.2). * Updated readme for model update with 2021 TCL (v.1.2.2). This is the model version as I ran it for the 2021 update.
- Loading branch information