Skip to content

Commit

Permalink
Formatting and fixing of links
Browse files Browse the repository at this point in the history
  • Loading branch information
NoraLoose committed Aug 22, 2023
1 parent ee74450 commit 5cb61b9
Show file tree
Hide file tree
Showing 4 changed files with 8 additions and 8 deletions.
4 changes: 2 additions & 2 deletions website/dataset.md
Original file line number Diff line number Diff line change
Expand Up @@ -21,10 +21,10 @@ Data from multi-scale climate model (E3SM-MMF) simulations were saved at 20-minu
- 1.9 MB per input file, 1.1 MB per output file

Input files are labeled ```E3SM-MMF.mli.YYYY-MM-DD-SSSSS.nc```, where ```YYYY-MM-DD-SSS``` corresponds to the simulation year (```YYYY```), month (``MM``), day of the month (``DD``), and seconds of the day (```SSSSS```), with timesteps being spaced 1,200 seconds (20 minutes) apart. Target files are labeled the same way, except ```mli``` is replaced by ```mlo```.
Scalar variables vary in time and "horizontal" grid ("ncol"), while vertically-resolved variables vary additionally in vertical space ("lev"). For vertically-resolved variables, lower indices of "lev" corresponds to higher levels in the atmosphere. This is because pressure decreases monotonically with altitude.
Scalar variables vary in time and "horizontal" grid (`ncol`), while vertically-resolved variables vary additionally in vertical space (`lev`). For vertically-resolved variables, lower indices of `lev` corresponds to higher levels in the atmosphere. This is because pressure decreases monotonically with altitude.

The full list of variables can be found in [Supplementary Information](https://arxiv.org/pdf/2306.08754.pdf), Table 1.

There is also a [Quickstart dataset](https://huggingface.co/datasets/LEAP/subsampled_low_res) that contains subsampled and prenormalized data. This data was used for training, validation, and metrics for the ClimSim paper and can be reproduced using the ```preprocessing/create_npy_data_splits.ipynb``` notebook.
There is also a [**Quickstart dataset**](https://huggingface.co/datasets/LEAP/subsampled_low_res) that contains subsampled and prenormalized data. This data was used for training, validation, and metrics for the ClimSim paper and can be reproduced using the [`preprocessing/create_npy_data_splits.ipynb`](https://github.com/leap-stc/ClimSim/tree/main/preprocessing/create_npy_data_splits.ipynb) notebook.


2 changes: 1 addition & 1 deletion website/evaluating.md
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@ Four different evaluation metrics were calculated:
3. Root mean squared error (RMSE)
4. Continuous ranked probability score (CRPS)

Evaluation and comparison of the different baseline models are found in the ```[evaluation/](https://github.com/leap-stc/ClimSim/tree/main/evaluation)``` folder on GitHub. All variables are converted to a common energy unit (i.e., W/m²) for scoring. The scoring is done using the functions in ```[climsim_utils/data_utils.py](https://github.com/leap-stc/ClimSim/tree/main/climsim_utils)```.
Evaluation and comparison of the different baseline models are found in the [```evaluation/```](https://github.com/leap-stc/ClimSim/tree/main/evaluation) folder on GitHub. All variables are converted to a common energy unit (i.e., W/m²) for scoring. The scoring is done using the functions in [`climsim_utils/data_utils.py`](https://github.com/leap-stc/ClimSim/tree/main/climsim_utils).

[This notebook](./evaluation/main_figure_generation.ipynb) calculates and plots MAE, R², RMSE, and CRPS scores for each baseline model. The separate R² for *longitudinally-averaged* and time-averaged 3D variables is found in [this notebook](./evaluation/plot_R2_analysis.ipynb).

Expand Down
2 changes: 1 addition & 1 deletion website/installation.md
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
# Installation & setup
# Code Installation

For preprocessing and evaluation, please install the `climsim_utils` python tools, by running the following code from the root of this repo:

Expand Down
8 changes: 4 additions & 4 deletions website/quickstart.md
Original file line number Diff line number Diff line change
Expand Up @@ -2,13 +2,13 @@

**Step 1**

The first step is to download the subsampled low-resolution real-geography version of the data [here](https://huggingface.co/datasets/LEAP/subsampled_low_res/tree/main). This contains subsampled and prenormalized data that was used for training, validation, and metrics for the ClimSim paper. It can be reproduced with the full version of the [dataset](https://huggingface.co/datasets/LEAP/ClimSim_low-res) using the [preprocessing/create_npy_data_splits.ipynb](https://github.com/leap-stc/ClimSim/blob/main/preprocessing/create_npy_data_splits.ipynb) notebook.
The first step is to download the subsampled low-resolution real-geography version of the data [here](https://huggingface.co/datasets/LEAP/subsampled_low_res/tree/main). This contains subsampled and prenormalized data that was used for training, validation, and metrics for the ClimSim paper. It can be reproduced with the full version of the [dataset](https://huggingface.co/datasets/LEAP/ClimSim_low-res) using the [`preprocessing/create_npy_data_splits.ipynb`](https://github.com/leap-stc/ClimSim/blob/main/preprocessing/create_npy_data_splits.ipynb) notebook.

Training data corresponds to **train_input.npy** and **train_target.npy**. Validation data corresponds to **val_input.npy** and **val_target.npy**. Scoring data (which can be treated as a test set) corresponds to **scoring_input.npy** and **scoring_target.npy**. We have an additional held-out test set that we will use for an upcoming online competition. Keep an eye out! 😉

**Step 2**

Install the `climsim_utils` python tools, by running the following code from the root of this repo:
Install the `climsim_utils` python tools, by running the following code from the root of the [GitHub repo](https://github.com/leap-stc/ClimSim/tree/main):

```
pip install .
Expand All @@ -22,9 +22,9 @@ pip install . --no-deps

**Step 3**

Train your model on the training data and validate using the validation data. If you wish to use something like a CNN, you will probably want to separate the variables into channels and broadcast scalars into vectors of the same dimension as vertically-resolved variables. Methods to do this can be found in the [climsim_utils/data_utils.py](https://github.com/leap-stc/ClimSim/blob/main/climsim_utils/data_utils.py) script.
Train your model on the training data and validate using the validation data. If you wish to use something like a CNN, you will probably want to separate the variables into channels and broadcast scalars into vectors of the same dimension as vertically-resolved variables. Methods to do this can be found in the [`climsim_utils/data_utils.py`](https://github.com/leap-stc/ClimSim/blob/main/climsim_utils/data_utils.py) script.

**Step 4**

Evaluation time! Use the [evaluation/main_figure_generation.ipynb](https://github.com/leap-stc/ClimSim/blob/main/evaluation/main_figure_generation.ipynb) notebook to see how your model does! Use the **calc_MAE**, **calc_RMSE**, and **calc_R2** methods in the [climsim_utils/data_utils.py](https://github.com/leap-stc/ClimSim/blob/main/climsim_utils/data_utils.py) script to see how your model does on point estimates and use the calc_CRPS method to check how well-calibrated your model is if it's stochastic. 😊
Evaluation time! Use the [`evaluation/main_figure_generation.ipynb`](https://github.com/leap-stc/ClimSim/blob/main/evaluation/main_figure_generation.ipynb) notebook to see how your model does! Use the **calc_MAE**, **calc_RMSE**, and **calc_R2** methods in the [`climsim_utils/data_utils.py`](https://github.com/leap-stc/ClimSim/blob/main/climsim_utils/data_utils.py) script to see how your model does on point estimates and use the `calc_CRPS` method to check how well-calibrated your model is if it's stochastic. 😊

0 comments on commit 5cb61b9

Please sign in to comment.