diff --git a/README.html b/README.html
index 941276f..26e7826 100644
--- a/README.html
+++ b/README.html
@@ -169,6 +169,7 @@
Quickstart
Dataset Information
Task Overview
+
Training
Evaluation
Baseline Models
@@ -372,6 +373,7 @@ Contents
Getting Started
Build Your Own Model
Benchmarking
+Motivation
@@ -388,10 +390,10 @@ ChaosBench: A Multi-Channel, Physics-Based Benchmark for Subseasonal-to-Seas
ChaosBench is a benchmark project to improve long-term forecasting of chaotic systems, in particular subseasonal-to-seasonal (S2S) climate, using ML approaches.
Homepage 🔗: https://leap-stc.github.io/ChaosBench
Paper 📚: https://arxiv.org/
-Dataset 🤗: https://huggingface.co/datasets/juannat7/ChaosBench
+Dataset 🤗: https://huggingface.co/datasets/LEAP/ChaosBench
Features
-
+
1️⃣ Extended Observations. Spanning over 45 years (1979 - 2023) of ERA5 reanalysis
2️⃣ Diverse Baselines. Wide selection of physics-based forecasts from leading national agencies in Europe, the UK, America, and Asia
3️⃣ Differentiable Physics Metrics. Introduces two differentiable physics-based metrics to minimize the decay of power spectra at long forecasting horizon (blurriness)
@@ -419,6 +421,13 @@ Benchmarking
+
+Motivation
+1️⃣ Collapse to Climatology. Performing comparable or worse than climatology renders these state-of-the-art-models operationally unusable
+
+2️⃣ Blurring Artifact. Averaged-out forecasts is of little use when one attempts to predict extreme events requiring high-fidelity on the S2S scale (e.g., droughts, hurricanes)
+
+
@@ -482,6 +491,7 @@ Build Your Own Model
Benchmarking
+Motivation
diff --git a/_images/all_rmse_sota.png b/_images/all_rmse_sota.png
new file mode 100644
index 0000000..af53a6f
Binary files /dev/null and b/_images/all_rmse_sota.png differ
diff --git a/_images/chaosbench_scheme.jpg b/_images/chaosbench_scheme.jpg
new file mode 100644
index 0000000..3450bde
Binary files /dev/null and b/_images/chaosbench_scheme.jpg differ
diff --git a/_images/preds_climax_q700_direct_Task1.png b/_images/preds_climax_q700_direct_Task1.png
new file mode 100644
index 0000000..07963a2
Binary files /dev/null and b/_images/preds_climax_q700_direct_Task1.png differ
diff --git a/_sources/README.md b/_sources/README.md
index 4be9024..ee346d4 100644
--- a/_sources/README.md
+++ b/_sources/README.md
@@ -7,8 +7,7 @@ Homepage 🔗: https://leap-stc.github.io/ChaosBench
Paper 📚: https://arxiv.org/
-Dataset 🤗: https://huggingface.co/datasets/juannat7/ChaosBench
-
+Dataset 🤗: https://huggingface.co/datasets/LEAP/ChaosBench
## Features
@@ -22,7 +21,6 @@ Dataset 🤗: https://huggingface.co/datasets/juannat7/ChaosBench
4️⃣ __Large-Scale Benchmarking__. Systematic evaluation for state-of-the-art ML-based weather models like PanguWeather, FourcastNetV2, ViT/ClimaX, and Graphcast
-
## Getting Started
- [Quickstart](https://leap-stc.github.io/ChaosBench/quickstart.html)
- [Dataset Overview](https://leap-stc.github.io/ChaosBench/dataset.html)
@@ -35,4 +33,12 @@ Dataset 🤗: https://huggingface.co/datasets/juannat7/ChaosBench
## Benchmarking
- [Baseline Models](https://leap-stc.github.io/ChaosBench/baseline.html)
-- [Leaderboard](https://leap-stc.github.io/ChaosBench/leaderboard.html)
\ No newline at end of file
+- [Leaderboard](https://leap-stc.github.io/ChaosBench/leaderboard.html)
+
+
+## Motivation
+1️⃣ __Collapse to Climatology__. Performing comparable or worse than climatology renders these state-of-the-art-models operationally unusable
+
+
+2️⃣ __Blurring Artifact__. Averaged-out forecasts is of little use when one attempts to predict extreme events requiring high-fidelity on the S2S scale (e.g., droughts, hurricanes)
+
\ No newline at end of file
diff --git a/_sources/baseline.md b/_sources/baseline.md
index 9077477..603f824 100644
--- a/_sources/baseline.md
+++ b/_sources/baseline.md
@@ -23,7 +23,7 @@ We differentiate between physics-based and data-driven models. The former is suc
- [x] GraphCast
## Model Checkpoints
-Checkpoints for data-driven models are accessible from [here](https://huggingface.co/datasets/juannat7/ChaosBench/tree/main/logs)
+Checkpoints for data-driven models are accessible from [here](https://huggingface.co/datasets/LEAP/ChaosBench/tree/main/logs)
- Data-driven models are indicated by the `_s2s` suffix (e.g., `unet_s2s`).
diff --git a/_sources/evaluation.md b/_sources/evaluation.md
index 387e1f0..0ab0267 100644
--- a/_sources/evaluation.md
+++ b/_sources/evaluation.md
@@ -22,7 +22,7 @@ __For example__, in our `unet_s2s` baseline model, we can run:
## Accessing Baseline Scores
-You can access the complete scores (in `.csv` format) for data-driven, physics-based models, climatology, and persistence [here](https://huggingface.co/datasets/juannat7/ChaosBench/tree/main/logs). Below is a snippet from `logs/climatology/eval/rmse_climatology.csv`, where each row represents ``, such as `RMSE`, at each future timestep.
+You can access the complete scores (in `.csv` format) for data-driven, physics-based models, climatology, and persistence [here](https://huggingface.co/datasets/LEAP/ChaosBench/tree/main/logs). Below is a snippet from `logs/climatology/eval/rmse_climatology.csv`, where each row represents ``, such as `RMSE`, at each future timestep.
| z-10 | z-50 | z-100 | z-200 | z-300 | ... | w-1000 |
|----------|----------|----------|----------|----------|-----|----------|
diff --git a/_sources/quickstart.md b/_sources/quickstart.md
index 3b7c96a..3779cf5 100644
--- a/_sources/quickstart.md
+++ b/_sources/quickstart.md
@@ -8,11 +8,11 @@ cd ChaosBench
mkdir data
```
-**Step 3**: Navigate to `chaosbench/config.py` and change the field `DATA_DIR = //ChaosBench/data` (_Provide absolute path_)
+**Step 3**: Navigate to `chaosbench/config.py` and change the field `DATA_DIR = ChaosBench/data`
**Step 4**: Initialize the space by running
```
-cd //ChaosBench/data/
+cd ChaosBench/data/
wget https://huggingface.co/datasets/juannat7/ChaosBench/blob/main/process.sh
chmod +x process.sh
```
diff --git a/_sources/task.md b/_sources/task.md
index 946338a..afb461f 100644
--- a/_sources/task.md
+++ b/_sources/task.md
@@ -14,6 +14,7 @@ __NOTE__: Before training your own model [instructions here](https://leap-stc.gi
- Task 2️⃣: `only_headline: True`. By default, it is going to optimize on {t-850, z-500, q-700}. To change this, modify the `HEADLINE_VARS` field in `chaosbench/config.py`
+# Training Strategies
In addition, we also provide flags to train the model either __autoregressively__ or __directly__.
- Autoregressive: Using current output as the next model input. The number of iterative steps is defined in the `n_step: ` field. For our baselines, we set `N_STEP = 5`.
diff --git a/_sources/training.md b/_sources/training.md
index 2d062fe..578dfdd 100644
--- a/_sources/training.md
+++ b/_sources/training.md
@@ -4,11 +4,14 @@
We will outline how one can implement their own data-driven models. Several examples, including ED, FNO, ResNet, and UNet have been provided.
-**Step 1**: Define your model class in `chaosbench/models/.py`. At present, we only support models built with `PyTorch`
+### Step 1
+Define your model class in `chaosbench/models/.py`
-**Step 2**: Initialize your model in `chaosbench/models/model.py` under `__init__` method in `S2SBenchmarkModel` class
+### Step 2
+Initialize your model in `chaosbench/models/model.py` under `S2SBenchmarkModel.__init__`
-**Step 3**: Write a configuration file in `chaosbench/configs/_s2s.yaml`. We recommend reading the details on the definition of [hyperparameters](https://leap-stc.github.io/ChaosBench/baseline.html) and the different [task]((https://leap-stc.github.io/ChaosBench/task.html) before training. Also change the `model_name: _s2s` to ensure correct pathing
+### Step 3
+Write a configuration file in `chaosbench/configs/_s2s.yaml`. Details on the definition of [hyperparameters](https://leap-stc.github.io/ChaosBench/baseline.html) and the different [task](https://leap-stc.github.io/ChaosBench/task.html). Also change the `model_name: _s2s` to ensure correct pathing
- Task 1️⃣ (autoregressive): `only_headline: False ; n_step: `
- Task 1️⃣ (direct): `only_headline: False ; n_step: 1 ; lead_time: `
@@ -17,8 +20,8 @@ We will outline how one can implement their own data-driven models. Several exam
- Task 2️⃣ (direct): `only_headline: True ; n_step: 1 ; lead_time: `
-**Step 4**: Train by running `python train.py --config_filepath chaosbench/configs/_s2s.yaml`
+### Step 4
+Train by running `python train.py --config_filepath chaosbench/configs/_s2s.yaml`
-**Step 5**: Done!
-__NOTE__: Remember to replace `` with your own model name, e.g., `unet`. Checkpoints and logs would be automatically generated in `logs/_s2s/`.
\ No newline at end of file
+Remember to replace `` with your own model name, e.g., `unet`. Checkpoints and logs would be automatically generated in `logs/_s2s/`.
\ No newline at end of file
diff --git a/baseline.html b/baseline.html
index 966db77..d98852f 100644
--- a/baseline.html
+++ b/baseline.html
@@ -170,6 +170,7 @@
Quickstart
Dataset Information
Task Overview
+
Training
Evaluation
Baseline Models
@@ -415,7 +416,7 @@ Model Definition