Datasets:
image imagewidth (px) 1.14k 2.4k | label class label 2 classes |
|---|---|
0validation_gik_vs_herbie_2024 | |
0validation_gik_vs_herbie_2024 | |
0validation_gik_vs_herbie_2024 | |
0validation_gik_vs_herbie_2024 | |
0validation_gik_vs_herbie_2024 | |
0validation_gik_vs_herbie_2024 | |
0validation_gik_vs_herbie_2024 | |
0validation_gik_vs_herbie_2024 | |
0validation_gik_vs_herbie_2024 | |
0validation_gik_vs_herbie_2024 | |
0validation_gik_vs_herbie_2024 | |
1validation_gik_vs_herbie_2025 | |
1validation_gik_vs_herbie_2025 | |
1validation_gik_vs_herbie_2025 | |
1validation_gik_vs_herbie_2025 | |
1validation_gik_vs_herbie_2025 | |
1validation_gik_vs_herbie_2025 | |
1validation_gik_vs_herbie_2025 | |
1validation_gik_vs_herbie_2025 | |
1validation_gik_vs_herbie_2025 | |
1validation_gik_vs_herbie_2025 | |
1validation_gik_vs_herbie_2025 | |
1validation_gik_vs_herbie_2025 |
GIK-ECMWF-PAR: Virtual Reference Parquets for ECMWF IFS Ensemble Forecasts
Lightweight parquet reference files that turn 3--4 GB ECMWF GRIB files into ~140 KB virtual datasets, enabling Dask-based parallel analysis without downloading the raw data.
What These Parquets Do
Each parquet file contains a table of [zarr_key, [s3_url, byte_offset, byte_length]] references pointing into ECMWF IFS ensemble GRIB files on AWS S3 (s3://ecmwf-forecasts/). Instead of downloading full GRIB files, analysis code performs targeted byte-range reads fetching only the specific variables, members, and timesteps needed -- typically 2--5% of the original data.
Without GIK: Download 340 GB of GRIB files per day --> then process
With GIK: Read 7 MB of parquets + stream ~5 GB --> direct analysis
Scale: Parquets vs Source GRIB Data
Source GRIB Scale (ECMWF IFS 00z only)
| Parameter | Value |
|---|---|
| Timesteps per run | 85 (0--144h at 3h, 150--360h at 6h) |
| Members per timestep | 51 (1 control + 50 perturbed) |
| File size per timestep | ~3--4 GB (all 51 members in one GRIB) |
| GRIB data per day (00z) | ~85 x 4 GB = ~340 GB |
Annual GRIB Data Referenced
| Year | Coverage | Days | Daily GRIB (00z) | Total GRIB Referenced |
|---|---|---|---|---|
| 2024 | March 1 -- December 31 | ~306 | ~340 GB | ~101 TB |
| 2025 | January 1 -- December 31 | 365 | ~340 GB | ~124 TB |
| 2026 | January 1 -- February 20 | 51 | ~340 GB | ~17 TB |
| Total | ~722 days | ~242 TB |
Parquet Reference Files in This Dataset
| Year | Parquets | Size | Compression vs GRIB |
|---|---|---|---|
| 2024 | ~15,606 | ~2.2 GB | 2.2 GB references ~101 TB (46,000x) |
| 2025 | ~18,615 | ~2.6 GB | 2.6 GB references ~124 TB (48,000x) |
| 2026 | ~1,581 | ~0.2 GB | 0.2 GB references ~17 TB (85,000x) |
| Total | ~35,802 | ~5.0 GB | 5 GB references ~242 TB |
Each parquet is ~140 KB and represents one ensemble member for one forecast date.
Dataset Structure
run_par_ecmwf/
{YYYY}/{MM}/{YYYYMMDD}/00z/
{YYYYMMDD}00z-control.parquet # Control member
{YYYYMMDD}00z-ens01.parquet # Ensemble member 1
{YYYYMMDD}00z-ens02.parquet # Ensemble member 2
...
{YYYYMMDD}00z-ens50.parquet # Ensemble member 50
validation_gik_vs_herbie_2024/ # Validation: 10 dates, r > 0.9999
validation_gik_vs_herbie_2025/ # Validation: 11 dates, r > 0.9999
Catalog / Index
A lightweight catalog.parquet (1.7 MB) at the repo root indexes all 144,228 parquet files across the dataset. Use it to discover available dates, runs, and members without listing the full repo tree.
Download: catalog.parquet
| Column | Example | Description |
|---|---|---|
year |
2024 |
Forecast year |
month |
03 |
Forecast month |
date |
20240301 |
Forecast date (YYYYMMDD) |
run |
00z |
Run hour |
member |
control |
Ensemble member name |
filename |
2024030100z-control.parquet |
Parquet filename |
hf_path |
run_par_ecmwf/2024/03/20240301/00z/... |
Full path in this repo |
size_bytes |
107520 |
File size in bytes |
Quick Start with the Catalog
import pandas as pd
from huggingface_hub import hf_hub_download
# 1. Load the catalog (1.7 MB, indexes all 144k+ files)
catalog_path = hf_hub_download(
repo_id="E4DRR/gik-ecmwf-par", repo_type="dataset",
filename="catalog.parquet"
)
catalog = pd.read_parquet(catalog_path)
# 2. Explore what's available
print(catalog.groupby(["year", "month"]).size()) # files per month
print(catalog["run"].unique()) # ['00z','06z','12z','18z']
print(catalog["member"].nunique()) # 51
# 3. Filter for a specific date + run
subset = catalog[(catalog["date"] == "20250101") & (catalog["run"] == "00z")]
print(subset[["member", "filename", "size_bytes"]])
# 4. Download a specific parquet using its hf_path
row = subset.iloc[0]
parquet_path = hf_hub_download(
repo_id="E4DRR/gik-ecmwf-par", repo_type="dataset",
filename=row["hf_path"]
)
Open Parquets as xarray Dataset
Each parquet contains zarr-style [key, value] references into remote GRIB files on S3.
Below is a minimal example that loads all 51 ensemble members for a single date and
streams one variable (total precipitation) into an xarray Dataset.
import json
import numpy as np
import pandas as pd
import xarray as xr
import fsspec
from concurrent.futures import ThreadPoolExecutor
from huggingface_hub import hf_hub_download
# ββ 1. Pick a date from the catalog ββββββββββββββββββββββββββββββ
catalog = pd.read_parquet(
hf_hub_download("E4DRR/gik-ecmwf-par", "catalog.parquet", repo_type="dataset")
)
date = "20250101"
subset = catalog[(catalog["date"] == date) & (catalog["run"] == "00z")]
# ββ 2. Helper: parquet β zstore dict βββββββββββββββββββββββββββββ
def parquet_to_zstore(parquet_path):
"""Read a GIK parquet and return a {zarr_key: value} dict."""
df = pd.read_parquet(parquet_path)
zstore = {}
for _, row in df.iterrows():
val = row["value"]
if isinstance(val, bytes):
val = val.decode("utf-8")
if isinstance(val, str) and val[0] in ("[", "{"):
val = json.loads(val)
zstore[row["key"]] = val
return zstore
# ββ 3. Helper: fetch one GRIB chunk from S3 ββββββββββββββββββββββ
s3 = fsspec.filesystem("s3", anon=True)
def fetch_step(ref):
"""Fetch raw GRIB bytes from S3 using [url, offset, length]."""
url, offset, length = ref[0], ref[1], ref[2]
if not url.endswith(".grib2"):
url += ".grib2"
with s3.open(url, "rb") as f:
f.seek(offset)
return f.read(length)
# ββ 4. Load all 51 members for one variable (tp) βββββββββββββββββ
STEPS = list(range(0, 145, 3)) + [150, 156, 162, 168] # 53 lead times
GRID = (721, 1440) # ECMWF 0.25 deg global
try:
import gribberish
decode = lambda b: gribberish.parse_grib_array(b, 0).reshape(GRID)
except ImportError:
# Fallback: cfgrib (slower)
import tempfile, os
def decode(b):
tmp = tempfile.NamedTemporaryFile(delete=False, suffix=".grib2")
tmp.write(b); tmp.close()
ds = xr.open_dataset(tmp.name, engine="cfgrib")
arr = ds[list(ds.data_vars)[0]].values.copy(); ds.close()
os.unlink(tmp.name)
return arr
members_data = {}
for _, row in subset.iterrows():
pq_path = hf_hub_download(
"E4DRR/gik-ecmwf-par", row["hf_path"], repo_type="dataset"
)
zstore = parquet_to_zstore(pq_path)
# Find tp references for each lead time
data = np.full((len(STEPS), *GRID), np.nan, dtype=np.float32)
member_key = row["member"].replace("_", "")
step_refs = []
for i, h in enumerate(STEPS):
for pattern in [
f"step_{h:03d}/tp/sfc/{member_key}/0.0.0",
f"step_{h:03d}/tp/surface/{member_key}/0.0.0",
]:
if pattern in zstore and isinstance(zstore[pattern], list):
step_refs.append((i, zstore[pattern]))
break
# Parallel S3 fetches (8 threads, I/O bound)
with ThreadPoolExecutor(8) as pool:
futs = {pool.submit(fetch_step, ref): i for i, ref in step_refs}
for fut in futs:
try:
data[futs[fut]] = decode(fut.result())
except Exception:
pass
members_data[row["member"]] = data
# ββ 5. Assemble xarray Dataset βββββββββββββββββββββββββββββββββββ
ds = xr.Dataset(
{"tp": (["member", "step", "latitude", "longitude"],
np.stack([members_data[m] for m in sorted(members_data)]))},
coords={
"member": sorted(members_data.keys()),
"step": STEPS,
"latitude": np.linspace(90, -90, 721),
"longitude": np.linspace(-180, 179.75, 1440),
},
)
print(ds)
# <xarray.Dataset> Dimensions: member=51, step=53, latitude=721, longitude=1440
Open Parquets as Lazy xarray Dataset (Virtual / On-Demand)
Same parquets, but opened lazily with dask β parquets are parsed instantly,
S3 byte-range reads happen only when you call .load(). This lets you build
a full 51-member dataset in seconds and then selectively materialize slices.
import json
import dask
import dask.array as da
import fsspec
import gribberish
import numpy as np
import pandas as pd
import xarray as xr
from huggingface_hub import hf_hub_download
# ββ 1. Setup βββββββββββββββββββββββββββββββββββββββββββββββββββββ
catalog = pd.read_parquet(
hf_hub_download("E4DRR/gik-ecmwf-par", "catalog.parquet", repo_type="dataset")
)
date, var = "20250101", "tp"
subset = catalog[(catalog["date"] == date) & (catalog["run"] == "00z")]
STEPS = list(range(0, 145, 3)) + [150, 156, 162, 168] # 53 lead times
GRID = (721, 1440)
s3 = fsspec.filesystem("s3", anon=True)
def parquet_to_zstore(path):
zstore = {}
for _, row in pd.read_parquet(path).iterrows():
val = row["value"]
if isinstance(val, bytes): val = val.decode("utf-8")
if isinstance(val, str) and len(val) > 0 and val[0] in ("[", "{"):
val = json.loads(val)
zstore[row["key"]] = val
return zstore
# ββ 2. Delayed fetcher: executes only when .load() is called βββββ
def make_lazy_chunk(ref):
@dask.delayed
def _fetch():
url, offset, length = ref[0], ref[1], ref[2]
if not url.endswith(".grib2"): url += ".grib2"
with s3.open(url, "rb") as f:
f.seek(offset)
return gribberish.parse_grib_array(f.read(length), 0) \
.reshape(GRID).astype(np.float32)
return da.from_delayed(_fetch(), shape=GRID, dtype=np.float32)
# ββ 3. Build lazy dataset (instant β no S3 reads) ββββββββββββββββ
member_arrays, member_names = [], []
for _, row in subset.iterrows():
pq = hf_hub_download("E4DRR/gik-ecmwf-par", row["hf_path"], repo_type="dataset")
zs = parquet_to_zstore(pq)
mk = row["member"].replace("_", "")
steps = []
for h in STEPS:
ref = None
for p in [f"step_{h:03d}/{var}/sfc/{mk}/0.0.0",
f"step_{h:03d}/{var}/surface/{mk}/0.0.0"]:
if p in zs and isinstance(zs[p], list):
ref = zs[p]; break
steps.append(make_lazy_chunk(ref) if ref else
da.full(GRID, np.nan, dtype=np.float32))
member_arrays.append(da.stack(steps, axis=0))
member_names.append(row["member"])
ds = xr.Dataset(
{var: (["member", "step", "latitude", "longitude"],
da.stack(member_arrays, axis=0))},
coords={"member": member_names, "step": STEPS,
"latitude": np.linspace(90, -90, 721),
"longitude": np.linspace(-180, 179.75, 1440)},
)
print(ds)
# <xarray.Dataset> Dimensions: member=51, step=53, latitude=721, longitude=1440
# tp: dask.array<...> β zero bytes in memory, all delayed
# ββ 4. Fetch only what you need ββββββββββββββββββββββββββββββββββ
# Each .load() triggers S3 byte-range reads (~2 MB per member per step)
step24 = ds["tp"].sel(step=24).load() # 1 step, all members
ea = ds["tp"].sel(step=24, latitude=slice(25, -14),
longitude=slice(19, 55)).load() # East Africa subset
Coverage Summary (from catalog)
| Year | Months | Dates | Files | Total Size |
|---|---|---|---|---|
| 2024 | Mar--Dec | 306 | ~62,424 | ~6.5 GB |
| 2025 | Jan--Dec | 365 | ~75,504 | ~7.8 GB |
| 2026 | Jan--Feb | 51 | ~7,344 | ~0.8 GB |
| Total | 720 | 144,228 | ~17.8 GB |
Catalog / Index
A lightweight catalog.parquet (~1.8 MB) at the repo root indexes all 150,246 parquet files across the dataset. Use it to discover available dates, runs, and members without listing the full repo tree.
Download: catalog.parquet
| Column | Example | Description |
|---|---|---|
year |
2024 |
Forecast year |
month |
03 |
Forecast month |
date |
20240301 |
Forecast date (YYYYMMDD) |
run |
00z |
Run hour |
member |
control |
Ensemble member name |
filename |
2024030100z-control.parquet |
Parquet filename |
hf_path |
run_par_ecmwf/2024/03/20240301/00z/... |
Full path in this repo |
size_bytes |
107520 |
File size in bytes |
Quick Start with the Catalog
import pandas as pd
from huggingface_hub import hf_hub_download
# 1. Load the catalog (~1.8 MB, indexes all 150k+ files)
catalog_path = hf_hub_download(
repo_id="E4DRR/gik-ecmwf-par", repo_type="dataset",
filename="catalog.parquet"
)
catalog = pd.read_parquet(catalog_path)
# 2. Explore what's available
print(catalog.groupby(["year", "month"]).size()) # files per month
print(catalog["run"].unique()) # ['00z','06z','12z','18z']
print(catalog["member"].nunique()) # 51
# 3. Filter for a specific date + run
subset = catalog[(catalog["date"] == "20250101") & (catalog["run"] == "00z")]
print(subset[["member", "filename", "size_bytes"]])
# 4. Download a specific parquet using its hf_path
row = subset.iloc[0]
parquet_path = hf_hub_download(
repo_id="E4DRR/gik-ecmwf-par", repo_type="dataset",
filename=row["hf_path"]
)
Coverage Summary (from catalog)
| Year | Months | Dates | Files | Total Size |
|---|---|---|---|---|
| 2024 | Mar--Dec | 306 | ~62,424 | ~6.5 GB |
| 2025 | Jan--Dec | 365 | ~75,504 | ~7.8 GB |
| 2026 | Jan--Mar 7 | 66 | ~12,318 | ~1.3 GB |
| Total | 737 | 150,246 | ~18.5 GB |
How It Works
The Grib-Index-Kerchunk (GIK) method applies the same principle as video streaming to weather data:
| Video Streaming | Weather Data Streaming (GIK) |
|---|---|
| Video split into segments | GRIB split into variable/member/timestep chunks |
| Manifest (.m3u8) lists segment URLs + byte ranges | Parquet lists GRIB URLs + byte ranges |
| Player fetches only visible segments | Analysis code fetches only needed variables |
| Full video never downloaded | Full GRIB (3--4 GB) never downloaded |
Usage with Dask
These parquets are designed for parallel analysis on Dask clusters. Each member's parquet is independent -- workers read their assigned parquet, fetch GRIB bytes via S3 byte-range reads, and decode using gribberish (Rust, ~25 ms/chunk).
| Approach | 12 vars x 51 members x 9 steps | Time |
|---|---|---|
| Download full GRIBs | ~340 GB transfer | Hours |
| GIK + single machine (8 threads) | ~4.5 GB byte-range reads | ~24 min |
| GIK + Coiled Dask (20 workers) | ~4.5 GB distributed reads | ~3 min |
Variables Available
All ECMWF IFS ensemble variables are referenced, including:
| Surface Variables | Pressure Level Variables |
|---|---|
| tp (Total Precipitation) | u/v wind at multiple levels |
| 2t (2m Temperature) | Temperature at multiple levels |
| sp (Surface Pressure) | Geopotential at multiple levels |
| ssr, ssrd (Solar Radiation) | Specific humidity |
| tcw, tcwv (Cloud Water) | |
| tcc (Total Cloud Cover) | |
| sf (Snowfall), ro (Runoff) |
Validation
GIK parquet-based data was validated against Herbie (independent ECMWF access library) across 21 dates spanning 2024--2025:
- Pearson r > 0.9999 for both ensemble mean and spread
- RMSE < 2e-04 m for total precipitation
- Tiny differences arise from decoder floating-point representation (gribberish vs cfgrib)
Project
Developed by ICPAC (IGAD Climate Prediction and Applications Centre) for continuous climate risk monitoring over East Africa.
- Funding: E4DRR (UN CRAF'd) and SEWAA projects
- Repository: icpac-igad/grib-index-kerchunk
- Method documentation: See
ecmwf/docs/gik_vs_herbie_comparison.mdin the repository
- Downloads last month
- 1,030