Overview¶
PyHazards brings together public dataset catalogs, registry-based models, benchmark families, experiment configs, and shared training or reporting workflows across wildfire, earthquake, flood, and tropical cyclone tasks.
It is designed for researchers and practitioners who need one coherent library for reproducing baselines, comparing methods, and extending hazard-ML workflows without rebuilding the software stack for each hazard family.
4
Wildfire, earthquake, flood, and tropical cyclone workflows under one library.
20
Curated dataset pages covering forcing sources and hazard-specific benchmark adapters.
24
Public implemented baselines and variants surfaced through the model catalog.
4
Shared evaluator families with linked ecosystems, smoke configs, and reports.
Start Here¶
Use one of these four paths to move from overview to action quickly.
Explore forcing sources, benchmark adapters, and inspection entrypoints.
Open: Datasets
Compare implemented baselines, variants, and benchmark-linked model detail pages.
Open: Models
Compare hazard benchmark families, ecosystem mappings, and smoke coverage.
Open: Benchmarks
Why PyHazards¶
Public datasets, forcing sources, and inspection surfaces are documented through one hazard-first catalog.
Shared benchmark families, smoke configs, and report exports make model comparisons more reproducible.
Baselines and adapters are exposed through a consistent build surface instead of one-off scripts.
One engine layer supports training, evaluation, prediction, and benchmark execution across hazard tasks.
Hazard Coverage¶
PyHazards spans four hazard families with public datasets, models, and benchmark pages designed to work together.
Waveform picking, dense-grid forecasting adapters, and linked benchmark ecosystems for phase-picking workflows.
Explore: Models | Benchmarks
Streamflow and inundation baselines with benchmark-backed datasets, configs, and evaluation coverage.
Explore: Datasets | Benchmarks
Track-and-intensity forecasting baselines plus shared benchmark ecosystems and experimental weather-model adapters.
Explore: Models | Benchmarks
Featured Example¶
Run a benchmark-aligned smoke configuration with one command, then move into the full Quick Start for model building and training workflows.
python scripts/run_benchmark.py --config pyhazards/configs/flood/hydrographnet_smoke.yaml
Next step: Quick Start for the first full workflow, or Models to browse benchmark-linked baselines.
Explore the Docs¶
Browse hazard-grouped dataset cards, detail pages, and inspection entrypoints.
Open: Datasets
Review benchmark families, ecosystem mappings, and smoke-config coverage.
Open: Benchmarks
For Contributors¶
PyHazards is registry-driven and uses dataset cards, model cards, and benchmark cards to generate the public catalogs. If you plan to extend the library, use Implementation Guide for the contributor workflow and Coverage Audit for the audited gap list behind the current roadmap work.
Citation¶
If you use PyHazards in your research, please cite:
@misc{pyhazards2025,
title = {PyHazards: An Open-Source Library for AI-Powered Hazard Prediction},
author = {Cheng et al.},
year = {2025},
howpublished = {\url{https://github.com/LabRAI/PyHazards}},
note = {GitHub repository}
}
Community¶
Use the RAI Lab Slack channel for project discussion and coordination.