Hi! This is work in progress - please check back later or use our Discussions tab if you're interested in this project and would like to say hi. The models here should be considered beta, and could improve over time, or they may change in structure, so any usage of these models should be for practice with Segmentation Gym. Better documentation is forthcoming.
- Segmentation Zoo is a repository of image segmentation models for broad use in the geosciences, pre-trained using Segmentation Gym. These models may be used for general scientific enquiries or for use in downstream applications (credit is always appreciated!)
- 'Broad use' is open to interpretation, but perhaps relates to 'broad classes' of widespread utility in geosciences (including geology, ecology, hydrology, oceanography and all related fields) such as water, sediment, vegetation, soil, rock, and landcover types. Such broad classes may also be broken down into other generally useful classes, such as water types (e.g. whitewater, turbid water), sediment types (sand, gravel, etc), vegetation types, soil types, and rock types. Examples of broad, established ontologies for geoscience disciplines are CMECS and GLanCE
- Segmentation Zoo promotes use of models published using Zenodo, along with a model card in a specific format that documents the model, its construction, and intendd uses.
- Finally, 'Zoo contains various examples of model implementations. 'Implementation' in this scope refers to the use of a model on unseen sample imagery. There are a number of ways that this may be acheived, and would differ depending on factors such as the type of imagery, amount of overlap in imagery, required accuracy, the need for probabilistic outputs, number of classes, etc. Whereas 'Segmentation Gym' contains a basic model implementation (called
seg_images_in_folder.py
), bespoke model applications may differ considerably. The notebooks and scripts folders in Zoo contain basic and more advanced examples of how to use a model for segmentation, including use of ensemble models. We also hope to demonstrate possible transfer learning scenarios
Package maintainers:
Contributions:
We welcome collaboration! Please use our Discussions tab if you're interested in this project. We welcome user-contributed models trained using Segmentation Gym, but please wait for announcements while we get our act together, and meanwhile we'll be happy to chat about your data, models, science, art, and life in general in Discussions.
If you already have the conda environment, gym
, installed from Segmentation Gym, you may use that. Otherwise, we advise creating a new conda environment to run the program.
Note that MACS are NOT SUPPORTED. Only Linux and WSL on Windows. Not sorry :)
- Clone the repo:
git clone --depth 1 https://github.com/Doodleverse/segmentation_zoo.git
(--depth 1
means "give me only the present code, not the whole history of git commits" - this saves disk space, and time)
- Create a conda environment called
zoo
[OPTIONAL] First you may want to do some conda and pip housekeeping (recommended)
conda update -n base conda
conda clean --all
python3 -m pip install --upgrade pip
[OPTIONAL] Set mamba to the default installer:
conda install -n base conda-libmamba-solver
conda config --set solver libmamba
(updated November 20, 2024)
conda env create --file ./install/zoo.yml
conda activate zoo
- User of models
- We anticipate that most Zoo users will simply want to use published models listed in the wiki, and adapt the example notebook workflows for their own purposes
- Contributor of models
- We welcome contributions of Gym models! The basic steps (to be outlined in the wiki) involve a) making a new Zenodo release including your model wrights file, config file, and modelcard file; b) cloning the Zoo wiki and adding a page explaining the model's purpose; and c) issuing a pull request for review by Doodleverse HQ
- Contributor of implementation workflows
- We welcome contributions of Gym model implementations! The basic steps (to be outlined in the wiki) involve a) cloning the Zoo repo and adding a notebook or script showing your workflow; and b) issuing a pull request for review by Doodleverse HQ
To run the example script, change directory (cd
) to the scripts
directory and run the python script using
python select_model_and_batch_process_folder.py
You select a task, then a model, then select a folder of images to apply it to. It will write output files into a subfolder of the input imagery called 'out'
The set of available tasks.models are listed below, by theme and more details may be found on the Models pages of the wiki for satellite, aerial, and orthomosaic imagery
- 4-class models
- "sat_RGB_4class_segformer_global_14183210", 4 class (water, whitewater, sediment, other), RGB satellite imagery, Global edition, segformer
- "sat_RGB_4class_segformer_AK_14037041", 4 class (water, whitewater, sediment, other), RGB satellite imagery, Alaska-only edition, segformer
- "sat_NDWI_4class_segformer_global_14172182", 4 class (water, whitewater, sediment, other), NDWI satellite imagery, Global edition, segformer
- "sat_NDWI_4class_segformer_AK_14183210", 4 class (water, whitewater, sediment, other), NDWI satellite imagery, Alaska-only edition, segformer
- "sat_MNDWI_4class_segformer_global_14183366", 4 class (water, whitewater, sediment, other), MNDWI satellite imagery, Global edition, segformer
- "sat_MNDWI_4class_segformer_AK_14187478", 4 class (water, whitewater, sediment, other), MNDWI satellite imagery, Alaska-only edition, segformer
- "pcmsc_wm_v2_7700430", 2 class (water, other), RGB aerial oblique imagery, 1024x768 pixel imagery, westcoast high-relief coasts
- "aerial_2class_6234122", 2 class (water, other), RGB aerial oblique imagery, 1024x768 pixel imagery
- "aerial_2class_6235090", 2 class (water, other), RGB aerial oblique imagery, 1024x768 pixel imagery
- "noaa_2class_7604083", 2 class (water, other), RGB aerial oblique imagery, 1024x768 pixel NOAA ERI imagery
- NOAA models
- "noaa_4class_7631354", 4 class (water, sediment, vegetation, developed), RGB aerial oblique imagery, 1024x768 pixel imagery
- FloodNet models
- "floodnet_10class_7566810", 768x512 pixel imagery, 10 class (Background, Building-flooded, Building-non-flooded, Road-flooded, Road-non-flooded, Water, Tree, Vehicle, Pool, Grass), RGB UAV imagery
- "floodnet_10class_7566810", 1024x768 pixel imagery, 10 class (Background, Building-flooded, Building-non-flooded, Road-flooded, Road-non-flooded, Water, Tree, Vehicle, Pool, Grass), RGB UAV imagery
- Elwha River alluvial models
- "elwha_alluvial_driftwood_segformer_7933013", 2 class (other, wood), RGB aerial orthomosaic imagery, 768x768 pixel imagery, segformer
- "elwha_alluvial_driftwood_resunet_8072293", 2 class (other, wood), RGB aerial orthomosaic imagery, 768x768 pixel imagery, resunet
- Elwha delta sediment models
- "elwha_beachsediment_4class_segformer_14199980", 4-class (sand, gravel, cobble, other) RGB close-range sediment imagery, 768x768 pixel imagery, segformer
-
Coast Train models
- "ortho_2class_7574784", 2 class (water, other), 768x768 pixel RGB NAIP/orthomosaic imagery
- "ortho_5class_7566992", 5 class (water, whitewater, sediment, bare terrain, other terrain), RGB NAIP/orthomosaic imagery
- "orthoCT_8class_7570583", 8 class (water, whitewater, sediment, bare terrain, marsh veg, terrestrial veg, ag., dev.), 768x768 pixel RGB NAIP/orthomosaic imagery
-
Chesapeake models
- "chesapeake_7class_7576904", 7 class (water, tree canopy / forest, low vegetation / field, barren land, impervious (other), impervious (road), no data), 512x512 pixel RGB NAIP/orthomosaic imagery
-
Barrier islands models
- Barrier islands: substrates (forthcoming)
- Barrier islands: geomorphic category (forthcoming)
- Barrier islands: vegetation type (forthcoming)
- Barrier islands: vegetation density (forthcoming)
- Barrier islands: general classification (forthcoming)
- OpenEarthMap models
- "openearthmap_9class_7576894", 512x512 pixel imagery, 9 class (bareland, rangeland, dev., road, tree, water, ag., building, nodata)
- DeepGlobe models
- "deepglobe_7class_7576898", 512x512 pixel imagery, 7 class (urban, ag., rangeland, forest, water, bare, unknown)
- EnviroAtlas models
- "enviroatlas_6class_7576909" 1024x1024 pixel imagery, 6 class (water, impervious, barren, trees, herbaceous, shrubland)
- AAAI/building models
- "aaai_building_7607895" 1024x1024 pixel imagery, 2 class (building, other)
- "aaai_floodbuilding_7622733" 1024x1024 pixel imagery, 2 class (flooded building, other)
- XBD/building models
- "xbd_building_7613212" 768x768 pixel imagery, 2 class (building, other)
- "xbd_damagedbuilding_7613175" 768x768 pixel imagery, 4 class (building/no-damage, building/minor-damage, building/major-damage, other)
To run the example notebooks, change directory (cd
) to the notebooks
directory and launch jupyter using
jupyter notebook
The menu of notebooks can be accessed in the browser at http://localhost:8888/tree
You should adapt the workflows shown in these notebooks to your own imagery
Check out our wiki for further instructions