Skip to content

DZGong/uSplit

 
 

Repository files navigation

μSplit

This is the official implementation of μSplit: image decomposition for fluorescence microscopy.

Installation

In case you don't have mamba, install it from here.

git clone https://github.com/juglab/uSplit.git
cd uSplit
mamba create -n usplit python=3.9
mamba activate usplit
./install_deps.sh
pip install -e .

In case, one wants to do training, then one also needs to create an account on wandb. This is used for logging training and evaluation metrics. In case you do not want to use wandb, you can replace the logger here with a logger of your choice and comment out here.

Data

For Hagen et al. dataset, please download the data from here. You need to download the following files:

1. actin-60x-noise2-highsnr.tif
2. mito-60x-noise2-highsnr.tif

For our PaviaATN dataset, please download the data from here.

For each dataset, create a separate directory and put the files in the directory. The directory should not contain any other files.

Network training

To train one of our LC variants on the Hagen dataset, run this command:

python /home/ubuntu/code/uSplit/uSplit/scripts/run.py --workdir=/home/ubuntu/training/uSplit/ -mode=train --datadir=/home/ubuntu/data/ventura_gigascience/ --config=/home/ubuntu/code/uSplit/usplit/configs/lc_hagen_config.py

For training our Lean-LC model, do the following hyperparameter assignment in lc_hagen_config.py

  1. Set model.decoder.multiscale_retain_spatial_dims=False.
  2. Set model.z_dims = [128, 128, 128, 128].

For training our Regular-LC model, do the following hyperparameter assignment in lc_hagen_config.py

  1. Set model.decoder.multiscale_retain_spatial_dims=True.
  2. Set model.z_dims = [128, 128, 128, 128].

For training our Deep-LC model, do the following hyperparameter assignment in lc_hagen_config.py

  1. Set model.decoder.multiscale_retain_spatial_dims=True.
  2. Set model.z_dims = [128, 128, 128, 128, 128, 128, 128, 128].

To train the Vanilla baseline, set data.multiscale_lowres_count = None. To train HAE version, set model.non_stochastic_version = True. For HVAE version, set model.non_stochastic_version = True.

We have also provided the configs for training on our PaviaATN dataset and SinosoidalCritters dataset. Above mentioned hyperparameter settings needs to be changed accordingly.

Evaluation

For evaluation, we have provided the pre-trained models here. We have provided one notebook here where one can inspect both qualitatively and quantitatively our models' performances. We have provided another notebook here where one can evaluate our models on input data where there is no ground truth. Note that in this case, model will perform well only if the input data is similar to input generated from the training data.

Training on your own data.

Codebase will require minimal changes if the data is organized in one of the three ways:

  1. A single tiff file with data shape (N,H,W,C) where C denotes the channel dimension which one intends to split. In this case, simply use DataType.OptiMEM100_014 as the data.data_type in the config.
  2. Two tiff files, each with data shape (N, H, W) and corresponding to one channel. Use DataType.SeparateTiffData as the data.data_type in the config.
  3. A .zarr file directory with data shape (N,H,W,C). Use DataType.SingleZarrData as the data.data_type in the config. In this case, we need to apriori split 'raw' .zarr into train/validation/test. To do this, we need to run the following command:
    python usplit/scripts/multiscale_zarr_data_generator.py /home/ubuntu/data/raw.zarr/ /home/ubuntu/data/microscopy_zarr ZHWC 5 --input_zarr_group='raw' --overwrite

No other changes should be required and at this point, one can start the training. Note that input will be created by summing the two channels present in the data.

Steps for data organized in a different way.

  1. Add an entry in DataType class for your own data. For example,
MyDataABC = 17
  1. In train_val_data.py, create a function which returns returns the train/val/test data with (N,H,W,C) dimension arrangement.
  2. In training.py:create_dataset() you may have to set dataclass variable appropriately in case your data is .zarr.

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 74.7%
  • Jupyter Notebook 25.1%
  • Shell 0.2%