Skip to content
This repository has been archived by the owner on Apr 25, 2023. It is now read-only.

how to use the pipeline #1

Open
bryantChhun opened this issue Sep 11, 2020 · 0 comments
Open

how to use the pipeline #1

bryantChhun opened this issue Sep 11, 2020 · 0 comments

Comments

@bryantChhun
Copy link
Contributor

@miaecle
Please see recent changes to readme.

Our current use mode is to call specific wrapper functions (run_preproc, run_segmentation, etc..) and pass file paths to data locations. There are a few problems with how we've implemented this that i'd like to document here, and address soon:

  1. The wrapper functions have hardcoded paths to data.
  2. The wrapper functions have multiple operations that are never run simultaneously and that are selected by commenting/uncommenting code. (example: run_patch.py Worker can both "extract_patches" and "build_trajectories")
  3. Some critical batches of code are standalone scripts or reference files that are generated outside of this codebase (example: NNSegmentation.run.py is a script that calls this file -> Annotations_8Sites.pkl, which is not written anywhere within the codebase)

potential solutions:

  1. wrapper functions could have main with CLI parsing.
  2. wrapper functions need to be rewritten to accept kwargs for the specific type of run. This works well with CLI
  3. The two files that do this are both for NN training -- the Unet and the VQ-VAE.

Additionally, I think it could be generally useful to have all functions that write critical files also return that data. This will make a full script "pipeline" feasible.

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant