Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Task killed when training for large zarr file #2

Open
DZGong opened this issue Sep 14, 2023 · 1 comment
Open

Task killed when training for large zarr file #2

DZGong opened this issue Sep 14, 2023 · 1 comment

Comments

@DZGong
Copy link

DZGong commented Sep 14, 2023

I was trying to train the model with a data set of 40x4096x4096x3 (NWHC), but the process was always killed, as shown in the following snapshot. This doesn't happen if I switch to a smaller data set (10x4096x4096) and the train goes well.

Screenshot 2023-09-13 at 3 00 36 PM

The dataset was originally a single tiff file and was transformed into a zarr file by using function zarr.convinient.save(). The data set was then split into train, val, and test by using the code multiscale_zarr_data_generator.py. Then started training by run.py.

The computation system I used has 68G(?) CPU memory and 16G GPU memory as shown in the following snapshot:
Screenshot 2023-09-13 at 9 59 50 PM

@DZGong
Copy link
Author

DZGong commented Sep 14, 2023

This is what goes after RuntimeError: DataLoader worker (pid 1749166) is killed by signal: Killed.
Screenshot 2023-09-13 at 10 22 34 PM

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant