NB: In order to run pipelines using this HPC cluster, you must first apply for access here.
Job limits are capped at 32 CPUs, 385 GB and a maximum run time of 14 days.
This profile is configured to run with Apptainer, which is pre-loaded and globally available to all users on the cluster.
First, Nextflow should be loaded into your interactive session or added to your Slurm script using:
module load Nextflow
Using -profile fub_curta
will download the config file, which has been pre-configured with a setup suitable for the Curta cluster.
The default partition for job submissions is main
. With the current max_memory
settings, the user has access to 162 nodes on main
. The profile automatically detects whether the user belongs to the BeGenDiv working group, and gives the user access to 4 more nodes by making use of the begendiv
partition together with main
, giving precedence to the former.
In addition, job submissions are assigned the appropriate quality of service (QOS) as such:
hiprio
fortask.time <= 3.h
prio
fortask.time <= 3.d
standard
fortask.time <= 14.d
Deactivating the cleanup of intermediate files is also possible. It is done by specifying -profile fub_curta,debug
. This simply turns off automatic clean up of intermediate files, which can be useful for debugging.
If computing time on Curta has contributed to a publication or the completion of a degree course, please send us the corresponding DOI or BibTeX entry and, if possible, an appropriate image to [email protected]. This is the ONLY way we can document the usefulness of our service.
The current list of publications is available here.
Please also acknowledge the use of our service in any publications you produce, using the following reporting guidelines, including the DOI 10.17169/refubium-26754.