Skip to content

Query for mixed precision training #27

Answered by iliaschalkidis
BenjaminKKK asked this question in Q&A
Discussion options

You must be logged in to vote

Hi @BenjaminKKK, as I mentioned in this recent issue #26, these two HF arguments/parameters ( --fp16, --fp16_full_eval) are only applicable (working) when there are available (and correctly configured) NVIDIA GPUs in a machine station (server or cluster) and also torch is correctly configured to use these compute resources.

So, in case you don't have such resources, just delete these two arguments to train models with standard fp32 precision. In case you have such resources, make sure to correctly install the NVIDIA CUDA drivers, and also correctly install torch (consider this page to figure out the appropriate steps: https://pytorch.org/get-started/locally/)

I'll update the README.md to …

Replies: 1 comment 2 replies

Comment options

You must be logged in to vote
2 replies
@BenjaminKKK
Comment options

@iliaschalkidis
Comment options

Answer selected by iliaschalkidis
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants