mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-31 18:22:34 +06:00
Corrected max number for bf16 in transformer/docs (#33658)
Update perf_train_gpu_one.md per issue https://github.com/huggingface/hub-docs/issues/1425 max number for bf16 should be 65,504 not 65,535
This commit is contained in:
parent
196d35ccfc
commit
ade9e0fe41
@ -186,7 +186,7 @@ If you prefer to use 🤗 Accelerate, find the 🤗 Accelerate example [further
|
||||
|
||||
If you have access to an Ampere or newer hardware you can use bf16 for mixed precision training and evaluation. While
|
||||
bf16 has a worse precision than fp16, it has a much bigger dynamic range. In fp16 the biggest number you can have
|
||||
is `65535` and any number above that will result in an overflow. A bf16 number can be as large as `3.39e+38` (!) which
|
||||
is `65504` and any number above that will result in an overflow. A bf16 number can be as large as `3.39e+38` (!) which
|
||||
is about the same as fp32 - because both have 8-bits used for the numerical range.
|
||||
|
||||
You can enable BF16 in the 🤗 Trainer with:
|
||||
|
Loading…
Reference in New Issue
Block a user