1
0
mirror of https://github.com/huggingface/diffusers.git synced 2026-01-27 17:22:53 +03:00

[Docs] add note on fp16 in fast diffusion (#6380)

add note on fp16
This commit is contained in:
Sayak Paul
2023-12-29 09:38:50 +05:30
committed by GitHub
parent e7044a4221
commit 203724e9d9

View File

@@ -96,6 +96,8 @@ bfloat16 reduces the latency from 7.36 seconds to 4.63 seconds:
</div>
_(We later ran the experiments in float16 and found out that the recent versions of torchao do not incur numerical problems from float16.)_
**Why bfloat16?**
* Using a reduced numerical precision (such as float16, bfloat16) to run inference doesnt affect the generation quality but significantly improves latency.