Tag: BF16

Mixed-Precision Training for Large Language Models: FP16, BF16, and Beyond 16 December 2025

Mixed-Precision Training for Large Language Models: FP16, BF16, and Beyond

Mixed-precision training using FP16 and BF16 cuts LLM training time by up to 70% and reduces memory use by half. Learn how it works, why BF16 is now preferred over FP16, and how to implement it safely with PyTorch.

Susannah Greenwood 8 Comments