使用Player FM应用程序离线!
Half precision
Manage episode 301973966 series 2921809
In this episode I talk about reduced precision floating point formats float16 (aka half precision) and bfloat16. I'll discuss what floating point numbers are, how these two formats vary, and some of the practical considerations that arise when you are working with numeric code in PyTorch that also needs to work in reduced precision. Did you know that we do all CUDA computations in float32, even if the source tensors are stored as float16? Now you know!
Further reading.
- The Wikipedia article on IEEE floating point is pretty great https://en.wikipedia.org/wiki/IEEE_754
- How bfloat16 works out when doing training https://arxiv.org/abs/1905.12322
- Definition of acc_type in PyTorch https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/AccumulateType.h
83集单集
Manage episode 301973966 series 2921809
In this episode I talk about reduced precision floating point formats float16 (aka half precision) and bfloat16. I'll discuss what floating point numbers are, how these two formats vary, and some of the practical considerations that arise when you are working with numeric code in PyTorch that also needs to work in reduced precision. Did you know that we do all CUDA computations in float32, even if the source tensors are stored as float16? Now you know!
Further reading.
- The Wikipedia article on IEEE floating point is pretty great https://en.wikipedia.org/wiki/IEEE_754
- How bfloat16 works out when doing training https://arxiv.org/abs/1905.12322
- Definition of acc_type in PyTorch https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/AccumulateType.h
83集单集
所有剧集
×欢迎使用Player FM
Player FM正在网上搜索高质量的播客,以便您现在享受。它是最好的播客应用程序,适用于安卓、iPhone和网络。注册以跨设备同步订阅。