Feb 14, 2020
Artificial Intelligence Gets Its Own System of Numbers
Posted by Genevieve Klien in categories: mathematics, robotics/AI
BF16, the new number format optimized for deep learning, promises power and compute savings with a minimal reduction in prediction accuracy.
BF16, sometimes called BFloat16 or Brain Float 16, is a new number format optimised for AI/deep learning applications. Invented at Google Brain, it has gained wide adoption in AI accelerators from Google, Intel, Arm and many others.
The idea behind BF16 is to reduce the compute power and energy consumption needed to multiply tensors together by reducing the precision of the numbers. A tensor is a three-dimensional matrix of numbers; multiplication of tensors is the key mathematical operation required for AI calculations.