- Sort Score
- Result 10 results
- Languages All
Results 1 - 1 of 1 for 23 (0.14 sec)
-
RELEASE.md
including matmuls and convolutions, due to the use of [TensorFloat-32](https://blogs.nvidia.com/blog/2020/05/14/tensorfloat-32-precision-format/). Specifically, inputs to such ops are rounded from 23 bits of precision to 10 bits of precision. This is unlikely to cause issues in practice for deep learning models. In some cases, TensorFloat-32 is also used for
Plain Text - Registered: Tue May 07 12:40:20 GMT 2024 - Last Modified: Mon Apr 29 19:17:57 GMT 2024 - 727.7K bytes - Viewed (8)