Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 1 of 1 for MatMul (0.34 sec)

  1. RELEASE.md

    * `tf.config.experimental.enable_tensor_float_32_execution`
        * Disabling TensorFloat-32 execution now causes TPUs to use float32 precision for float32 matmuls and other ops. TPUs have always used bfloat16 precision for certain ops, like matmul, when such ops had float32 inputs. Now, disabling TensorFloat-32 by calling `tf.config.experimental.enable_tensor_float_32_execution(False)` will cause TPUs to use float32 precision for such ops instead of bfloat16.
    
    Registered: Tue Sep 09 12:39:10 UTC 2025
    - Last Modified: Mon Aug 18 20:54:38 UTC 2025
    - 740K bytes
    - Viewed (1)
Back to top