- Sort Score
- Result 10 results
- Languages All
Results 1 - 3 of 3 for support (0.21 sec)
-
tensorflow/BUILD
config_setting( name = "stackdriver_support", define_values = {"stackdriver_support": "true"}, visibility = ["//visibility:public"], ) config_setting( name = "with_xla_support", define_values = {"with_xla_support": "true"}, visibility = ["//visibility:public"], ) # By default, XLA GPU is compiled into tensorflow when building with # --config=cuda even when `with_xla_support` is false. The config setting
Plain Text - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Tue Apr 09 18:15:11 GMT 2024 - 53.4K bytes - Viewed (8) -
.bazelrc
# v2: Build TF v2 # # Feature and Third party library support options: # xla: Build TF with XLA # tpu: Build TF with TPU support # cuda: Build with CUDA support. # cuda_clang Build with CUDA Clang support. # rocm: Build with AMD GPU support (rocm) # mkl: Enable full mkl support. # tensorrt: Enable Tensorrt support. # noaws: Disable AWS S3 storage support
Plain Text - Registered: Tue May 07 12:40:20 GMT 2024 - Last Modified: Thu May 02 19:34:20 GMT 2024 - 52.8K bytes - Viewed (2) -
configure.py
write_to_bazelrc( 'build --copt="-DEIGEN_ALTIVEC_ENABLE_MMA_DYNAMIC_DISPATCH=1"') with_xla_support = environ_cp.get('TF_ENABLE_XLA', None) if with_xla_support is not None: write_to_bazelrc('build --define=with_xla_support=%s' % ('true' if int(with_xla_support) else 'false')) set_action_env_var( environ_cp, 'TF_NEED_ROCM', 'ROCm', False, bazel_config_name='rocm')
Python - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 18:25:36 GMT 2024 - 53.8K bytes - Viewed (1)