Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 3 of 3 for with_xla_support (0.07 sec)

  1. .bazelrc

    # Enable all targets in XLA
    build:cpu_cross --define=with_cross_compiler_support=true
    
    # Disable XLA on mobile.
    build:xla     --define=with_xla_support=true # TODO: remove, it's on by default.
    build:android --define=with_xla_support=false
    build:ios     --define=with_xla_support=false
    
    # BEGIN TF REMOTE BUILD EXECUTION OPTIONS
    # Options when using remote execution
    Registered: Tue Sep 09 12:39:10 UTC 2025
    - Last Modified: Fri Aug 22 21:03:34 UTC 2025
    - 56K bytes
    - Viewed (0)
  2. configure.py

            write_to_bazelrc(
                'build --copt="-DEIGEN_ALTIVEC_ENABLE_MMA_DYNAMIC_DISPATCH=1"')
    
      with_xla_support = environ_cp.get('TF_ENABLE_XLA', None)
      if with_xla_support is not None:
        write_to_bazelrc('build --define=with_xla_support=%s' %
                         ('true' if int(with_xla_support) else 'false'))
    
      set_action_env_var(
          environ_cp, 'TF_NEED_ROCM', 'ROCm', False, bazel_config_name='rocm')
    Registered: Tue Sep 09 12:39:10 UTC 2025
    - Last Modified: Wed Apr 30 15:18:54 UTC 2025
    - 48.3K bytes
    - Viewed (0)
  3. tensorflow/BUILD

        visibility = ["//visibility:public"],
    )
    
    config_setting(
        name = "with_xla_support",
        define_values = {"with_xla_support": "true"},
        visibility = ["//visibility:public"],
    )
    
    # By default, XLA GPU is compiled into tensorflow when building with
    # --config=cuda even when `with_xla_support` is false. The config setting
    # here allows us to override the behavior if needed.
    config_setting(
    Registered: Tue Sep 09 12:39:10 UTC 2025
    - Last Modified: Thu Aug 28 19:11:51 UTC 2025
    - 53.4K bytes
    - Viewed (0)
Back to top