Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 6 of 6 for tf_xla_enable_xla_devices (0.39 sec)

  1. tensorflow/compiler/jit/flags.cc

                "Switch a device into 'on-demand' mode, where instead of "
                "autoclustering ops are compiled one by one just-in-time."),
    
           Flag("tf_xla_enable_xla_devices",
                &device_flags->tf_xla_enable_xla_devices,
                "Generate XLA_* devices, where placing a computation on such a "
                "device"
                "forces compilation by XLA. Deprecated."),
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Apr 17 18:52:57 UTC 2024
    - 24.5K bytes
    - Viewed (0)
  2. tensorflow/compiler/jit/xla_platform_info_test.cc

        DeviceCompiler<xla::PjRtLoadedExecutable, xla::PjRtClient>;
    
    class XlaPlatformInfoTest : public ::testing::Test {
     protected:
      void SetUp() override {
        tensorflow::GetXlaDeviceFlags()->tf_xla_enable_xla_devices = true;
        tensorflow::GetMarkForCompilationPassFlags()
            ->tf_xla_persistent_cache_directory = "";
        tensorflow::GetMarkForCompilationPassFlags()
            ->tf_xla_persistent_cache_device_types = "";
      }
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sun Jan 14 15:17:12 UTC 2024
    - 13.6K bytes
    - Viewed (0)
  3. tensorflow/compiler/jit/xla_launch_util_gpu_test.cc

        rollout_config.enabled_for_gpu_ = true;
    
        // Set flag to enable using XLA devices. PJRT currently is only supported
        // for XLA devices.
        GetXlaDeviceFlags()->tf_xla_enable_xla_devices = true;
    
        // Add and setup the GPU device.
        auto device_type = DeviceType(DEVICE_GPU);
        auto jit_device_type = DeviceType(DEVICE_GPU);
        rollout_config.AllowForDeviceInXlaLaunch(device_type);
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Sep 06 19:12:29 UTC 2023
    - 10K bytes
    - Viewed (0)
  4. tensorflow/compiler/jit/xla_compiler_options_util_test.cc

    }
    
    class XlaCompilerOptionsTest : public ::testing::Test {
     protected:
      void SetUp() override {
        tensorflow::GetXlaDeviceFlags()->tf_xla_enable_xla_devices = true;
      }
    
      DeviceSetup device_setup_;
    };
    
    TEST_F(XlaCompilerOptionsTest, PjRtOptionsXlaDevice) {
      device_setup_.AddDevicesAndSetUp({DEVICE_XLA_GPU});
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Dec 29 01:41:20 UTC 2023
    - 14.8K bytes
    - Viewed (0)
  5. tensorflow/compiler/jit/flags.h

      // feature is battle-tested, we will switch this to be a session option.
      bool tf_xla_compile_on_demand;
    
      // Enables "XLA" devices if this flag is set.
      bool tf_xla_enable_xla_devices;
    };
    
    // Flags common to the _Xla* ops and their kernels.
    struct XlaOpsCommonFlags {
      // If true, _XlaCompile always refuses to compile the cluster, which means the
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Apr 17 18:52:57 UTC 2024
    - 14.5K bytes
    - Viewed (0)
  6. tensorflow/compiler/jit/xla_launch_util_test.cc

        rollout_config.enabled_for_compile_on_demand_ = true;
    
        // Set flag to enable using XLA devices. PJRT currently is only supported
        // for XLA devices.
        GetXlaDeviceFlags()->tf_xla_enable_xla_devices = true;
    
        // Add and setup the XLA_CPU device.
        auto device_type = DeviceType(DEVICE_XLA_CPU);
        rollout_config.AllowForDeviceInXlaLaunch(device_type);
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 21 09:53:30 UTC 2024
    - 28.8K bytes
    - Viewed (0)
Back to top