- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 18 for tpu (0.03 sec)
-
ci/official/containers/linux_arm64/devel.usertools/aarch64.bazelrc
# bazel test invocation as normal. test:nonpip_filters --test_tag_filters=-no_oss,-oss_serial,-gpu,-tpu,-benchmark-test,-v1only,-no_aarch64,-no_oss_py38,-no_oss_py39,-no_oss_py310 test:nonpip_filters --build_tag_filters=-no_oss,-oss_serial,-gpu,-tpu,-benchmark-test,-v1only,-no_aarch64,-no_oss_py38,-no_oss_py39,-no_oss_py310
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Fri Jul 12 20:16:57 UTC 2024 - 5.7K bytes - Viewed (0) -
ci/official/containers/linux_arm64/devel.usertools/aarch64_clang.bazelrc
# bazel test invocation as normal. test:nonpip_filters --test_tag_filters=-no_oss,-oss_serial,-gpu,-tpu,-benchmark-test,-v1only,-no_aarch64,-no_oss_py38,-no_oss_py39,-no_oss_py310 test:nonpip_filters --build_tag_filters=-no_oss,-oss_serial,-gpu,-tpu,-benchmark-test,-v1only,-no_aarch64,-no_oss_py38,-no_oss_py39,-no_oss_py310
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Fri Jul 12 20:16:57 UTC 2024 - 6.2K bytes - Viewed (0) -
ci/official/envs/linux_x86_tpu_build
source ci/official/envs/linux_x86_build TFCI_BAZEL_COMMON_ARGS="--repo_env=HERMETIC_PYTHON_VERSION=$TFCI_PYTHON_VERSION --config release_cpu_linux --config=tpu" TFCI_BAZEL_TARGET_SELECTING_CONFIG_PREFIX=linux_tpu TFCI_BUILD_PIP_PACKAGE_ARGS="--repo_env=WHEEL_NAME=tensorflow_tpu" TFCI_LIB_SUFFIX="-tpu-linux-x86_64" TFCI_WHL_BAZEL_TEST_ENABLE=0 TFCI_WHL_IMPORT_TEST_ENABLE=0 TFCI_WHL_SIZE_LIMIT=580M
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Tue Nov 05 08:00:04 UTC 2024 - 1.2K bytes - Viewed (0) -
.bazelrc
# dbg: Build with debug info # # TF version options; # v2: Build TF v2 # # Feature and Third party library support options: # xla: Build TF with XLA # tpu: Build TF with TPU support # cuda: Build with CUDA support. # cuda_clang Build with CUDA Clang support. # rocm: Build with AMD GPU support (rocm) # mkl: Enable full mkl support.
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Mon Oct 28 22:02:31 UTC 2024 - 51.3K bytes - Viewed (0) -
ci/official/envs/linux_x86_tpu
source ci/official/envs/linux_x86 TFCI_BAZEL_COMMON_ARGS="--repo_env=HERMETIC_PYTHON_VERSION=$TFCI_PYTHON_VERSION --config release_cpu_linux --config=tpu" TFCI_BAZEL_TARGET_SELECTING_CONFIG_PREFIX=linux_tpu TFCI_BUILD_PIP_PACKAGE_ARGS="--repo_env=WHEEL_NAME=tensorflow_tpu" TFCI_LIB_SUFFIX="-tpu-linux-x86_64" TFCI_WHL_BAZEL_TEST_ENABLE=0 TFCI_WHL_IMPORT_TEST_ENABLE=0 TFCI_WHL_SIZE_LIMIT=580M
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Mon Oct 14 23:45:36 UTC 2024 - 1.2K bytes - Viewed (0) -
.github/bot_config.yml
* It has an added advantage since you can you easily switch to different hardware accelerators (cpu, gpu, tpu) as per the task. * All you need is a good internet connection and you are all set. * Try to build TF from sources by changing CPU optimization flags. *Please let us know if this helps.*
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Mon Jul 15 05:00:54 UTC 2024 - 4K bytes - Viewed (0) -
tensorflow/c/eager/c_api_test_util.h
TFE_TensorHandle* axis); // If there is a device of type `device_type`, returns true // and sets 'device_name' accordingly. // `device_type` must be either "GPU" or "TPU". bool GetDeviceName(TFE_Context* ctx, tensorflow::string* device_name, const char* device_type); // Create a ServerDef with the given `job_name` and add `num_tasks` tasks in it.
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Mon Jul 17 23:43:59 UTC 2023 - 7.7K bytes - Viewed (0) -
tensorflow/BUILD
config_setting( name = "disable_mlir_bridge", define_values = {"enable_mlir_bridge": "false"}, visibility = ["//visibility:public"], ) # This flag enables experimental TPU support bool_flag( name = "enable_tpu_support", build_setting_default = False, ) config_setting( name = "with_tpu_support_define", define_values = {"with_tpu_support": "true"},
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Wed Oct 16 05:28:35 UTC 2024 - 53.5K bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device_test.cc
TFE_NewContext(opts.get(), status.get()), TFE_DeleteContext); ASSERT_EQ(TF_GetCode(status.get()), TF_OK) << TF_Message(status.get()); // Skip the test if no TPU is available. std::unique_ptr<TF_DeviceList, decltype(&TF_DeleteDeviceList)> devices( TFE_ContextListDevices(context.get(), status.get()), TF_DeleteDeviceList);
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Tue Aug 06 23:56:17 UTC 2024 - 29.4K bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device.cc
const TFE_OpAttrs* attributes, int expected_max_outputs, TF_Status* status) { absl::optional<std::vector<MaybeParallelTensorOwned>> result; // TODO(allenl): We should remove "TPU" from these op names at the very least, // or consider other ways of packing/unpacking parallel tensors. if (operation_name == std::string("TPUReplicatedInput")) {
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Mon Oct 21 04:14:14 UTC 2024 - 18.3K bytes - Viewed (0)