- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 41 for Suda (0.04 sec)
-
tensorflow/compiler/jit/tests/auto_clustering_test.cc
} }; TEST_F(AutoClusteringTestImpl, KerasImagenetMain) { // Generated from // // TARGET_PATH=tensorflow_models/official/legacy/image_classification \ // bazel run -c opt --config=cuda ${TARGET_PATH}:resnet_imagenet_main \ // -- --skip_eval --num_gpus=1 --dtype=fp16 --batch_size=192 \ // --train_steps=210 --enable_xla --enable_eager=true // // At CL 245846452
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jan 13 20:13:03 UTC 2022 - 3.6K bytes - Viewed (0) -
.github/ISSUE_TEMPLATE/tflite-other.md
- type: input id: Compiler attributes: label: GCC/Compiler version description: if compiling from source placeholder: validations: required: false - type: input id: Cuda attributes: label: CUDA/cuDNN version description: placeholder: validations: required: false - type: input id: Gpu attributes: label: GPU model and memory description: if
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Dec 29 22:28:29 UTC 2022 - 3.4K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_host_send_recv_device_context_test.cc
Tensor dest_cpu_tensor(host_allocator_, DT_FLOAT, TensorShape({2, 2})); stream_executor::Platform* platform = stream_executor::PlatformManager::PlatformWithName("CUDA").value(); stream_executor::StreamExecutor* executor = platform->ExecutorForDevice(0).value(); TF_ASSERT_OK_AND_ASSIGN(auto stream, executor->CreateStream());
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 22:46:36 UTC 2024 - 7.2K bytes - Viewed (0) -
ci/official/libtensorflow.sh
# limitations under the License. # ============================================================================== source "${BASH_SOURCE%/*}/utilities/setup.sh" # Record GPU count and CUDA version status if [[ "$TFCI_NVIDIA_SMI_ENABLE" == 1 ]]; then tfrun nvidia-smi fi # Update the version numbers for Nightly only if [[ "$TFCI_NIGHTLY_UPDATE_VERSION_ENABLE" == 1 ]]; then
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Jan 19 19:07:48 UTC 2024 - 1.5K bytes - Viewed (0) -
.github/ISSUE_TEMPLATE/tensorflow_issue_template.yaml
description: If compiling from source - type: input id: Compiler attributes: label: GCC/compiler version description: If compiling from source - type: input id: Cuda attributes: label: CUDA/cuDNN version - type: input id: Gpu attributes: label: GPU model and memory description: If compiling from source - type: textarea id: what-happened
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 28 18:25:42 UTC 2023 - 3.7K bytes - Viewed (0) -
ci/official/wheel.sh
# limitations under the License. # ============================================================================== source "${BASH_SOURCE%/*}/utilities/setup.sh" # Record GPU count and CUDA version status if [[ "$TFCI_NVIDIA_SMI_ENABLE" == 1 ]]; then tfrun nvidia-smi fi # Update the version numbers for Nightly only if [[ "$TFCI_NIGHTLY_UPDATE_VERSION_ENABLE" == 1 ]]; then
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Mar 06 21:54:13 UTC 2024 - 1.8K bytes - Viewed (0) -
tensorflow/BUILD
# Config setting that is satisfied when TensorFlow is being built with CUDA # support through e.g. `--config=cuda` (or `--config=cuda_clang` in OSS). alias( name = "is_cuda_enabled", actual = if_oss( "@local_config_cuda//:is_cuda_enabled", "@local_config_cuda//cuda:using_clang", ), ) # Config setting that is satisfied when CUDA device code should be compiled
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 31 16:51:59 UTC 2024 - 53.5K bytes - Viewed (0) -
ci/official/README.md
These "env" files match up with an environment matrix that roughly covers: - Different Python versions - Linux, MacOS, and Windows machines (these pool definitions are internal) - x86 and arm64 - CPU-only, or with NVIDIA CUDA support (Linux only), or with TPUs ## How to Test Your Changes to TensorFlow You may check how your changes will affect TensorFlow by: 1. Creating a PR and observing the presubmit test results
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 01 03:21:19 UTC 2024 - 8K bytes - Viewed (0) -
ci/official/envs/linux_arm64
TFCI_BAZEL_TARGET_SELECTING_CONFIG_PREFIX=linux_arm64 # Note: this is not set to "--cpu", because that changes the package name # to tensorflow_cpu. These ARM builds are supposed to have the name "tensorflow" # despite lacking Nvidia CUDA support. TFCI_BUILD_PIP_PACKAGE_ARGS="--repo_env=WHEEL_NAME=tensorflow" TFCI_DOCKER_ENABLE=1 TFCI_DOCKER_IMAGE=us-central1-docker.pkg.dev/tensorflow-sigs/tensorflow/build-arm64:tf-2-16-multi-python TFCI_DOCKER_PULL_ENABLE=1
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Jun 03 16:53:33 UTC 2024 - 1.5K bytes - Viewed (0) -
CONTRIBUTING.md
export flags="--config=opt -k" ``` If the tests are to be run on the GPU, add CUDA paths to LD_LIBRARY_PATH and add the `cuda` option flag ```bash export LD_LIBRARY_PATH="${LD_LIBRARY_PATH}:/usr/local/cuda/lib64:/usr/local/cuda/extras/CUPTI/lib64:$LD_LIBRARY_PATH" export flags="--config=opt --config=cuda -k" ``` For example, to run all tests under tensorflow/python, do: ```bash
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Mar 21 11:45:51 UTC 2024 - 15.6K bytes - Viewed (0)