- Sort Score
- Result 10 results
- Languages All
Results 11 - 18 of 18 for Jeep (0.41 sec)
-
ci/official/utilities/code_check_full.bats
# package bazel cquery --keep_going 'deps(//tensorflow/tools/pip_package:wheel)' | sort -u > $BATS_TEST_TMPDIR/pip_deps # Find all Python py_test targets not tagged "no_pip" or "manual", excluding # any targets in ignored packages. Combine this list of targets into a bazel # query list (e.g. the list becomes "target+target2+target3")
Plain Text - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Wed Mar 06 21:54:13 GMT 2024 - 13.2K bytes - Viewed (0) -
tensorflow/c/c_api_experimental.cc
// These XLA flags are needed to trigger XLA properly from C (more generally // non-Python) clients. If this API is called again with `enable` set to // false, it is safe to keep these flag values as is. tensorflow::MarkForCompilationPassFlags* flags = tensorflow::GetMarkForCompilationPassFlags(); flags->tf_xla_cpu_global_jit = true; flags->tf_xla_min_cluster_size = 1; } else {
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 03:35:10 GMT 2024 - 29.4K bytes - Viewed (0) -
.bazelrc
build:release_cpu_linux_base --repo_env=CC="/usr/lib/llvm-17/bin/clang" build:release_cpu_linux_base --repo_env=BAZEL_COMPILER="/usr/lib/llvm-17/bin/clang" # Test-related settings below this point. test:release_linux_base --build_tests_only --keep_going --test_output=errors --verbose_failures=true test:release_linux_base --local_test_jobs=HOST_CPUS test:release_linux_base --test_env=LD_LIBRARY_PATH # Give only the list of failed tests at the end of the log
Plain Text - Registered: Tue May 07 12:40:20 GMT 2024 - Last Modified: Thu May 02 19:34:20 GMT 2024 - 52.8K bytes - Viewed (2) -
tensorflow/c/eager/tape.h
template <typename Gradient, typename BackwardFunction, typename TapeTensor> class GradientTape { public: // If `persistent` is true, GradientTape will not eagerly delete backward // functions (and hence the tensors they keep alive). Instead, everything // is deleted in ~GradientTape. Persistent GradientTapes are useful when // users want to compute multiple gradients over the same tape.
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Tue Apr 02 12:40:29 GMT 2024 - 47.2K bytes - Viewed (1) -
ci/official/utilities/setup.sh
# functions (except for tfrun); use dedicated scripts instead and reference them # specifically. Use your best judgment to keep the scripts in this directory # lean and easy to follow. When in doubt, remember that for CI scripts, "keep it # simple" is MUCH more important than "don't repeat yourself." # -e: abort script if one command fails # -u: error if undefined variable used # -x: log all commands
Shell Script - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri Jan 26 00:33:34 GMT 2024 - 5.2K bytes - Viewed (1) -
ci/official/envs/ci_default
# (error on undefined variables). This ensures that (a) every TFCI variable # has an explicit default value, and (b) no script can accidentally use a # variable that doesn't exist. Please keep this list in alphabetical order. # # Find usage in scripts with e.g.: # cd ci/official # ls *.sh utilities/*.sh | xargs grep -H --color=always TFCI_ARG_HERE # You may also get an overview, e.g.: # cd ci/official
Plain Text - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Wed Mar 27 21:16:27 GMT 2024 - 2.5K bytes - Viewed (0) -
tensorflow/c/eager/gradient_checker.cc
sum_dims.reset(sum_dims_raw); } // Reduce sum the output on all dimensions. TF_RETURN_IF_ERROR(ops::Sum(ctx, model_out.get(), sum_dims.get(), &outputs[0], /*keep_dims=*/false, "sum_output")); return absl::OkStatus(); } // ========================= End Helper Functions============================== Status CalcNumericalGrad(AbstractContext* ctx, Model forward,
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 7.3K bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device_lib.cc
// Outputs: std::vector<TensorHandlePtr> op_outputs_ TF_GUARDED_BY(execution_mutex_); // TF_Status is an incomplete type and so can't be stack allocated. To avoid // unnecessary allocations each Execute call, we keep one heap-allocated // version for the thread. StatusPtr status_ TF_GUARDED_BY(execution_mutex_); const std::string device_; ExecutorPtr executor_ TF_GUARDED_BY(execution_mutex_);
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri Feb 09 07:47:20 GMT 2024 - 25.4K bytes - Viewed (1)