- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 26 for courant (0.18 sec)
-
tensorflow/c/eager/tape.h
} } } auto usage_count_it = state.tensor_usage_counts.find(id); if (usage_count_it == state.tensor_usage_counts.end()) { VLOG(1) << "Tensor " << id << " not used"; continue; } usage_count_it->second--; if (usage_count_it->second > 0) { VLOG(1) << "Tensor " << id << " usage count " << usage_count_it->second; continue; }
C - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Tue Apr 02 12:40:29 GMT 2024 - 47.2K bytes - Viewed (1) -
ci/official/utilities/code_check_full.bats
"@local_config_cuda//cuda:cudart + "\ "@local_config_cuda//cuda:cudart + "\ "@local_config_cuda//cuda:cuda_driver + "\ "@local_config_cuda//cuda:cudnn + "\ "@local_config_cuda//cuda:curand + "\ "@local_config_cuda//cuda:cusolver + "\ "@local_config_tensorrt//:tensorrt)" --keep_going > $BATS_TEST_TMPDIR/out cat <<EOF There was a path found connecting //tensorflow/tools/pip_package:wheel
Plain Text - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Wed Mar 06 21:54:13 GMT 2024 - 13.2K bytes - Viewed (0) -
configure.py
# ROCm / CUDA are mutually exclusive. # At most 1 GPU platform can be configured. gpu_platform_count = 0 if environ_cp.get('TF_NEED_ROCM') == '1': gpu_platform_count += 1 if environ_cp.get('TF_NEED_CUDA') == '1': gpu_platform_count += 1 if gpu_platform_count >= 2: raise UserInputError('CUDA / ROCm are mututally exclusive. '
Python - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 18:25:36 GMT 2024 - 53.8K bytes - Viewed (0) -
tensorflow/c/c_api_experimental.h
// a) ConfigProto.optimizer_options.global_jit_level is set to ON_1 if // `enable_xla_compilation` is non-zero, and OFF otherwise. // b) ConfigProto.gpu_options.allow_growth is set to `gpu_memory_allow_growth`. // c) ConfigProto.device_count is set to `num_cpu_devices`. TF_CAPI_EXPORT extern TF_Buffer* TF_CreateConfig( unsigned char enable_xla_compilation, unsigned char gpu_memory_allow_growth, unsigned int num_cpu_devices);
C - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Thu Apr 27 21:07:00 GMT 2023 - 15.1K bytes - Viewed (0) -
tensorflow/c/c_api_experimental.cc
} auto* gpu_options = config.mutable_gpu_options(); gpu_options->set_allow_growth(gpu_memory_allow_growth); (*config.mutable_device_count())["CPU"] = num_cpu_devices; // TODO(b/113217601): This is needed for EagerContext::runner_ to use a // threadpool, so that we avoid the possibility of running the runner_ in the
C++ - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 03:35:10 GMT 2024 - 29.4K bytes - Viewed (0) -
tensorflow/c/env.cc
void TF_DeleteRecursively(const char* dirname, int64_t* undeleted_file_count, int64_t* undeleted_dir_count, TF_Status* status) { ::int64_t f, d; TF_SetStatus(status, TF_OK, ""); ::tensorflow::Set_TF_Status_from_Status( status, ::tensorflow::Env::Default()->DeleteRecursively(dirname, &f, &d)); *undeleted_file_count = f; *undeleted_dir_count = d; }
C++ - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Wed Aug 11 01:20:50 GMT 2021 - 7K bytes - Viewed (0) -
tensorflow/c/experimental/filesystem/plugins/gcs/expiring_lru_cache.h
/// that no entry is ever placed in the cache. const uint64_t max_age_; /// The maximum number of entries in the cache. A value of 0 means there is no /// limit on entry count. const size_t max_entries_; /// The callback to read timestamps. std::function<uint64_t()> timer_seconds_; /// Guards access to the cache and the LRU list. absl::Mutex mu_;
C - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Thu Jul 09 19:31:22 GMT 2020 - 6.3K bytes - Viewed (0) -
tensorflow/c/eager/c_api_experimental.h
typedef struct TFE_MonitoringBuckets TFE_MonitoringBuckets; TF_CAPI_EXPORT extern TFE_MonitoringBuckets* TFE_MonitoringNewExponentialBuckets(double scale, double growth_factor, int bucket_count); TF_CAPI_EXPORT extern void TFE_MonitoringDeleteBuckets( TFE_MonitoringBuckets* buckets); // APIs for Sampler without label. typedef struct TFE_MonitoringSampler0 TFE_MonitoringSampler0;
C - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Wed Feb 21 22:37:46 GMT 2024 - 39.5K bytes - Viewed (0) -
ci/official/libtensorflow.sh
# limitations under the License. # ============================================================================== source "${BASH_SOURCE%/*}/utilities/setup.sh" # Record GPU count and CUDA version status if [[ "$TFCI_NVIDIA_SMI_ENABLE" == 1 ]]; then tfrun nvidia-smi fi # Update the version numbers for Nightly only if [[ "$TFCI_NIGHTLY_UPDATE_VERSION_ENABLE" == 1 ]]; then
Shell Script - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Fri Jan 19 19:07:48 GMT 2024 - 1.5K bytes - Viewed (0) -
ci/official/wheel.sh
# limitations under the License. # ============================================================================== source "${BASH_SOURCE%/*}/utilities/setup.sh" # Record GPU count and CUDA version status if [[ "$TFCI_NVIDIA_SMI_ENABLE" == 1 ]]; then tfrun nvidia-smi fi # Update the version numbers for Nightly only if [[ "$TFCI_NIGHTLY_UPDATE_VERSION_ENABLE" == 1 ]]; then
Shell Script - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Wed Mar 06 21:54:13 GMT 2024 - 1.8K bytes - Viewed (0)