- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 12 for locales (0.21 sec)
-
ci/official/wheel_test/README.md
[requirements updater readme](https://github.com/tensorflow/tensorflow/blob/master/ci/official/requirements_updater/README.md) ### Prerequisites for Local Testing To run tests locally, follow these steps: 1. Navigate to the relevant directory: ``` cd ci/official/wheel_test ``` 2. Run a script for creating requirements file: ```
Plain Text - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Aug 31 18:17:57 GMT 2023 - 3.3K bytes - Viewed (0) -
ci/official/containers/linux_arm64/builder.devtoolset/gcc9-fixups.patch
{ cp += 2 * nhere; continue; diff --git a/locale/xlocale.h b/locale/xlocale.h index 98c080b..843bd45 100644 --- a/locale/xlocale.h +++ b/locale/xlocale.h @@ -20,6 +20,9 @@ #ifndef _XLOCALE_H #define _XLOCALE_H 1 +#ifndef _BITS_TYPES___LOCALE_T_H +#define _BITS_TYPES___LOCALE_T_H 1 + /* Structure for reentrant locale using functions. This is an
Others - Registered: Tue May 07 12:40:20 GMT 2024 - Last Modified: Mon Sep 18 14:52:45 GMT 2023 - 8.9K bytes - Viewed (0) -
tensorflow/api_template.__init__.py
TensorFlow via the command `import tensorflow as tf`. The primary function of this module is to import all of the public TensorFlow interfaces into a single place. The interfaces themselves are located in sub-modules, as described below. Note that the file `__init__.py` in the TensorFlow source code tree is actually only a placeholder to enable test cases to run. The TensorFlow build replaces
Python - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Tue Mar 05 06:27:59 GMT 2024 - 6.7K bytes - Viewed (3) -
ci/official/requirements_updater/README.md
for an example. See [this commit](https://github.com/tensorflow/tensorflow/commit/5f7f05a80aac9b01325a78ec3fcff0dbedb1cc23) as a rough example of the steps below. All the files referenced below are located in the same directory as this README, unless indicated otherwise. 1) Add the new version to the `VERSIONS` variable inside `tensorflow/tools/toolchains/python/python_repo.bzl`. \
Plain Text - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Tue Jan 23 02:14:00 GMT 2024 - 3.9K bytes - Viewed (1) -
ci/official/README.md
# Ex. linux_x86_cuda -- x86_64 Linux platform, with Nvidia CUDA support # Ex. macos_arm64 -- arm64 MacOS platform # 3. Add modifiers. Some modifiers for local execution are: # Ex. disk_cache -- Use a local cache # Ex. public_cache -- Use TF's public cache (read-only) # Ex. public_cache_push -- Use TF's public cache (read and write, Googlers only)
Plain Text - Registered: Tue May 07 12:40:20 GMT 2024 - Last Modified: Thu Feb 01 03:21:19 GMT 2024 - 8K bytes - Viewed (0) -
CONTRIBUTING.md
``` If the tests are to be run on the GPU, add CUDA paths to LD_LIBRARY_PATH and add the `cuda` option flag ```bash export LD_LIBRARY_PATH="${LD_LIBRARY_PATH}:/usr/local/cuda/lib64:/usr/local/cuda/extras/CUPTI/lib64:$LD_LIBRARY_PATH" export flags="--config=opt --config=cuda -k" ``` For example, to run all tests under tensorflow/python, do: ```bash
Plain Text - Registered: Tue May 07 12:40:20 GMT 2024 - Last Modified: Thu Mar 21 11:45:51 GMT 2024 - 15.6K bytes - Viewed (0) -
tensorflow/c/eager/c_api_experimental.h
// EagerService to maintain consistency between client and worker. The // context_id is initialized with a dummy value and is later set when the worker // is initialized (either locally or remotely). The context_id can change during // the process lifetime although this should cause the worker to be // reinitialized (e.g. cleared caches) as well. TF_CAPI_EXPORT extern uint64_t TFE_GetContextId(TFE_Context* ctx);
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Wed Feb 21 22:37:46 GMT 2024 - 39.5K bytes - Viewed (0) -
.bazelrc
build:unsupported_gpu_linux --config=tensorrt build:unsupported_gpu_linux --action_env=CUDA_TOOLKIT_PATH="/usr/local/cuda-11.2" build:unsupported_gpu_linux --action_env=LD_LIBRARY_PATH="/usr/local/cuda:/usr/local/cuda/lib64:/usr/local/cuda/extras/CUPTI/lib64:/usr/local/cuda-11.1/lib64:/usr/local/tensorrt/lib" build:unsupported_gpu_linux --action_env=GCC_HOST_COMPILER_PATH="/dt9/usr/bin/gcc"
Plain Text - Registered: Tue May 07 12:40:20 GMT 2024 - Last Modified: Thu May 02 19:34:20 GMT 2024 - 52.8K bytes - Viewed (2) -
configure.py
if not is_linux(): raise ValueError('Currently NCCL is only supported on Linux platform.') if 'TF_NCCL_VERSION' in environ_cp: return ask_nccl_version = ( 'Please specify the locally installed NCCL version you want to use. ' '[Leave empty to use http://github.com/nvidia/nccl]: ') tf_nccl_version = get_from_env_or_user_or_default(environ_cp,
Python - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 18:25:36 GMT 2024 - 53.8K bytes - Viewed (1) -
ci/official/utilities/setup.sh
# $TFCI_OUTPUT_DIR/script.log exec > >(tee "$TFCI_OUTPUT_DIR/script.log") 2>&1 # Setup tfrun, a helper function for executing steps that can either be run # locally or run under Docker. setup_docker.sh, below, redefines it as "docker # exec". # Important: "tfrun foo | bar" is "( tfrun foo ) | bar", not "tfrun (foo | bar)". # Therefore, "tfrun" commands cannot include pipes -- which is
Shell Script - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri Jan 26 00:33:34 GMT 2024 - 5.2K bytes - Viewed (1)