- Sort Score
- Result 10 results
- Languages All
Results 1 - 7 of 7 for fakefs (0.28 sec)
-
tensorflow/c/eager/unified_api_testutil.h
Status CreateParamsForInputs(AbstractContext* ctx, absl::Span<AbstractTensorHandle* const> inputs, std::vector<AbstractTensorHandle*>* params); // A callable that takes tensor inputs and returns zero or more tensor outputs. using Model = std::function<Status(AbstractContext*, absl::Span<AbstractTensorHandle* const>,
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Tue Feb 27 13:57:45 GMT 2024 - 4K bytes - Viewed (0) -
ci/official/containers/linux_arm64/devel.usertools/aarch64.bazelrc
# installed into a virtual environment, and then that venv is used to run all # bazel tests with a special flag "--define=no_tensorflow_py_deps=true", which # drops all the bazel dependencies for each py_test; this makes all the tests # use the wheel's TensorFlow installation instead of the one made available # through bazel. This must be done in a different root directory, //bazel_pip/...,
Plain Text - Registered: Tue May 07 12:40:20 GMT 2024 - Last Modified: Tue Nov 21 12:25:39 GMT 2023 - 5.8K bytes - Viewed (0) -
ci/official/README.md
1. Creating a PR and observing the presubmit test results 2. Running the CI scripts locally, as explained below 3. **Google employees only**: Google employees can use an internal-only tool called "MLCI" that makes testing more convenient: it can execute any full CI job against a pending change. Search for "MLCI" internally to find it. You may invoke a CI script of your choice by following these instructions: ```bash
Plain Text - Registered: Tue May 07 12:40:20 GMT 2024 - Last Modified: Thu Feb 01 03:21:19 GMT 2024 - 8K bytes - Viewed (0) -
ci/official/requirements_updater/README.md
``` --repo_env=TF_PYTHON_VERSION=3.11 ``` ## Requirements updater Requirements updater is a standalone tool, intended to simplify process of updating requirements for multiple minor versions of Python. It takes in a file with a set of dependencies, and produces a more detailed requirements file for each version, with hashes specified for each dependency required, as well as their sub-dependencies. ### How to update/add requirements
Plain Text - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Tue Jan 23 02:14:00 GMT 2024 - 3.9K bytes - Viewed (1) -
tensorflow/c/experimental/gradients/array_grad.cc
for (int i = 0; i < grad_outputs.size(); i++) { auto grad_input = grad_outputs[i]; // TODO(srbs): Should we add a copy contructor to AbstractTensorHandle // that takes care of this similar to `Tensor`? if (grad_input) { grad_input->Ref(); } grad_inputs[i] = grad_input; } return absl::OkStatus(); } ~IdentityNGradientFunction() override {}
C++ - Registered: Tue Apr 09 12:39:09 GMT 2024 - Last Modified: Wed Feb 28 13:53:47 GMT 2024 - 1.6K bytes - Viewed (0) -
ci/official/containers/linux_arm64/devel.usertools/squash_testlogs.py
r._elem.remove(testsuite._elem) if len(r) > 0: # pylint: disable=g-explicit-length-test result += r # Insert the number of failures for each test to help identify flakes # need to clarify for shard for p in result._elem.xpath(".//error | .//failure"): key = re.sub(r"0x\w+", "", p.getparent().get("name", "")) + p.text p.text = runfiles_matcher.sub("[testroot]/", p.text)
Python - Registered: Tue May 07 12:40:20 GMT 2024 - Last Modified: Mon Sep 18 19:00:37 GMT 2023 - 4.8K bytes - Viewed (0) -
ci/official/upload.sh
# "/path/bar". This script uses "gsutil rsync" instead, which acts on directory # contents. About arguments to gsutil: # "gsutil -m rsync" runs in parallel. # "gsutil rsync -r" is recursive and makes directories work. # "gsutil rsync -d" is "sync and delete files from destination if not present in source" DOWNLOADS="$(mktemp -d)" mkdir -p "$DOWNLOADS" gsutil -m rsync -r "$TFCI_ARTIFACT_STAGING_GCS_URI" "$DOWNLOADS"
Shell Script - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Wed Jan 24 20:52:12 GMT 2024 - 2.8K bytes - Viewed (0)