- Sort Score
- Result 10 results
- Languages All
Results 1 - 9 of 9 for logs (0.18 sec)
-
.github/ISSUE_TEMPLATE/tflite-other.md
share a link to Colab/Jupyter/any notebook. placeholder: Tell us what you see! value: render: shell validations: required: true - type: textarea id: logs attributes: label: Relevant log output description: Please copy and paste any relevant log output. This will be automatically
Plain Text - Registered: Tue May 07 12:40:20 GMT 2024 - Last Modified: Thu Dec 29 22:28:29 GMT 2022 - 3.4K bytes - Viewed (1) -
ci/official/utilities/generate_index_html.sh
# presubmits. # Usage: generate_index_html.sh /path/to/output/index.html cat > "$1" <<EOF <html> <head> <title>$(basename "$KOKORO_JOB_NAME")</title> </head> <body> <h1>TensorFlow Job Logs and Links</h1> <h2>Job Details</h2> <ul> <li>Job name: $KOKORO_JOB_NAME</li> <li>Job pool: $KOKORO_JOB_POOL</li> <li>Job ID: $KOKORO_BUILD_ID</li>
Shell Script - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri Sep 29 20:26:13 GMT 2023 - 2.3K bytes - Viewed (0) -
.github/ISSUE_TEMPLATE/tflite-in-play-services.md
the problem. If possible, please share a link to or attach code demonstrating the problem. **Any other info / logs** Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files
Plain Text - Registered: Tue May 07 12:40:20 GMT 2024 - Last Modified: Wed Jun 15 03:35:58 GMT 2022 - 880 bytes - Viewed (0) -
.github/ISSUE_TEMPLATE/tflite-converter-issue.md
### 4. (optional) RNN conversion support If converting TF RNN to TFLite fused RNN ops, please prefix [RNN] in the title. ### 5. (optional) Any other info / logs
Plain Text - Registered: Tue May 07 12:40:20 GMT 2024 - Last Modified: Wed Jun 15 03:35:58 GMT 2022 - 2.1K bytes - Viewed (0) -
ci/official/envs/versions_upload
# limitations under the License. # ============================================================================== # Release jobs are very basic. They don't use any caching or RBE, # but they do upload logs to resultstore. # IMPORTANT: trailing slash is required on GCS URIs, as it tells gcloud to # pretend the path is a directory. TFCI_ARTIFACT_FINAL_GCS_ENABLE=1
Plain Text - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri Jan 19 19:07:48 GMT 2024 - 1.6K bytes - Viewed (0) -
.github/ISSUE_TEMPLATE/tflite-op-request.md
Also, please include a link to a GraphDef or the model if possible. **Any other info / logs** Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files
Plain Text - Registered: Tue May 07 12:40:20 GMT 2024 - Last Modified: Wed Jun 15 03:35:58 GMT 2022 - 879 bytes - Viewed (0) -
.github/ISSUE_TEMPLATE/tensorflow_issue_template.yaml
placeholder: Tell us what you see! value: render: shell validations: required: true - type: textarea id: logs attributes: label: Relevant log output description: Please copy and paste any relevant log output. This will be automatically formatted into code, so no need for backticks.
Others - Registered: Tue May 07 12:40:20 GMT 2024 - Last Modified: Wed Jun 28 18:25:42 GMT 2023 - 3.7K bytes - Viewed (0) -
tensorflow/c/experimental/gradients/nn_grad_test.cc
AbstractContext* ctx, absl::Span<AbstractTensorHandle* const> inputs, absl::Span<AbstractTensorHandle*> outputs) { AbstractTensorHandle* loss; AbstractTensorHandle* backprop; TF_RETURN_IF_ERROR(ops::SparseSoftmaxCrossEntropyWithLogits( ctx, inputs[0], inputs[1], &loss, &backprop, "SparseSoftmaxCrossEntropyWithLogits")); // `gradient_checker` only works with model that returns only 1 tensor.
C++ - Registered: Tue Mar 26 12:39:09 GMT 2024 - Last Modified: Wed Feb 28 13:53:47 GMT 2024 - 8.3K bytes - Viewed (0) -
ci/official/containers/linux_arm64/devel.usertools/squash_testlogs.py
files = subprocess.check_output( ["grep", "-rlE", '(failures|errors)="[1-9]', sys.argv[1]] ) except subprocess.CalledProcessError as e: print("No failures found to log!") exit(0) # For test cases, only show the ones that failed that have text (a log) seen = collections.Counter() runfiles_matcher = re.compile(r"(/.*\.runfiles/)") for f in files.strip().splitlines(): # Just ignore any failures, they're probably not important
Python - Registered: Tue May 07 12:40:20 GMT 2024 - Last Modified: Mon Sep 18 19:00:37 GMT 2023 - 4.8K bytes - Viewed (0)