Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 4 of 4 for connectEnd (0.18 sec)

  1. ci/official/containers/linux_arm64/devel.usertools/code_check_full.bats

        "@local_config_cuda//cuda:curand + "\
        "@local_config_cuda//cuda:cusolver + "\
        "@local_config_tensorrt//:tensorrt)" --keep_going > $BATS_TEST_TMPDIR/out
    
      cat <<EOF
    There was a path found connecting //tensorflow/tools/pip_package:build_pip_package
    to a banned CUDA dependency. Here's the output from bazel query:
    EOF
      cat $BATS_TEST_TMPDIR/out
      [[ ! -s $BATS_TEST_TMPDIR/out ]]
    }
    
    Plain Text
    - Registered: Tue May 07 12:40:20 GMT 2024
    - Last Modified: Mon Sep 18 14:52:45 GMT 2023
    - 12.7K bytes
    - Viewed (0)
  2. RELEASE.md

        *   Add SeparableConv1D layer.
        *   Add convolutional Flipout layers.
        *   When both inputs of `tf.matmul` are bfloat16, it returns bfloat16,
            instead of float32.
        *   Added `tf.contrib.image.connected_components`.
        *   Add `tf.contrib.framework.CriticalSection` that allows atomic variable
            access.
        *   Output variance over trees predictions for classifications tasks.
    Plain Text
    - Registered: Tue May 07 12:40:20 GMT 2024
    - Last Modified: Mon Apr 29 19:17:57 GMT 2024
    - 727.7K bytes
    - Viewed (8)
  3. ci/official/utilities/code_check_full.bats

        "@local_config_cuda//cuda:curand + "\
        "@local_config_cuda//cuda:cusolver + "\
        "@local_config_tensorrt//:tensorrt)" --keep_going > $BATS_TEST_TMPDIR/out
    
      cat <<EOF
    There was a path found connecting //tensorflow/tools/pip_package:wheel
    to a banned CUDA dependency. Here's the output from bazel query:
    EOF
      cat $BATS_TEST_TMPDIR/out
      [[ ! -s $BATS_TEST_TMPDIR/out ]]
    }
    
    Plain Text
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Wed Mar 06 21:54:13 GMT 2024
    - 13.2K bytes
    - Viewed (0)
  4. tensorflow/c/eager/tape.h

            // operations which supposedly both created the same Tensor. It comes up
            // in recompute_grad, where the gradients have the same value. However,
            // only the original gradient is connected to everything else, so we
            // should still use that.
            vspace_.DeleteGradient(forward_grads[i]);
          } else {
            accumulated_gradients_[output_tensors[i].GetID()] = forward_grads[i];
          }
    C
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Tue Apr 02 12:40:29 GMT 2024
    - 47.2K bytes
    - Viewed (1)
Back to top