Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 4 of 4 for we (0.11 sec)

  1. tensorflow/c/experimental/gradients/array_grad.cc

                     absl::Span<AbstractTensorHandle*> grad_inputs) override {
        for (int i = 0; i < grad_outputs.size(); i++) {
          auto grad_input = grad_outputs[i];
          // TODO(srbs): Should we add a copy contructor to AbstractTensorHandle
          // that takes care of this similar to `Tensor`?
          if (grad_input) {
            grad_input->Ref();
          }
          grad_inputs[i] = grad_input;
        }
    C++
    - Registered: Tue Apr 09 12:39:09 GMT 2024
    - Last Modified: Wed Feb 28 13:53:47 GMT 2024
    - 1.6K bytes
    - Viewed (0)
  2. tensorflow/c/experimental/gradients/array_grad_test.cc

    };
    
    TEST_P(CppGradients, TestIdentityNGrad) {
      // This test is interesting because the current implementation of GradientTape
      // would return [0, 1] whereas we use build_default_zeros_grads=false here
      // so we get back [nullptr, 1].
    
      AbstractTensorHandlePtr x1;
      {
        AbstractTensorHandle* x1_raw = nullptr;
        status_ = TestScalarTensorHandle<float, TF_FLOAT>(
    C++
    - Registered: Tue Mar 26 12:39:09 GMT 2024
    - Last Modified: Wed Feb 28 13:53:47 GMT 2024
    - 5K bytes
    - Viewed (0)
  3. tensorflow/c/eager/unified_api_testutil.cc

        const char* fn_name = "test_fn";
        core::RefCountPtr<AbstractFunction> scoped_func;
        // Returning null tensors from a tf.function is not supported, so we keep
        // track of indices in the model's outputs are nullptr in this set.
        // The FunctionDef only outputs the non-null tensors. We later pad the
        // function op outputs to have nullptrs at the `null_indices`.
        absl::flat_hash_set<int> null_indices;
        {
    C++
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Tue Feb 27 13:57:45 GMT 2024
    - 5.7K bytes
    - Viewed (0)
  4. tensorflow/c/experimental/gradients/nn_grad_test.cc

      // `gradient_checker` only works with model that returns only 1 tensor.
      // Although, `ops::SparseSoftmaxCrossEntropyWithLogits` returns 2 tensors, the
      // second tensor isn't needed for computing gradient so we could safely drop
      // it.
      outputs[0] = loss;
      backprop->Unref();
      return absl::OkStatus();
    }
    
    Status BiasAddModel(AbstractContext* ctx,
                        absl::Span<AbstractTensorHandle* const> inputs,
    C++
    - Registered: Tue Mar 26 12:39:09 GMT 2024
    - Last Modified: Wed Feb 28 13:53:47 GMT 2024
    - 8.3K bytes
    - Viewed (0)
Back to top