Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 4 of 4 for what (0.27 sec)

  1. tensorflow/c/experimental/gradients/array_grad.cc

        for (int i = 0; i < grad_outputs.size(); i++) {
          auto grad_input = grad_outputs[i];
          // TODO(srbs): Should we add a copy contructor to AbstractTensorHandle
          // that takes care of this similar to `Tensor`?
          if (grad_input) {
            grad_input->Ref();
          }
          grad_inputs[i] = grad_input;
        }
        return absl::OkStatus();
      }
    C++
    - Registered: Tue Apr 09 12:39:09 GMT 2024
    - Last Modified: Wed Feb 28 13:53:47 GMT 2024
    - 1.6K bytes
    - Viewed (0)
  2. tensorflow/c/eager/immediate_execution_tensor_handle.cc

    }
    
    Status ImmediateExecutionTensorHandle::SummarizeValue(
        std::string& summary) const {
      Status status;
      AbstractTensorPtr resolved(
          // TODO(allenl): Resolve should be const, and the caches that get updated
          // marked mutable.
          const_cast<ImmediateExecutionTensorHandle*>(this)->Resolve(&status));
      if (!status.ok()) {
        return status;
      }
      summary = resolved->SummarizeValue();
    C++
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Thu Feb 15 09:49:45 GMT 2024
    - 2.1K bytes
    - Viewed (0)
  3. tensorflow/c/eager/unified_api_test.cc

      }
    
     public:
      bool UseMlir() const { return strcmp(std::get<0>(GetParam()), "mlir") == 0; }
      bool UseFunction() const { return std::get<2>(GetParam()); }
    };
    
    // Checks that inputs[0] is a scalar.
    Status TestScalarShape(AbstractContext* ctx,
                           absl::Span<AbstractTensorHandle* const> inputs,
                           absl::Span<AbstractTensorHandle*> outputs) {
    C++
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Tue Feb 27 13:57:45 GMT 2024
    - 6.7K bytes
    - Viewed (0)
  4. tensorflow/c/experimental/gradients/nn_grad_test.cc

      TF_RETURN_IF_ERROR(ops::SparseSoftmaxCrossEntropyWithLogits(
          ctx, inputs[0], inputs[1], &loss, &backprop,
          "SparseSoftmaxCrossEntropyWithLogits"));
      // `gradient_checker` only works with model that returns only 1 tensor.
      // Although, `ops::SparseSoftmaxCrossEntropyWithLogits` returns 2 tensors, the
      // second tensor isn't needed for computing gradient so we could safely drop
      // it.
      outputs[0] = loss;
    C++
    - Registered: Tue Mar 26 12:39:09 GMT 2024
    - Last Modified: Wed Feb 28 13:53:47 GMT 2024
    - 8.3K bytes
    - Viewed (0)
Back to top