Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 15 for removed (0.17 sec)

  1. tensorflow/c/eager/parallel_device/parallel_device_remote_test.cc

    A. Unique TensorFlower <******@****.***> 1682633218 -0700
    C++
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Thu Apr 27 22:09:57 GMT 2023
    - 6.7K bytes
    - Viewed (0)
  2. tensorflow/c/experimental/gradients/nn_grad.cc

         *
         *    dA = U
         *    dbias = reduceSum(U, dims = channel_dim)
         */
    
        AbstractTensorHandle* upstream_grad = grad_outputs[0];
        DCHECK(upstream_grad);
    
        // Recover data format from forward pass for gradient.
        std::string data_format;
        TF_RETURN_IF_ERROR(forward_attrs_.Get("data_format", &data_format));
    
        // Grad for A
        grad_inputs[0] = upstream_grad;
    C++
    - Registered: Tue Mar 26 12:39:09 GMT 2024
    - Last Modified: Wed Feb 28 13:53:47 GMT 2024
    - 5.7K bytes
    - Viewed (0)
  3. tensorflow/c/c_api.cc

    namespace {
    
    #if !defined(IS_MOBILE_PLATFORM) && !defined(IS_SLIM_BUILD)
    
    // Creates a placeholder representing an input to the cond or body graph.
    // TODO(skyewm): remove these from final graph
    bool CreateInput(const TF_Output& parent_input, TF_Graph* g, const char* name,
                     TF_Output* input, TF_Status* status) {
    C++
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Mon Apr 15 03:35:10 GMT 2024
    - 102.3K bytes
    - Viewed (0)
  4. tensorflow/c/experimental/gradients/math_grad_test.cc

          immediate_execution_ctx_.get(), {x.get()}, UseFunction()));
    }
    
    TEST_P(CppGradients, TestMatMulGrad) {
      // TODO(vnvo2409): Figure out why `gradient_checker` does not work very
      // well with `MatMul` and remove `TestMatMul*` in
      // `mnist_gradients_test` when done.
      GTEST_SKIP();
    
      float A_vals[] = {1.0f, 2.0f, 3.0f, 4.0f, 5.0f, 6.0f, 7.0f, 8.0f, 9.0f};
      int64_t A_dims[] = {3, 3};
      AbstractTensorHandlePtr A;
      {
    C++
    - Registered: Tue Mar 26 12:39:09 GMT 2024
    - Last Modified: Thu Apr 13 17:32:14 GMT 2023
    - 16.3K bytes
    - Viewed (0)
  5. tensorflow/c/eager/c_api_cluster_test.cc

      // Check that copying it to the old remote device (named localhost) fails.
      TFE_TensorHandleCopyToDevice(h0_task0_new, ctx, remote_device_name, status);
      EXPECT_NE(TF_OK, TF_GetCode(status)) << TF_Message(status);
    
      // Copying and executing on the new remote device works.
      const char new_remote_device_name[] =
          "/job:worker/replica:0/task:1/device:CPU:0";
    C++
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Fri Apr 14 10:03:59 GMT 2023
    - 19.3K bytes
    - Viewed (0)
  6. tensorflow/c/eager/immediate_execution_tensor_handle.cc

        std::string& summary) const {
      Status status;
      AbstractTensorPtr resolved(
          // TODO(allenl): Resolve should be const, and the caches that get updated
          // marked mutable.
          const_cast<ImmediateExecutionTensorHandle*>(this)->Resolve(&status));
      if (!status.ok()) {
        return status;
      }
      summary = resolved->SummarizeValue();
      return absl::OkStatus();
    }
    
    C++
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Thu Feb 15 09:49:45 GMT 2024
    - 2.1K bytes
    - Viewed (0)
  7. tensorflow/c/eager/gradients.cc

      // and not for correctness. The only downside of keeping this 1 seems to be
      // that the gradient accumulation is unbounded and we will never
      // aggressively aggregate accumulated gradients to recover memory.
      // Revisit and fix.
      return 1;
    }
    
    // Consumes references to the tensors in the gradient_tensors list and returns
    // a tensor with the result.
    AbstractTensorHandle* TapeVSpace::AggregateGradients(
    C++
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Thu Feb 15 09:49:45 GMT 2024
    - 19.3K bytes
    - Viewed (0)
  8. tensorflow/c/eager/c_api_distributed_test.cc

      EXPECT_EQ(TF_OK, TF_GetCode(status)) << TF_Message(status);
    
      // Pack 3 variable handles into one TFE_TensorHandle.
      // When remote is false, function device is placed on task0. Handle types are
      // REMOTE, REMOTE, LOCAL on task0. When remote is true, function device is
      // placed on task1, Handle types are LOCAL, REMOTE, LOCAL on task1.
      int num_replicas = 3;
      std::vector<TFE_TensorHandle*> handles = {h0, h1, h2};
    C++
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Thu Feb 15 09:49:45 GMT 2024
    - 23.5K bytes
    - Viewed (0)
  9. tensorflow/c/eager/parallel_device/parallel_device.cc

        const TFE_OpAttrs* attributes, int expected_max_outputs,
        TF_Status* status) {
      absl::optional<std::vector<MaybeParallelTensorOwned>> result;
      // TODO(allenl): We should remove "TPU" from these op names at the very least,
      // or consider other ways of packing/unpacking parallel tensors.
      if (operation_name == std::string("TPUReplicatedInput")) {
    C++
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Wed Mar 29 22:05:31 GMT 2023
    - 18.3K bytes
    - Viewed (0)
  10. tensorflow/c/eager/parallel_device/parallel_device_lib.cc

        per_device_output_tensors.push_back(device_thread->Join(status));
        // We will run every Join even if there are bad statuses in case the user
        // wants to recover and continue running ops on the parallel device (which
        // would otherwise deadlock).
        if (TF_GetCode(status) != TF_OK &&
            (first_bad_status == nullptr
    C++
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Fri Feb 09 07:47:20 GMT 2024
    - 25.4K bytes
    - Viewed (1)
Back to top