Search Options

Results per page
Sort
Preferred Languages
Advance

Results 31 - 33 of 33 for Zach (0.13 sec)

  1. tensorflow/c/c_api.h

    // of fetched mappings is returned in `num_missing_unused_input_mappings`. The
    // array of each mapping's source node name is returned in `src_names`, and the
    // array of each mapping's source index is returned in `src_indexes`.
    //
    // `*src_names`, `*src_indexes`, and the memory backing each string in
    // `src_names` are owned by and have the lifetime of `results`.
    C
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Thu Oct 26 21:08:15 GMT 2023
    - 82.3K bytes
    - Viewed (3)
  2. tensorflow/c/experimental/filesystem/plugins/gcs/gcs_filesystem_test.cc

        status_ = TF_NewStatus();
        filesystem_ = new TF_Filesystem;
        filesystem_->plugin_filesystem = nullptr;
        // Because different tests requires different setup for filesystem. We
        // initialize filesystem in each testcase.
      }
      void TearDown() override {
        TF_DeleteStatus(status_);
        if (filesystem_->plugin_filesystem != nullptr)
          tf_gcs_filesystem::Cleanup(filesystem_);
        delete filesystem_;
      }
    
    C++
    - Registered: Tue Apr 23 12:39:09 GMT 2024
    - Last Modified: Mon Aug 31 12:04:23 GMT 2020
    - 24.9K bytes
    - Viewed (0)
  3. tensorflow/c/eager/parallel_device/parallel_device_test.cc

      TensorHandlePtr parallel_value = CreatePerDeviceValues(
          context.get(), components, device_name, status.get());
      ASSERT_EQ(TF_GetCode(status.get()), TF_OK) << TF_Message(status.get());
    
      // Run a collective sum, so each component should now be the same.
      TensorHandlePtr reduced(
          CollectiveSum(context.get(), parallel_value.get(), 2, status.get()));
      ASSERT_EQ(TF_GetCode(status.get()), TF_OK) << TF_Message(status.get());
    
    C++
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Thu Jul 08 23:47:35 GMT 2021
    - 29.3K bytes
    - Viewed (1)
Back to top