Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 4 of 4 for asynchronously (8.93 sec)

  1. tensorflow/c/eager/parallel_device/parallel_device_lib.h

      // re-use a thread.
      std::vector<std::unique_ptr<DeviceThread>> device_threads_;
      // A cancellation manager to use if the caller does not provide one. When ops
      // are executed asynchronously this must outlive the queued op, so it can't be
      // function-local to Execute.
      mutable std::unique_ptr<CancellationManager> default_cancellation_manager_;
    };
    
    C
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Tue Apr 25 15:21:13 GMT 2023
    - 12.9K bytes
    - Viewed (0)
  2. tensorflow/c/eager/parallel_device/parallel_device_test.cc

      ExpectScalarEq<float>(result_components[1].get(), 3.);
    }
    
    TEST(PARALLEL_DEVICE, TestCollectiveSync) { TestCollective(/*async=*/false); }
    
    // Note that ops on the parallel device currently don't execute
    // asynchronously. The test is just that we don't get deadlocks.
    TEST(PARALLEL_DEVICE, TestCollectiveAsync) { TestCollective(/*async=*/true); }
    
    void RegisterCollectiveMulFunction(TFE_Context* context,
    C++
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Thu Jul 08 23:47:35 GMT 2021
    - 29.3K bytes
    - Viewed (1)
  3. RELEASE.md

    *   Introduce TFDecorator.
    *   Added an Mfcc op for speech feature generation.
    *   Improved DirectSession::Run() overhead and error checking. Feeding a value
        of the wrong type will now synchronously raise an INVALID_ARGUMENT error
        instead of asynchronously raising an INTERNAL error. Code that depends on
        the (undefined) behavior when feeding a tensor of the wrong type may need to
        be updated.
    Plain Text
    - Registered: Tue May 07 12:40:20 GMT 2024
    - Last Modified: Mon Apr 29 19:17:57 GMT 2024
    - 727.7K bytes
    - Viewed (8)
  4. tensorflow/c/experimental/next_pluggable_device/tensor_pjrt_buffer_util_test.cc

      auto allocator = std::make_unique<AsyncValueAllocator>();
      tensorflow::Tensor tensor(allocator.get(), DT_FLOAT, {1});
      TF_ASSERT_OK_AND_ASSIGN(
          auto pjrt_client,
          xla::GetTfrtCpuClient(/*asynchronous=*/true, /*cpu_device_count=*/1));
      std::vector<int32_t> data(1, 0);
      xla::Shape shape = xla::ShapeUtil::MakeShape(xla::S32, {1});
      TF_ASSERT_OK_AND_ASSIGN(
          auto buffer,
          pjrt_client->BufferFromHostBuffer(
    C++
    - Registered: Tue Feb 27 12:39:08 GMT 2024
    - Last Modified: Mon Oct 30 19:20:20 GMT 2023
    - 7.2K bytes
    - Viewed (0)
Back to top