Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 12 for can (0.13 sec)

  1. tensorflow/c/eager/c_api_cluster_test.cc

    TEST(CAPI, TestRemoteExecuteUpdateServerDefResourceAccessAsync) {
      TestRemoteExecuteUpdateServerDefResourceAccess(true);
    }
    
    void TestRemoteExecuteUpdateServerDefWithFailures(bool async) {
      // Fail fast on GetStatus requests so we can get errors instead of timeout
      // when updating cluster with non-exsitent worker
      tensorflow::setenv("GRPC_FAIL_FAST", "TRUE", /*overwrite=*/1);
    
      tensorflow::ServerDef server_def = GetServerDef(2);
    C++
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Fri Apr 14 10:03:59 GMT 2023
    - 19.3K bytes
    - Viewed (0)
  2. tensorflow/c/c_api_experimental.cc

      // TODO(b/113217601): This is needed for EagerContext::runner_ to use a
      // threadpool, so that we avoid the possibility of running the runner_ in the
      // threadpool of GPU event mgr, as that can trigger more callbacks to be
      // scheduled on that same threadpool, causing a deadlock in cases where the
      // caller of event_mgr->ThenExecute() blocks on the completion of the callback
    C++
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Mon Apr 15 03:35:10 GMT 2024
    - 29.4K bytes
    - Viewed (0)
  3. tensorflow/c/eager/parallel_device/parallel_device_lib_test.cc

          ParallelTensor::FromTensorHandles(parallel_device,
                                            std::move(mixed_handles), status.get());
      ASSERT_TRUE(TF_GetCode(status.get()) == TF_OK);
      // Can't take the shape of a parallel tensor with varying numbers of axes, but
      // running operations on them is OK.
      TF_ASSERT_OK(unknown_length_vector->Shape(&shape));
      EXPECT_THAT(*shape, ElementsAre(-1));
    C++
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Thu Jul 08 23:47:35 GMT 2021
    - 15.3K bytes
    - Viewed (0)
  4. tensorflow/c/eager/dlpack.cc

      int64_t expected_stride = 1;
      for (int i = ndim - 1; i >= 0; --i) {
        // Empty tensors are always compact regardless of strides.
        if (shape_arr[i] == 0) return true;
        // Note that dimensions with size=1 can have any stride.
        if (shape_arr[i] != 1 && stride_arr[i] != expected_stride) {
          valid = false;
        }
        expected_stride *= shape_arr[i];
      }
      return valid;
    }
    }  // namespace
    
    C++
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Thu Feb 15 09:49:45 GMT 2024
    - 12.8K bytes
    - Viewed (0)
  5. tensorflow/c/eager/custom_device_test.cc

      num_retvals = 1;
      executed = false;
      TFE_Execute(matmul.get(), &retval, &num_retvals, status.get());
      EXPECT_FALSE(executed);
      ASSERT_FALSE(TF_GetCode(status.get()) == TF_OK);
    
      // Custom devices can refuse to do type-based dispatch (as hcustom1 is
      // configured to do)
      matmul.reset(MatMulOp(context.get(), hcustom1.get(), hcpu.get()));
      num_retvals = 1;
      executed = false;
    C++
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Thu Aug 27 23:39:24 GMT 2020
    - 18.4K bytes
    - Viewed (0)
  6. tensorflow/c/experimental/filesystem/plugins/gcs/ram_file_block_cache.cc

        // Copy the relevant portion of the block into the result buffer.
        const auto& data = block->data;
        if (offset >= pos + data.size()) {
          // The requested offset is at or beyond the end of the file. This can
          // happen if `offset` is not block-aligned, and the read returns the last
          // block in the file, which does not extend all the way out to `offset`.
          std::stringstream os;
    C++
    - Registered: Tue Apr 23 12:39:09 GMT 2024
    - Last Modified: Thu Jul 16 01:39:09 GMT 2020
    - 11.1K bytes
    - Viewed (0)
  7. tensorflow/c/eager/parallel_device/parallel_device_lib.cc

      CancellationManager* cancellation_manager_ TF_GUARDED_BY(execution_mutex_);
      //   Outputs:
      std::vector<TensorHandlePtr> op_outputs_ TF_GUARDED_BY(execution_mutex_);
      // TF_Status is an incomplete type and so can't be stack allocated. To avoid
      // unnecessary allocations each Execute call, we keep one heap-allocated
      // version for the thread.
      StatusPtr status_ TF_GUARDED_BY(execution_mutex_);
    
      const std::string device_;
    C++
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Fri Feb 09 07:47:20 GMT 2024
    - 25.4K bytes
    - Viewed (1)
  8. tensorflow/c/eager/c_api_distributed_test.cc

        }
        return tensorflow::errors::Internal("Graph pass runs for more than once!");
      }
    
     private:
      bool first_call_ = true;
    };
    
    // After the graph pass is registered, it takes effect globally and can affect
    // other test cases. Define a static variable to switch it on and off.
    bool GraphErrorInjectionPass::enabled_ = false;
    
    // Test to ensure that a registered graph optimization pass is only executed
    C++
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Thu Feb 15 09:49:45 GMT 2024
    - 23.5K bytes
    - Viewed (0)
  9. tensorflow/c/experimental/filesystem/modular_filesystem_registration.cc

    //     TensorFlow will know to not call these on behalf of users;
    //   * increased security as plugins will not be able to alter function table
    //     after loading up. Thus, malicious plugins can't alter functionality to
    //     probe for gadgets inside core TensorFlow. We can even protect the area
    //     of memory where the copies reside to not allow any more writes to it
    //     after all copies are created.
    template <typename T>
    C++
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Tue Mar 07 22:08:43 GMT 2023
    - 12.8K bytes
    - Viewed (0)
  10. tensorflow/c/experimental/filesystem/plugins/gcs/ram_file_block_cache_test.cc

                  ReadCache(&cache, "a", i * block_size, block_size, &out));
              std::vector<char> x(block_size, 'x');
              EXPECT_EQ(out, x);
            }));
      }
      // The `threads` destructor blocks until the threads can be joined, once their
      // respective reads finish (which happens once they are all concurrently being
      // executed, or 10 seconds have passed).
    }
    
    TEST(RamFileBlockCacheTest, CoalesceConcurrentReads) {
    C++
    - Registered: Tue Apr 23 12:39:09 GMT 2024
    - Last Modified: Fri Oct 15 03:16:57 GMT 2021
    - 23.2K bytes
    - Viewed (0)
Back to top