Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 5 of 5 for device_index (0.23 sec)

  1. tensorflow/c/eager/parallel_device/parallel_device_lib.cc

          default_cancellation_manager_(absl::make_unique<CancellationManager>()) {
      device_threads_.reserve(devices.size());
      for (int device_index = 0; device_index < devices.size(); ++device_index) {
        device_threads_.emplace_back(new DeviceThread(
            devices[device_index].c_str(), is_async, in_flight_nodes_limit));
      }
    }
    
    // Necessary for a unique_ptr to a forward-declared type.
    C++
    - Registered: Tue Apr 23 12:39:09 GMT 2024
    - Last Modified: Fri Feb 09 07:47:20 GMT 2024
    - 25.4K bytes
    - Viewed (1)
  2. tensorflow/c/eager/parallel_device/parallel_device_lib.h

        return nullptr;
      }
      TF_DataType datatype_enum(
          static_cast<TF_DataType>(DataTypeToEnum<DataType>().value));
      for (int device_index = 0; device_index < num_underlying_devices();
           ++device_index) {
        auto device_value = absl::make_unique<DataType>();
        *device_value = values[device_index];
        std::unique_ptr<TF_Tensor, decltype(&TF_DeleteTensor)> tensor(
            TF_NewTensor(
    C
    - Registered: Tue Apr 23 12:39:09 GMT 2024
    - Last Modified: Tue Apr 25 15:21:13 GMT 2023
    - 12.9K bytes
    - Viewed (0)
  3. tensorflow/c/eager/parallel_device/parallel_device.cc

      std::vector<std::string> underlying_devices_vector;
      underlying_devices_vector.reserve(num_underlying_devices);
      for (int device_index = 0; device_index < num_underlying_devices;
           ++device_index) {
        underlying_devices_vector.push_back(underlying_devices[device_index]);
      }
      std::unique_ptr<ParallelDevice> parallel_device(
          new ParallelDevice(underlying_devices_vector));
      *device_info =
    C++
    - Registered: Tue Apr 23 12:39:09 GMT 2024
    - Last Modified: Wed Mar 29 22:05:31 GMT 2023
    - 18.3K bytes
    - Viewed (0)
  4. tensorflow/c/eager/parallel_device/parallel_device_test.cc

      ASSERT_EQ(TF_GetCode(status.get()), TF_OK) << TF_Message(status.get());
      bool has_tpu = false;
      for (int device_index = 0; device_index < TF_DeviceListCount(devices.get());
           ++device_index) {
        std::string device_type =
            TF_DeviceListType(devices.get(), device_index, status.get());
        ASSERT_EQ(TF_GetCode(status.get()), TF_OK) << TF_Message(status.get());
        if (device_type == "TPU") {
    C++
    - Registered: Tue Apr 23 12:39:09 GMT 2024
    - Last Modified: Thu Jul 08 23:47:35 GMT 2021
    - 29.3K bytes
    - Viewed (1)
  5. RELEASE.md

            treated as tensors, so if you still want them to be treated that way,
            you need to wrap them with `tf.convert_to_tensor`.
        *   No lowering on gradient case op when input is `DeviceIndex` op.
        *   Extend the ragged version of `tf.gather` to support `batch_dims` and
            `axis` args.
        *   Update `tf.map_fn` to support RaggedTensors and SparseTensors.
    Plain Text
    - Registered: Tue Apr 23 12:39:09 GMT 2024
    - Last Modified: Wed Apr 03 20:27:38 GMT 2024
    - 727.4K bytes
    - Viewed (8)
Back to top