Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 45 for casa (0.22 sec)

  1. tensorflow/c/experimental/next_pluggable_device/c_api.cc

        TF_OpKernelContext* ctx) {
      auto* cc_ctx = reinterpret_cast<tensorflow::OpKernelContext*>(ctx);
      return reinterpret_cast<TF_CoordinationServiceAgent*>(
          cc_ctx->coordination_service_agent());
    }
    
    bool TF_CoordinationServiceIsInitialized(TF_CoordinationServiceAgent* agent) {
      if (agent == nullptr) return false;
      auto* cc_agent = reinterpret_cast<tsl::CoordinationServiceAgent*>(agent);
      return cc_agent->IsInitialized();
    C++
    - Registered: Tue Feb 27 12:39:08 GMT 2024
    - Last Modified: Tue Jan 09 00:52:04 GMT 2024
    - 13.9K bytes
    - Viewed (1)
  2. tensorflow/c/eager/custom_device_testutil.cc

      return TFE_TensorHandleDim(reinterpret_cast<LoggedTensor*>(data)->tensor,
                                 dim_index, status);
    }
    
    int LoggedTensorNumDims(void* data, TF_Status* status) {
      return TFE_TensorHandleNumDims(reinterpret_cast<LoggedTensor*>(data)->tensor,
                                     status);
    }
    
    void LoggedTensorDeallocator(void* data) {
      delete reinterpret_cast<LoggedTensor*>(data);
    }
    
    C++
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Wed Mar 03 20:47:31 GMT 2021
    - 8.3K bytes
    - Viewed (0)
  3. tensorflow/c/experimental/grappler/grappler_test.cc

        EXPECT_EQ(nodes_preserved.find(string(static_cast<const char*>(values[i]),
                                              lens[i])) != nodes_preserved.end(),
                  true);
      }
      TF_DeleteStatus(status);
    }
    
    TEST(TF_GrapplerItem, FetchNodes) {
      GrapplerItem item;
      item.fetch = std::vector<string>{"Conv", "BiasAdd"};
      TF_GrapplerItem* c_item = reinterpret_cast<TF_GrapplerItem*>(&item);
    
      int list_total_size = 0;
    C++
    - Registered: Tue Feb 27 12:39:08 GMT 2024
    - Last Modified: Thu Apr 13 22:30:58 GMT 2023
    - 11.6K bytes
    - Viewed (0)
  4. ci/official/utilities/extract_resultstore_links.py

        command_type = invocation_results.get('command_type')
    
        case_attrib = attrib.copy()
        if command_type:
          command_type = command_type.title()
          case_name = f'{command_type} invocation {invocation_id}'
        else:
          case_name = f' Invocation {invocation_id}'
        case_attrib.update({'name': case_name,
                            'status': 'run', 'result': 'completed'})
    
    Python
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Wed Nov 08 17:50:27 GMT 2023
    - 10.9K bytes
    - Viewed (0)
  5. tensorflow/c/eager/parallel_device/parallel_device.cc

    void ParallelTensorDeallocator(void* data) {
      delete reinterpret_cast<ParallelTensor*>(data);
    }
    
    // Used as an argument to TFE_NewCustomDeviceTensorHandle, for computing the
    // number of dimensions of a parallel tensor.
    int ParallelTensorNumDims(void* data, TF_Status* status) {
      const std::vector<int64_t>* shape;
      Status s = reinterpret_cast<ParallelTensor*>(data)->Shape(&shape);
      if (!s.ok()) {
    C++
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Wed Mar 29 22:05:31 GMT 2023
    - 18.3K bytes
    - Viewed (0)
  6. tensorflow/c/experimental/next_pluggable_device/tensor_pjrt_buffer_util_test.cc

      if (!status.ok()) {
        CHECK_OK(pjrt::SetPjrtApi(DEVICE_CPU, GetPjrtApi()));
      }
      auto pjrt_client = xla::GetCApiClient(DEVICE_CPU);
      CHECK_OK(pjrt_client.status());
      auto c_api_client = down_cast<xla::PjRtCApiClient*>(pjrt_client->get());
      std::vector<int32_t> data(1, 0);
      xla::Shape shape = xla::ShapeUtil::MakeShape(xla::S32, {1});
    
      auto buffer = c_api_client->pjrt_c_client()->client->BufferFromHostBuffer(
    C++
    - Registered: Tue Feb 27 12:39:08 GMT 2024
    - Last Modified: Mon Oct 30 19:20:20 GMT 2023
    - 7.2K bytes
    - Viewed (0)
  7. tensorflow/c/experimental/filesystem/plugins/gcs/gcs_filesystem_test.cc

      ASSERT_EQ(read, -1) << "Read: " << read;
      ASSERT_EQ(TF_GetCode(status_), TF_NOT_FOUND) << TF_Message(status_);
      TF_SetStatus(status_, TF_OK, "");
    
      auto gcs_file =
          static_cast<tf_gcs_filesystem::GCSFile*>(filesystem_->plugin_filesystem);
      ASSERT_TRUE(WriteToServer(filepath, 0, content_view.length(),
                                &gcs_file->gcs_client, status_));
    
    C++
    - Registered: Tue Apr 23 12:39:09 GMT 2024
    - Last Modified: Mon Aug 31 12:04:23 GMT 2020
    - 24.9K bytes
    - Viewed (0)
  8. .github/ISSUE_TEMPLATE/tflite-op-request.md

    - TensorFlow version (or github SHA if from source):
    
    
    **Provide the text output from tflite_convert**
    
    ```
    # Copy and paste here
    ```
    
    **Standalone code to reproduce the issue** 
    Provide a reproducible test case that is the bare minimum necessary to generate
    the problem. If possible, please share a link to Colab/Jupyter/any notebook.
    
    Also, please include a link to a GraphDef or the model if possible.
    
    **Any other info / logs**
    
    Plain Text
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Wed Jun 15 03:35:58 GMT 2022
    - 879 bytes
    - Viewed (0)
  9. tensorflow/c/eager/parallel_device/parallel_device_testlib.h

      ASSERT_EQ(TF_GetCode(status.get()), TF_OK) << TF_Message(status.get());
      ASSERT_EQ(TF_TensorType(actual_value.get()),
                static_cast<TF_DataType>(DataTypeToEnum<value_type>().value));
      EXPECT_EQ(expected_value,
                *static_cast<value_type*>(TF_TensorData(actual_value.get())));
    }
    
    template <std::size_t num_devices>
    void RegisterParallelDevice(
    C
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Tue Feb 09 01:12:35 GMT 2021
    - 6.9K bytes
    - Viewed (0)
  10. ci/official/envs/macos_x86

    TFCI_MACOS_TWINE_INSTALL_ENABLE=1
    TFCI_OUTPUT_DIR=build_output
    TFCI_WHL_BAZEL_TEST_ENABLE=1
    TFCI_WHL_SIZE_LIMIT=255M
    TFCI_WHL_SIZE_LIMIT_ENABLE=1
    
    # 3.11 is the system python on our images
    case $TFCI_PYTHON_VERSION in
    3.11)
      TFCI_MACOS_PYENV_INSTALL_ENABLE=0
      TFCI_MACOS_UPGRADE_PYENV_ENABLE=0
      ;;
    *)
      TFCI_MACOS_PYENV_INSTALL_ENABLE=1
      TFCI_MACOS_UPGRADE_PYENV_ENABLE=1
      ;;
    Plain Text
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Tue Jan 30 19:47:28 GMT 2024
    - 1.6K bytes
    - Viewed (0)
Back to top