- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 20 for Rust (0.18 sec)
-
tensorflow/c/eager/c_api_unified_experimental.cc
TracingContext* tracing_ctx = dyn_cast<TracingContext>(unwrap(func)); if (!tracing_ctx) { tsl::Set_TF_Status_from_Status( s, tensorflow::errors::InvalidArgument( "TF_AddFunctionParameter must be called on a TracingContext.")); return nullptr; } tensorflow::PartialTensorShape partial_shape; if (shape.num_dims != -1) { DCHECK(shape.dim_sizes != nullptr);
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 9K bytes - Viewed (0) -
tensorflow/c/c_api.cc
opts->tensor_id_data.push_back(src_name); const string& src_name_str = opts->tensor_id_data.back(); // We don't need to store dst's name in tensor_id_data, since `dst` must // outlive the ImportGraphDef call. opts->opts.input_map[TensorId(src_name_str, src_index)] = ToTensorId(dst); } void TF_ImportGraphDefOptionsRemapControlDependency(
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 03:35:10 GMT 2024 - 102.3K bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device.cc
// copies/broadcasts. Other implicit copies may be supported eventually, // but need special handling for gradients (gradient of copy-on is not // just copy-off but includes a sum) and consideration of performance. // // TODO(allenl): There may be smarter ways to do this copy in some // cases, i.e. with a collective broadcast. We'll need to be careful
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Wed Mar 29 22:05:31 GMT 2023 - 18.3K bytes - Viewed (0) -
tensorflow/c/c_api_function.cc
// artificial restriction and require that when num_opers=-1, such // nodes must have a single output. if (node->num_outputs() != 1) { return InvalidArgument( "When `num_opers` is set to -1, nodes referenced in `inputs` " "must have a single output. Node ", node->name(), " has ", node->num_outputs(),
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 03:35:10 GMT 2024 - 13.6K bytes - Viewed (2) -
tensorflow/c/eager/parallel_device/parallel_device_test.cc
TensorHandlePtr cpu_value(FloatTensorHandle(3., status.get())); ASSERT_EQ(TF_GetCode(status.get()), TF_OK) << TF_Message(status.get()); // Copying on to a parallel device must be explicit. TensorHandlePtr failed_copy_on_result(TFE_TensorHandleCopyToDevice( cpu_value.get(), context.get(), device_name, status.get())); EXPECT_EQ(TF_GetCode(status.get()), TF_UNIMPLEMENTED);
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Jul 08 23:47:35 GMT 2021 - 29.3K bytes - Viewed (1) -
tensorflow/c/experimental/grappler/grappler_test.cc
tensorflow::Status status = InitGraphPlugin(plugin_init); ASSERT_EQ(status.code(), tensorflow::error::FAILED_PRECONDITION); ASSERT_EQ( status.message(), "'device_type' field in TP_OptimizerRegistrationParams must be set."); } TEST(Grappler, OptimizeFuncNotSet) { auto plugin_init = [](TP_OptimizerRegistrationParams* const params, TF_Status* const status) -> void { TF_SetStatus(status, TF_OK, "");
C++ - Registered: Tue Feb 27 12:39:08 GMT 2024 - Last Modified: Thu Apr 13 22:30:58 GMT 2023 - 11.6K bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device_lib.cc
// Requests that the worker thread execute the specified operation. Blocks // until the previously pending operation (a StartExecute without a Join) has // finished, if any. // // `cancellation_manager` must live until after `Join` finishes and pending // `is_async` operations finish. In addition to allowing the caller to cancel // the operation, its `StartCancel` method will be called if op execution
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri Feb 09 07:47:20 GMT 2024 - 25.4K bytes - Viewed (1) -
tensorflow/c/eager/custom_device_testutil.cc
LoggingDevice* dev = reinterpret_cast<LoggingDevice*>(device_info); if (dev->strict_scope_placement && *requested_placement == '\0') { TF_SetStatus(s, TF_INTERNAL, "Ops must be placed on the device explicitly, or their inputs " "first copied to other devices."); return; } TFE_Context* context = TFE_OpGetContext(original_op, s); if (TF_GetCode(s) != TF_OK) return;
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Wed Mar 03 20:47:31 GMT 2021 - 8.3K bytes - Viewed (0) -
tensorflow/c/experimental/filesystem/plugins/gcs/gcs_filesystem_test.cc
ASSERT_TF_OK(status_); ASSERT_EQ(bucket, "bucket"); ParseGCSPath("bucket/path/to/object", false, &bucket, &object, status_); ASSERT_EQ(TF_GetCode(status_), TF_INVALID_ARGUMENT); // bucket name must end with "/" ParseGCSPath("gs://bucket", true, &bucket, &object, status_); ASSERT_EQ(TF_GetCode(status_), TF_INVALID_ARGUMENT); ParseGCSPath("gs://bucket/", false, &bucket, &object, status_);
C++ - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Mon Aug 31 12:04:23 GMT 2020 - 24.9K bytes - Viewed (0) -
tensorflow/c/c_api_experimental.cc
memcpy(ret, debug_str.c_str(), *len + 1); return ret; } // On success, returns a set of TF_Function instances from `text_proto` of // GraphDef type. These functions must be deleted by calling TF_DeleteFunction. // // If `mutate_proto_func` is non-NULL, run it over each FunctionDef proto, // before creating a TF_Function out of the possibly mutated proto.
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 03:35:10 GMT 2024 - 29.4K bytes - Viewed (0)