- Sort Score
- Result 10 results
- Languages All
Results 1 - 7 of 7 for y_shapes (0.07 sec)
-
tensorflow/c/eager/c_api_unified_experimental_graph.cc
std::vector<PartialTensorShape> shapes; shapes.reserve(num_values); for (int i = 0; i < num_values; ++i) { if (num_dims[i] < 0) { shapes.emplace_back(); } else { shapes.emplace_back(ArraySlice<int64_t>( reinterpret_cast<const int64_t*>(dims[i]), num_dims[i])); } } op_->node_builder.Attr(attr_name, shapes); return absl::OkStatus(); }
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Sat Oct 12 05:11:17 UTC 2024 - 15.7K bytes - Viewed (0) -
tensorflow/c/c_api.cc
std::vector<PartialTensorShape> shapes; shapes.reserve(num_shapes); for (int i = 0; i < num_shapes; ++i) { if (num_dims[i] < 0) { shapes.emplace_back(); } else { shapes.emplace_back(ArraySlice<int64_t>( reinterpret_cast<const int64_t*>(dims[i]), num_dims[i])); } } desc->node_builder.Attr(attr_name, shapes); }
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Sat Oct 12 16:27:48 UTC 2024 - 102.3K bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device_lib.h
// run and have results. If is_async=true it means that all of the // device-specific executors have scheduled the op. // // Accepts inferred shapes for outputs (`expected_output_shapes`), which if // fully defined will avoid querying the shapes of the underlying // TensorHandles when ParallelTensor::Shape is called. This allows async // computation to continue without blocking. //
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Mon Oct 21 04:14:14 UTC 2024 - 13.1K bytes - Viewed (0) -
tensorflow/c/eager/immediate_execution_tensor_handle.h
// // -1 indicates an unknown axis length; this is unreachable for most standard // ImmediateExecutionTensorHandles, but comes up for example when computing // the shape of a parallel tensor with component shapes differing across // devices. virtual absl::Status Dim(int dim_index, int64_t* dim) const = 0; // Returns the device which created the handle. virtual const char* DeviceName(absl::Status* status) const = 0;
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Sat Oct 12 05:11:17 UTC 2024 - 4.3K bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device_lib.cc
if (!status->status.ok()) { return nullptr; } TF_DataType dtype = TFE_TensorHandleDataType(components[0].get()); // Verify that the TensorHandle's shape and dtype match all of the component // shapes and dtypes. for (TensorHandlePtr& component : components) { TFE_TensorHandleGetStatus(component.get(), status); if (!status->status.ok()) { return nullptr; }
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Mon Oct 21 04:14:14 UTC 2024 - 25.9K bytes - Viewed (0) -
RELEASE.md
expressions. * Fix spurious background colors in some text terminals. * `tf.contrib`: * Add meta-distribution BatchReshape which reshapes batch dimensions. * `tf.contrib.layers.recompute_grad` works for explicit gradient checkpointing on TPU. * Add `tf.contrib.framework.argsort`.
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Tue Oct 22 14:33:53 UTC 2024 - 735.3K bytes - Viewed (0) -
tensorflow/c/c_api_test.cc
// be inferred. TF_Status* status = TF_NewStatus(); TF_Graph* graph = TF_NewGraph(); // Create this failure by trying to add two nodes with incompatible shapes // (A tensor with shape [2] and a tensor with shape [3] cannot be added). const char data[] = {1, 2, 3}; const int64_t vec2_dims[] = {2}; unique_tensor_ptr vec2_tensor(
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Sat Oct 12 16:27:48 UTC 2024 - 97K bytes - Viewed (0)