- Sort Score
- Num 10 results
- Language All
Results 11 - 20 of 89 for shape_2 (0.05 seconds)
-
tensorflow/c/eager/c_api_experimental.h
int version = TFE_CUSTOM_DEVICE_VERSION; // Computes the rank of the tensor handle. // // Shapes are specified via callbacks because retrieving the shape of a tensor // is a blocking operation for async eager; custom devices should avoid // retrieving shapes of tensors they wrap until the custom device tensor's // shape is explicitly requested where possible. int (*num_dims)(void* data, TF_Status* status);
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Wed Feb 21 22:37:46 GMT 2024 - 39.5K bytes - Click Count (1) -
tensorflow/c/eager/c_api_experimental_test.cc
ASSERT_TRUE(TF_GetCode(status.get()) == TF_OK) << TF_Message(status.get()); TFE_Op* shape_op = ShapeOp(ctx, hgpu); TFE_OpSetDevice(shape_op, gpu_device_name.c_str(), status.get()); ASSERT_TRUE(TF_GetCode(status.get()) == TF_OK) << TF_Message(status.get()); TFE_TensorHandle* retvals[1]; int num_retvals = 1; TFE_Execute(shape_op, &retvals[0], &num_retvals, status.get());
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Thu Oct 09 05:56:18 GMT 2025 - 31.5K bytes - Click Count (0) -
tensorflow/c/eager/c_api_test.cc
ASSERT_TRUE(TF_GetCode(status.get()) == TF_OK) << TF_Message(status.get()); TFE_Op* shape_op = ShapeOp(ctx, hgpu); TFE_OpSetDevice(shape_op, gpu_device_name.c_str(), status.get()); ASSERT_TRUE(TF_GetCode(status.get()) == TF_OK) << TF_Message(status.get()); TFE_TensorHandle* retvals[1]; int num_retvals = 1; TFE_Execute(shape_op, &retvals[0], &num_retvals, status.get());
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Thu Oct 09 05:56:18 GMT 2025 - 94.6K bytes - Click Count (0) -
tensorflow/c/eager/c_api_test_util.cc
TF_DeleteStatus(status); TFE_OpSetAttrType(op, "T", TFE_TensorHandleDataType(a)); return op; } TFE_Op* ShapeOp(TFE_Context* ctx, TFE_TensorHandle* a) { TF_Status* status = TF_NewStatus(); TFE_Op* op = TFE_NewOp(ctx, "Shape", status); CHECK_EQ(TF_OK, TF_GetCode(status)) << TF_Message(status); TFE_OpAddInput(op, a, status);
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Thu Oct 09 05:56:18 GMT 2025 - 23.4K bytes - Click Count (0) -
src/test/java/jcifs/smb1/smb1/NetShareEnumResponseTest.java
SmbShareInfo share2 = (SmbShareInfo) response.results[1]; assertEquals("IPC$", share2.getName().trim(), "Share 2 name should be correct."); assertEquals(3, share2.type, "Share 2 raw type should be 3."); assertEquals("Inter-Process Communication", share2.remark.trim(), "Share 2 remark should be correct."); } /**Created: Sun Apr 05 00:10:12 GMT 2026 - Last Modified: Thu Aug 14 05:31:44 GMT 2025 - 5.4K bytes - Click Count (0) -
tensorflow/c/eager/c_api_debug.cc
std::vector<int64_t> shape; int rank = -1; *status = handle.NumDims(&rank); if (!status->ok()) { return shape; } shape.reserve(rank); for (int i = 0; i < rank; ++i) { int64_t dim; *status = handle.Dim(i, &dim); if (!status->ok()) { return shape; } shape.push_back(dim); } return shape; } } // namespace extern "C" {
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Sat Oct 12 05:11:17 GMT 2024 - 2.5K bytes - Click Count (0) -
tensorflow/c/eager/dlpack_test.cc
dltensor_in->device = {kDLCPU, 0}; dltensor_in->ndim = static_cast<int32_t>(shape.size()); dltensor_in->dtype = {kDLFloat, 32, 1}; dltensor_in->shape = shape.data(); dltensor_in->strides = strides.data(); TFE_TensorHandle* handle = TFE_HandleFromDLPack(&dlm_in, status, ctx); ASSERT_NE(handle, nullptr) << TF_Message(status) << " (shape=[" << absl::StrJoin(shape, ",") << "], strides=[" << absl::StrJoin(strides, ",") << "])";Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Fri Jun 30 03:04:46 GMT 2023 - 4.4K bytes - Click Count (0) -
tensorflow/c/eager/abstract_tensor_handle.h
virtual absl::Status TensorHandleStatus() const; // Returns tensor shape. If tensor has unknown rank, shape remains untouched. virtual absl::Status Shape(tensorflow::PartialTensorShape* shape) const = 0; // Returns tensor (full) type. // While there is no immediate plan to deprecate dtype and shape in favor // of only using full type type information, this is a future possibility. //
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Sat Oct 12 05:11:17 GMT 2024 - 3K bytes - Click Count (0) -
RELEASE.md
* Tracing/timeline support for distributed runtime (no GPU profiler yet). * C API gives access to inferred shapes with `TF_GraphGetTensorNumDims` and `TF_GraphGetTensorShape`. * Shape functions for core ops have moved to C++ via `REGISTER_OP(...).SetShapeFn(...)`. Python shape inference RegisterShape calls use the C++ shape functions with `common_shapes.call_cpp_shape_fn`. A future release will remove `RegisterShape` from python.
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Mon Mar 30 18:31:38 GMT 2026 - 746.5K bytes - Click Count (3) -
src/test/java/jcifs/smb1/smb1/SmbShareInfoTest.java
void testEquals() { SmbShareInfo instance1 = new SmbShareInfo("SHARE1", 0, "remark1"); SmbShareInfo instance2 = new SmbShareInfo("SHARE1", 1, "remark2"); SmbShareInfo instance3 = new SmbShareInfo("SHARE2", 0, "remark1"); Object notAShareInfo = new Object(); assertTrue(instance1.equals(instance2)); // Should be equal based on netName assertFalse(instance1.equals(instance3)); // Should not be equal
Created: Sun Apr 05 00:10:12 GMT 2026 - Last Modified: Thu Aug 14 05:31:44 GMT 2025 - 5K bytes - Click Count (0)