- Sort Score
- Result 10 results
- Languages All
Results 21 - 30 of 290 for shake (0.12 sec)
-
tensorflow/compiler/jit/xla_host_send_recv_device_context_test.cc
TF_ASSERT_OK_AND_ASSIGN(auto stream, executor->CreateStream()); se::DeviceMemoryBase gpu_dst{device_tensor.data(), 4 * sizeof(float)}; xla::Shape shape; TF_ASSERT_OK(TensorShapeToXLAShape(DT_FLOAT, TensorShape({2, 2}), &shape)); // Copy the cpu_tensor to the GPU first before trying to copy it back. TF_ASSERT_OK( stream->Memcpy(&gpu_dst, origin_cpu_tensor.data(), gpu_dst.size()));
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 22:46:36 UTC 2024 - 7.2K bytes - Viewed (0) -
src/runtime/stubs.go
import ( "internal/abi" "unsafe" ) // Should be a built-in for unsafe.Pointer? // // add should be an internal detail, // but widely used packages access it using linkname. // Notable members of the hall of shame include: // - fortio.org/log // // Do not remove or change the type signature. // See go.dev/issue/67401. // //go:linkname add //go:nosplit func add(p unsafe.Pointer, x uintptr) unsafe.Pointer {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 20.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/flatbuffer_operator.cc
std::vector<int64_t> shape; shape.push_back(static_cast<int64_t>(data.size())); attributes.emplace_back(builder.getNamedAttr( "window_strides", BuildVhloTensorV1Attr(shape, data, builder))); } if (!(op->padding.empty())) { std::vector<int64_t> shape; shape.push_back(static_cast<int64_t>(op->padding.size()) / 2); shape.push_back(2);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 21 18:21:50 UTC 2024 - 38K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/python/integration_test/quantize_model_test_base.py
tensor_spec.TensorSpec( shape=shape, dtype=dtypes.float32, name='input_tensor' ) ), ) return model # Prepares sample einsum input data shapes. # This function returns: # 1. Shape for input 1 # 2. Shape for input 2 # 3. Shape for bias # 4. Signature for input 1 (Could contain None dimension)
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 14 06:31:57 UTC 2024 - 18.2K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_host_recv_device_context.h
// xla::Shape shape(xla::F32, {2, 2}, {}, {}) // tsl::AsyncValueRef<std::unique_ptr<se::Event>> done_event = // tsl::MakeConstructedAsyncValueRef<std::unique_ptr<se::Event>>(stream.parent()); // done_event->Init(); // Tensor dest_cpu_tensor; // // XlaHostRecvDeviceContext device_context(&stream, gpu_dst, // shape, done_event); // device_context.CopyDeviceTensorToCPUSync(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 22:46:36 UTC 2024 - 3.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/optimize_patterns.td
// tail of the other operand and the intermediate result isn't used by other // ops. // $rhs is required to be the tail shape of $lhs, so after transformation the // shape of the binary op result is valid. For example, assume the shapes of // $input, $lhs and $rhs are [1600], [1,40,40] and [40x1]. After the // transformation, the shape of the binary op result is [40x1600], which
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 16 20:31:41 UTC 2024 - 66.4K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_device_context.cc
xla_tensor->WaitForDefinitionEventOnStream(device_to_host_stream.get()); // Transfer manager requires the shape of the shaped buffer to be the same as // literal shape except for the layout. Set the literal to use xla_tensor's // shape as it is derived from the cpu_tensor's shape using // shape_representation_fn_. xla::MutableBorrowingLiteral literal; TF_CHECK_OK(HostTensorToMutableBorrowingLiteral(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 16 00:36:08 UTC 2024 - 12.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/ir/tf_generated_ops.td
merged[indices[m][i], ...] = data[m][i, ...] ``` Each `data[i].shape` must start with the corresponding `indices[i].shape`, and the rest of `data[i].shape` must be constant w.r.t. `i`. That is, we must have `data[i].shape = indices[i].shape + constant`. In terms of this `constant`, the output shape is merged.shape = [max(indices) + 1] + constant
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 23:24:08 UTC 2024 - 793K bytes - Viewed (0) -
src/net/http/clone.go
"mime/multipart" "net/textproto" "net/url" _ "unsafe" // for linkname ) // cloneURLValues should be an internal detail, // but widely used packages access it using linkname. // Notable members of the hall of shame include: // - github.com/searKing/golang // // Do not remove or change the type signature. // See go.dev/issue/67401. // //go:linkname cloneURLValues func cloneURLValues(v url.Values) url.Values { if v == nil {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 01:17:26 UTC 2024 - 3K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/utils/tf_to_xla_attribute_utils.cc
auto reshape_op = [&](Value value, const SmallVector<int64_t> &shape) { const int64_t rank = shape.size(); return builder.create<TF::ReshapeOp>( loc, RankedTensorType::get(shape, builder.getI32Type()), value, CreateConstValue<int64_t>(builder, loc, {rank}, shape)); }; ShapedType filter_shape = mlir::cast<ShapedType>(filter.getType());
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 17:58:54 UTC 2024 - 13.3K bytes - Viewed (0)