- Sort Score
- Result 10 results
- Languages All
Results 21 - 30 of 828 for shake (0.06 sec)
-
tensorflow/c/eager/unified_api_test.cc
absl::Span<AbstractTensorHandle*> outputs) { PartialTensorShape shape; TF_RETURN_IF_ERROR(inputs[0]->Shape(&shape)); if (shape.dims() != 0) { return errors::InvalidArgument( "Tensor expected to have scalar shape found rank: ", shape.dims()); } return absl::OkStatus(); } TEST_P(UnifiedAPI, TestTensorShapeScalar) { if (UseFunction() && UseMlir()) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Feb 27 13:57:45 UTC 2024 - 6.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/experimental/tac/transforms/device_transform_patterns.cc
if (pack_axis < 0) { pack_axis += rank; } // Concat out shape. for (int i = 0; i < rank; ++i) { int64_t dim_size = input_type.getDimSize(i); if (i == pack_axis) { dim_size *= count; } concat_out_shape.push_back(dim_size); } // Pack out shape. int j = 0; for (int i = 0; i < rank + 1; ++i) { if (i == pack_axis) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 25.4K bytes - Viewed (0) -
tensorflow/compiler/jit/tests/device_compiler_test_helper.cc
TF_RETURN_IF_ERROR(session->Create(graph)); RunOptions run_options; Tensor input_a = CreateInputTensor(shape, 0); Tensor input_b = CreateInputTensor(shape, shape.num_elements()); Tensor input_c = CreateInputTensor(shape, 2 * shape.num_elements()); TF_RETURN_IF_ERROR(session->Run( run_options, {std::make_pair("a", input_a), std::make_pair("b", input_b),
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Feb 09 08:24:16 UTC 2024 - 6.2K bytes - Viewed (0) -
tensorflow/c/experimental/next_pluggable_device/tensor_pjrt_buffer_util_test.cc
auto c_api_client = down_cast<xla::PjRtCApiClient*>(pjrt_client->get()); std::vector<int32_t> data(1, 0); xla::Shape shape = xla::ShapeUtil::MakeShape(xla::S32, {1}); auto buffer = c_api_client->pjrt_c_client()->client->BufferFromHostBuffer( data.data(), shape.element_type(), shape.dimensions(), /*byte_strides=*/std::nullopt, xla::PjRtClient::HostBufferSemantics::kImmutableOnlyDuringCall, nullptr,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Apr 01 16:29:40 UTC 2024 - 7.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/utils/utils.h
return transposed_type; } // Returns shape of a ranked tensor. // Precondition: output_val's is ranked tensor. // Returns a truncated shape when `truncate` is set to true. inline DenseElementsAttr GetShape(Value output_val, bool truncate = false) { auto output_shape = output_val.getType().dyn_cast<ShapedType>().getShape(); SmallVector<int32_t> shape; shape.reserve(output_shape.size());
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Apr 30 00:40:15 UTC 2024 - 11.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/api/v1/compile_mlir_util.cc
// Extracts shape from XlaArgument as TensorShape. If shape is a xla::Shape, // that is converted to a TensorShape. absl::StatusOr<TensorShape> GetTensorShapeFromXlaArgument( const XlaArgument& arg) { if (absl::holds_alternative<xla::Shape>(arg.shape)) { TensorShape arg_shape; TF_RETURN_IF_ERROR( XLAShapeToTensorShape(std::get<xla::Shape>(arg.shape), &arg_shape)); return arg_shape;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 21 17:24:39 UTC 2024 - 45.3K bytes - Viewed (0) -
tensorflow/c/kernels/ops/bitcast.cc
if (input_type_size < output_type_size) { TF_ShapeInferenceContextWithRankAtLeast(ctx, shape, 1, shape, status); if (TF_GetCode(status) == TF_OK) { TF_DimensionHandle* last_dim = TF_NewDimensionHandle(); size_t divisor_val = output_type_size / input_type_size; TF_ShapeInferenceContextDim(ctx, shape, -1, last_dim); if (!TF_DimensionHandleValueKnown(last_dim) ||
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 07:51:50 UTC 2024 - 5.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/flatbuffer2mlir/lstm.json
{ "tensors": [ { "shape": [1, 5, 2], "name": "input0" }, { "shape": [2, 5], "buffer": 1, "name": "input2input_weights1" }, { "shape": [2, 5], "buffer": 2, "name": "input2forget_weights2" }, { "shape": [2, 5], "buffer": 3,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 01 06:25:50 UTC 2024 - 9.1K bytes - Viewed (0) -
tensorflow/compiler/jit/shape_inference.cc
// Merge node causes a loop so we remove NextIteration->Merge edge before // performing shape inference. But removing those edges also prevents us // from inferring output shape for Merge node (we need shapes for all its // inputs). // For loop invariant resource input's Merge node, we set output resource // shape as Enter node's resource shape. // TODO(b/129367850): clean this up.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 31 00:41:19 UTC 2024 - 13K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_host_send_device_context.h
// se::DeviceMemoryBase gpu_dst{device_tensor.data(), 4 * sizeof(float)}; // xla::Shape shape(xla::F32, {2, 2}, {}, {}) // tsl::AsyncValueRef<std::unique_ptr<se::Event>> done_event = // tsl::MakeConstructedAsyncValueRef<std::unique_ptr<se::Event>>(stream.parent()); // done_event->Init(); // // XlaHostSendDeviceContext device_context(&stream, &gpu_dst, // shape, done_event); // device_context.CopyCPUTensorToDeviceSync(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 22:46:36 UTC 2024 - 3.7K bytes - Viewed (0)