- Sort Score
- Result 10 results
- Languages All
Results 41 - 50 of 51 for s1_shape (0.19 sec)
-
tensorflow/compiler/jit/xla_launch_util.cc
output.set_shapes(output_device_shape, output_device_shape); for (int i = 0; i < ctx->num_outputs(); ++i) { const xla::Shape& subshape = xla::ShapeUtil::GetSubshape(output_device_shape, {i}); TensorShape shape; TF_RETURN_IF_ERROR(XLAShapeToTensorShape(subshape, &shape)); output_tensor_shapes.push_back(shape); } } else { for (int i = 0; i < ctx->num_outputs(); ++i) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 16 00:36:08 UTC 2024 - 40.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/ir/tf_ops_n_z.cc
} else { output_ty = tensorflow::GetTypeFromTFTensorShape( {static_cast<int64_t>(sub_shape.size())}, output_elt_ty); } } // Down-cast to 32 bit int if needed. if (output_elt_ty.isInteger(32)) { SmallVector<int32_t, 2> sub_shape_i32(sub_shape.size()); std::transform(sub_shape.begin(), sub_shape.end(), sub_shape_i32.begin(), [](int64_t d) { return static_cast<int32_t>(d); });
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 09 22:07:10 UTC 2024 - 170.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/python/integration_test/quantize_model_test.py
has_bias: bool, use_kernel: bool, ): n = 5 x_shape = [v if v is not None else n for v in shapes[0]] y_shape = [v if v is not None else n for v in shapes[1]] class MatmulModel(module.Module): def __init__(self, bias: Optional[core.Tensor]): self._bias = bias self._kernel = np.random.uniform(size=y_shape).astype('f4') self._min = (-0.8, -0.8, -0.9)
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 03:36:50 UTC 2024 - 235.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/optimize.cc
q_op.setOperand(new_bias_op); auto new_q_op_type = RankedTensorType::Builder( mlir::cast<RankedTensorType>(q_op.getResult().getType())) .setShape(mlir::cast<ShapedType>(new_bias_type).getShape()); q_op.getResult().setType(new_q_op_type); auto attr = TypeAttr::get(q_op.getResult().getType()); q_op.setQtypeAttr(attr);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Apr 30 00:40:15 UTC 2024 - 102.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/schema/schema_generated.h
typedef VariantSubTypeBuilder Builder; enum FlatBuffersVTableOffset FLATBUFFERS_VTABLE_UNDERLYING_TYPE { VT_SHAPE = 4, VT_TYPE = 6, VT_HAS_RANK = 8 }; const ::flatbuffers::Vector<int32_t> *shape() const { return GetPointer<const ::flatbuffers::Vector<int32_t> *>(VT_SHAPE); } tflite::TensorType type() const { return static_cast<tflite::TensorType>(GetField<int8_t>(VT_TYPE, 0));
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 21 18:21:50 UTC 2024 - 1M bytes - Viewed (0) -
tensorflow/c/c_api.cc
return; } tensorflow::shape_inference::ShapeHandle new_shape = tensorflow::ShapeHandleFromDims(ic, num_dims, dims); status->status = graph->refiner.SetShape(node, output.index, new_shape); } int TF_GraphGetTensorNumDims(TF_Graph* graph, TF_Output output, TF_Status* status) { Node* node = &output.oper->node;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Apr 15 03:35:10 UTC 2024 - 102.3K bytes - Viewed (0) -
tensorflow/c/c_api_test.cc
TF_Buffer* buf = TF_GetAllOpList(); tensorflow::OpList op_list; EXPECT_TRUE(op_list.ParseFromArray(buf->data, buf->length)); EXPECT_GT(op_list.op_size(), 0); TF_DeleteBuffer(buf); } TEST(CAPI, SetShape) { TF_Status* s = TF_NewStatus(); TF_Graph* graph = TF_NewGraph(); TF_Operation* feed = Placeholder(graph, s); ASSERT_EQ(TF_OK, TF_GetCode(s)) << TF_Message(s); TF_Output feed_out_0 = TF_Output{feed, 0};
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Apr 15 03:35:10 UTC 2024 - 96.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/transforms/legalize_tf.cc
Value result = reduction.getResult(0); // The mean op needs to divide by the product of the reduced dimensions. if (std::is_same<OpTy, TF::MeanOp>::value) { Value in_shape = rewriter.create<shape::ShapeOfOp>(loc, op.getInput()); Value divisor_count = rewriter.create<arith::ConstantIndexOp>(loc, 1); for (size_t i = 0; i < input_shape.size(); ++i) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 20:00:43 UTC 2024 - 291.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/g3doc/_includes/tf_passes.md
} return } ``` ### `-tf-rewrite-tpu-embedding-ops` _Rewrites TPU embedding send/recv ops by adding TPU embedding deduplication data_ ### `-tf-shape-inference` _Shape inference on TF dialect and ops implementing InferTypeOpInterface_ Fixed point shape refinement pass that utilizes the shape functions registered on ops using the InferTypeOpInterface as well as by bridging to
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Aug 02 02:26:39 UTC 2023 - 96.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/ir/tf_generated_ops.td
Arg<TF_Int64Tensor, [{2-D. The `indices` of the second `SparseTensor`, size `[nnz, ndims]` Matrix.}]>:$b_indices, Arg<TF_NumberTensor, [{1-D. The `values` of the second `SparseTensor`, size `[nnz]` Vector.}]>:$b_values, Arg<TF_Int64Tensor, [{1-D. The `shape` of the second `SparseTensor`, size `[ndims]` Vector.}]>:$b_shape,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 23:24:08 UTC 2024 - 793K bytes - Viewed (0)