- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 137 for Toutput_types (0.27 sec)
-
tensorflow/compiler/mlir/tensorflow/tests/graphdef2mlir/tf-data-pipeline.pbtxt
} } } } node { name: "TensorSliceDataset" op: "TensorSliceDataset" input: "tensors/normalize_tensors/component_0" attr { key: "Toutput_types" value { list { type: DT_INT32 } } } attr { key: "output_shapes" value { list { shape { } } } } }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jul 29 04:41:05 UTC 2021 - 4K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/utils/perception_ops_utils_test.cc
auto indices_type = RankedTensorType::get(input_shape, builder->getI64Type()); auto output_type = RankedTensorType::get(output_shape, builder->getF32Type()); SmallVector<mlir::Type, 2> input_types{input_type, indices_type}; SmallVector<mlir::Type, 1> output_types{output_type}; return createMaxUnpoolingFunc<2, 1>(builder, input_types, output_types); } template <int N>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Sep 29 21:02:21 UTC 2022 - 7.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/utils/lstm_utils_test.cc
auto output_types = fused_lstm_func_.getFunctionType().getResults(); SmallVector<int64_t, 2> output_shape{1, mlir::ShapedType::kDynamic}; EXPECT_EQ(mlir::cast<RankedTensorType>(output_types[0]).getShape().size(), output_shape.size()); for (int i = 0; i < output_shape.size(); i++) { EXPECT_EQ(mlir::cast<RankedTensorType>(output_types[0]).getDimSize(i), output_shape[i]);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 10K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/post_quantize.cc
Value dequantized_result = dequantize_op.getInput(); output_types.push_back(dequantized_result.getType()); terminator->setOperand(i, dequantized_result); returned_op->erase(); } else { output_types.push_back(returned_value.getType()); } } auto new_func_type = builder.getFunctionType(input_types, output_types); func.setType(new_func_type); } enum RemoveVolatileOpsType {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 17.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/utils/xla_sharding_util.cc
shape[split_dimension] = shape[split_dimension] / num_split; output_type = mlir::RankedTensorType::get(shape, input_type.getElementType()); } } else { output_type = input_type; } // Creates a split op that splits |src_input| along |split_dimension|. llvm::SmallVector<mlir::Type, 4> output_types(num_split, output_type); *split_op = builder->create<mlir::TF::SplitOp>(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 22 21:28:13 UTC 2024 - 34K bytes - Viewed (0) -
tensorflow/compiler/jit/encapsulate_xla_computations_pass.cc
} // Outputs. const int num_outputs = launch->output_types().size(); absl::flat_hash_set<Node*> control_outputs; std::vector<std::vector<std::pair<Node*, int>>> data_outputs(num_outputs); const DataTypeVector& output_types(launch->output_types()); for (const Edge* le : launch->out_edges()) { if (le->IsControlEdge()) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Mar 12 06:33:33 UTC 2024 - 15.1K bytes - Viewed (0) -
tensorflow/compiler/jit/build_xla_ops_pass.cc
MemoryTypeVector input_mtypes, output_mtypes; DeviceType device_type(""); TF_RETURN_IF_ERROR( DeviceNameToDeviceType(n->assigned_device_name(), &device_type)); TF_RETURN_IF_ERROR(MemoryTypesForNode(root.graph()->op_registry(), device_type, n->def(), &input_mtypes, &output_mtypes)); return output_mtypes; }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Mar 12 06:33:33 UTC 2024 - 24.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/utils/nms_utils.cc
Value iou_threshold = func_.getArgument(3); Value score_threshold = func_.getArgument(4); auto output_type0 = func_.getFunctionType().getResult(0); auto output_type1 = func_.getFunctionType().getResult(1); OpBuilder builder(func_.getBody()); auto op = builder.create<mlir::TFL::NonMaxSuppressionV4Op>( func_.getLoc(), output_type0, output_type1, boxes, scores, max_output_size, iou_threshold, score_threshold);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 8.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/utils/lstm_utils.cc
size_of_splits = dim / splits; output_shape.push_back(size_of_splits); } else { output_shape.push_back(dim); } } SmallVector<mlir::Type, 4> output_types; for (int i = 0; i < splits; ++i) { output_types.push_back( mlir::RankedTensorType::get(output_shape, input_type.getElementType())); } auto size_of_splits_op = Create1DConstantOp(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 36.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/quantize_composite_functions.cc
} output_types = {new_output_type}; } else { output_types = {output_type.clone(elem_type.getStorageType())}; } SmallVector<Value> args = {q_op.getArg(), scale, zero_point}; FlatSymbolRefAttr func_name = FlatSymbolRefAttr::get(rewriter.getStringAttr(kQuantizeFuncName)); auto quantize_call = rewriter.create<TF::PartitionedCallOp>( loc, output_types, args, func_name,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 54.5K bytes - Viewed (0)