- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 27 for _input_shapes (0.4 sec)
-
tensorflow/compiler/mlir/tensorflow/transforms/tf_saved_model_freeze_variables.cc
if (input_shapes_attr.size() != func_op.getNumArguments()) { return func_op->emitError( "Number of arguments and 'tf._input_shapes' " "attribute size do not match. ") << "Num args: " << func_op.getNumArguments() << ", tf._input_shapes size: " << input_shapes_attr.size(); } return success(); } // Validates ModuleOp. Returns `failure` if the module op is invalid.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 09:56:53 UTC 2024 - 19.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/tests/quantize_xla.mlir
func.func private @conv(%input: tensor<1x3x4x3xf32> {tf._user_specified_name = "input_tensor"}) -> tensor<*xf32> attributes {tf._construction_context = "kEagerRuntime", tf._input_shapes = [#tf_type.shape<1x3x4x3>]} { %weight = arith.constant dense_resource<__elided__> : tensor<2x3x3x2xf32> %bias = arith.constant dense<[7.11401462, 7.05456924]> : tensor<2xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 08 19:32:28 UTC 2024 - 11.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfrt/tests/hoist_invariant_ops.mlir
// Test hoist var handle op and read variable op in the batch function. // CHECK-LABEL: func private @batched_function func.func private @batched_function(%arg0: tensor<1x3xf32>) -> tensor<1x3xf32> attributes {tf._input_shapes = [#tf_type.shape<1x3>, #tf_type.shape<*>], tf.signature.is_stateful} { // CHECK-NOT: tf.VarHandleOp // CHECK-NOT: tf.ReadVariableOp // CHECK: "tf._TfrtGetResource"()
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Apr 01 23:54:14 UTC 2024 - 18.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/tests/passes/insert_calibration_statistics_saver.mlir
func.func private @cond_false_80(%arg0: tensor<1x4xf32> {tf._user_specified_name = "x"}) -> (tensor<i1>, tensor<1x3xf32>) attributes {tf._construction_context = "kEagerRuntime", tf._input_shapes = [#tf_type.shape<1x4>], tf._original_func_name = "cond_false_8"} { %cst = "tf.Const"() <{value = dense<true> : tensor<i1>}> {device = ""} : () -> tensor<i1>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 01:09:50 UTC 2024 - 24.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/tests/insert_custom_aggregation_ops.mlir
return %3 : tensor<1x3xf32> } func.func private @cond_false_80(%arg0: tensor<1x4xf32> {tf._user_specified_name = "x"}) -> (tensor<i1>, tensor<1x3xf32>) attributes {tf._construction_context = "kEagerRuntime", tf._input_shapes = [#tf_type.shape<1x4>], tf._original_func_name = "cond_false_8"} { %cst = "tf.Const"() <{value = dense<true> : tensor<i1>}> {device = ""} : () -> tensor<i1>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 10 04:07:09 UTC 2024 - 32.1K bytes - Viewed (0) -
tensorflow/c/c_api_experimental.cc
// Set input_shapes. for (int i = 0; i < num_inputs; ++i) { std::vector<DimensionHandle> dims; const TF_ShapeAndType& input_shape = input_shapes->items[i]; if (input_shape.num_dims == InferenceContext::kUnknownRank) { c.SetInput(i, c.UnknownShape()); continue; } dims.reserve(input_shape.num_dims); for (int j = 0; j < input_shape.num_dims; ++j) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Apr 15 03:35:10 UTC 2024 - 29.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/translate/tf_mlir_translate.cc
const std::vector<std::optional<std::vector<int>>>& input_shapes, const std::vector<std::string>& output_arrays, const std::vector<std::string>& control_output_arrays, const GraphdefToMlirOptions& import_options, mlir::MLIRContext* context) { auto module_or = GraphdefToMlirImport( input, input_arrays, input_dtypes, input_shapes, output_arrays, control_output_arrays, import_options, context);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 07 11:51:44 UTC 2024 - 14.1K bytes - Viewed (0) -
tensorflow/compiler/jit/increase_dynamism_for_auto_jit_pass.cc
constant_pool.Get1DHostConstant(i), constant_pool.Get1DHostConstant(1)); Output input_shape_i = ops::Slice( host_scope.WithOpName("input_shape_", i), input_shape, constant_pool.Get1DHostConstant(i), constant_pool.Get1DHostConstant(1)); slice_size.push_back(ops::Sub(host_scope.WithOpName("slice_size_", i),
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Mar 12 06:33:33 UTC 2024 - 13.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/tpu_space_to_depth_pass.cc
void HandleConv2DInput(TF::Conv2DOp conv2d, int64_t block_size) { auto input = conv2d.getInput(); auto input_shape = mlir::cast<RankedTensorType>(input.getType()).getShape(); SmallVector<int64_t, 4> transform_shape = { input_shape[0], input_shape[1] / block_size, input_shape[2] / block_size, input_shape[3] * block_size * block_size}; auto transform_result_type =
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 29.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tf_tfl_translate.cc
// Graphdef import path. module = tensorflow::LoadFromGraphdefOrMlirSource( input_file_name, input_mlir, use_splatted_constant, custom_opdefs, specs, debug_info_file, input_arrays, input_dtypes, input_shapes, output_arrays, control_output_arrays, &source_mgr, &context); } // If errors occur, the library call in the above already logged the error // message. So we can just return here.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 03 18:01:23 UTC 2024 - 14K bytes - Viewed (0)