- Sort Score
- Result 10 results
- Languages All
Results 31 - 40 of 209 for mhlo (0.06 sec)
-
tensorflow/compiler/mlir/quantization/stablehlo/tests/passes/unfuse_mhlo_batch_norm.mlir
// CHECK: %[[X_CENTER:.+]] = mhlo.subtract %[[X]], %[[MEAN_BCAST]] : tensor<4x256xf32> // CHECK: %[[X_SCALED:.+]] = mhlo.multiply %[[X_CENTER]], %[[SCALE_BCAST]] : tensor<4x256xf32> // CHECK: %[[X_NORMED:.+]] = mhlo.divide %[[X_SCALED]], %[[STDDEV_BCAST]] : tensor<4x256xf32> // CHECK: %[[RESULT:.+]] = mhlo.add %[[X_NORMED]], %[[OFFSET_BCAST]] : tensor<4x256xf32> %0 = "mhlo.batch_norm_inference"(%x, %scale, %offset, %mean, %variance)
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Apr 06 15:32:52 UTC 2024 - 2.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/tests/bridge/convert_tf_quant_ops_to_mhlo.mlir
// CHECK-DAG: %[[LHS_1:.*]] = mhlo.convert %arg0 : tensor<3x2xi32> // CHECK-DAG: %[[LHS_2:.*]] = mhlo.bitcast_convert %[[LHS_1]] : (tensor<3x2xi32>) -> tensor<3x2x!quant.uniform<i32:f32, 2.000000e+00:4>> // CHECK-DAG: %[[RHS_1:.*]] = mhlo.convert %arg1 : tensor<2xi32> // CHECK-DAG: %[[RHS_2:.*]] = mhlo.bitcast_convert %[[RHS_1]] : (tensor<2xi32>) -> tensor<2x!quant.uniform<i32:f32, 2.000000e+00:4>>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Apr 06 15:32:52 UTC 2024 - 7.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/stablehlo/transforms/composite_avg_pool.h
#include "xla/mlir_hlo/mhlo/IR/hlo_ops.h" // IWYU pragma: keep namespace mlir { namespace odml { // Given a Composite op that wraps a core.aten.avg_pool2d, returns the padding // configuration required for the `tfl.pad` if the padding part of the op is // to be done before average pooling. DenseIntElementsAttr GetPadOpAttr(Builder& builder, mhlo::CompositeOp op);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 28 23:16:05 UTC 2024 - 2.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/passes/bridge/convert_tf_quant_ops_to_mhlo.cc
if (failed(output_type)) { return failure(); } auto input_quant = rewriter.create<mhlo::BitcastConvertOp>( op->getLoc(), *input_quant_type, input); auto result = rewriter.create<mhlo::UniformQuantizeOp>( op->getLoc(), *output_type, input_quant); rewriter.replaceOpWithNewOp<mhlo::BitcastConvertOp>( op, output_type->clone(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 17:58:54 UTC 2024 - 30.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/tools/stablehlo_quant_opt.cc
#include "tensorflow/compiler/mlir/tensorflow/ir/tf_saved_model.h" #include "tensorflow/compiler/mlir/tensorflow/transforms/passes.h" #include "xla/mlir_hlo/mhlo/IR/register.h" #include "xla/mlir_hlo/mhlo/transforms/passes.h" #include "tensorflow/core/ir/types/dialect.h" int main(int argc, char** argv) { tensorflow::InitMlir y(&argc, &argv); mlir::registerAllPasses();
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 07:37:34 UTC 2024 - 3.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/transforms/legalize_tf.cc
} }; // Converts TensorFlow DiagPartOp to HLO ops using reduction on masked matrix. // For a Rank-2 input, it creates the following ops: // %1 = "mhlo.iota"() {iota_dimension = 0 : i64} // %2 = "mhlo.iota"() {iota_dimension = 1 : i64} // %3 = "mhlo.compare"(%1, %2) {comparison_direction = "EQ"} // %4 = mhlo.constant dense<0.000000e+00> : tensor<f32> // %5 = "mhlo.broadcast"(%4) // %6 = "mhlo.select"(%3, %input, %5)
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 20:00:43 UTC 2024 - 291.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/stablehlo/transforms/legalize_hlo_conversions/dot_general.cc
lhs_flattend = rewriter.create<mhlo::ReshapeOp>( loc, RankedTensorType::get(lhs_flattened_shape, lhs_type.getElementType()), lhs_transposed.getResult()); } else { auto lhs_flattend_shape_op = BuildDotOperandFlattenedShapeOp( lhs, lhs_dot_dimensions_info, builder, /*is_lhs=*/true); lhs_flattend = rewriter.create<mhlo::DynamicReshapeOp>( loc,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 19.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/stablehlo/transforms/legalize_hlo.cc
// Converts a mhlo.reduce op with a mlho binary operation into a tensorflow // reduction operation. If the initial value can be ignored, then convert it // into a single TfReduceOp. Otherwise, convert it into a TfReduceOp followed by // a TfBinaryOp. // For example: // 1) A mhlo::ReduceOp on value `x` with a mhlo::AndOp and a constant initial // value `true` is converted to a TF::Any on value `x`.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 154.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/transforms/legalize_tf_communication.cc
#include "xla/client/sharding_builder.h" #include "xla/mlir_hlo/mhlo/IR/hlo_ops.h" #include "xla/primitive_util.h" #include "xla/side_effect_util.h" #include "xla/translate/mhlo_to_hlo/type_to_shape.h" namespace mlir { using func::FuncOp; namespace mhlo { namespace { constexpr char kShardingAttr[] = "mhlo.sharding"; constexpr char kFrontendAttributesAttr[] = "mhlo.frontend_attributes";
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 40.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/stablehlo/transforms/legalize_hlo_conversions/scatter.h
rewriter.replaceOpWithNewOp<mhlo::TransposeOp>( scatter_op, scatter_op.getResult(0).getType(), tf_scatter_op, inverse_permutation); return success(); } } }; using ConvertScatterAddOp = ConvertScatterOp<mhlo::AddOp, TF::TensorScatterAddOp>; using ConvertScatterMaxOp = ConvertScatterOp<mhlo::MaxOp, TF::TensorScatterMaxOp>; using ConvertScatterMinOp =
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 10.1K bytes - Viewed (0)