- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 21 for add_0 (0.12 sec)
-
tensorflow/compiler/mlir/quantization/stablehlo/tests/passes/defer_activation_transpose.mlir
%2 = stablehlo.add %1, %0 : tensor<1x3x2x4xf32> return %2 : tensor<1x3x2x4xf32> } // CHECK: %[[TRANSPOSE_0:.+]] = stablehlo.transpose // CHECK: %[[ADD_0:.+]] = stablehlo.add %[[TRANSPOSE_0]], {{.*}} // CHECK: return %[[ADD_0]] // ----- // [No change] Tests that the activation transpose whose rank is not 4 is not // deferred. // CHECK-LABEL: add_with_activation_transpose_rank_two
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 18 20:32:46 UTC 2024 - 14.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/tests/passes/quantize_composite_functions.mlir
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 09 05:56:10 UTC 2024 - 91.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/passes/lift_quantizable_spots_as_functions_fusion.td
[(IsNotInLiftedFunc $res), (FloatValueEquals<"0"> $cst), (IsStableHLOConstantOp $bias), (AreTheSameValue $conv_0, $conv_1), (AreTheSameValue $add_0, $add_1)], [], (addBenefit 15)>; def LiftDotGeneralWithBiasAndReluDynamic : Pat< (StableHLO_MaxOp:$res (StableHLO_AddOp:$add_0 (StableHLO_DotGeneralOp:$dot_general_0 $lhs, $rhs, $dot_dimension_numbers, $precision_config), (StableHLO_DynamicBroadcastInDimOp
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 04 07:19:09 UTC 2024 - 23.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/quantization/tensorflow/tests/fallback_to_flex_ops_legacy.mlir
%0 = "tf.AddV2"(%arg0, %arg1) : (tensor<1xf32>, tensor<1xf32>) -> tensor<1xf32> func.return %0: tensor<1xf32> // CHECK: %[[ADD_0:.*]] = "tf.AddV2"(%arg0, %arg1) : (tensor<1xf32>, tensor<1xf32>) -> tensor<1xf32> // CHECK: return %[[ADD_0]] : tensor<1xf32> } // CHECK-LABEL: softmax func.func @softmax(%arg0: tensor<8x16xf32>) -> tensor<8x16xf32> {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 5.8K bytes - Viewed (0) -
tensorflow/compiler/jit/mark_for_compilation_pass_test.cc
ops::Switch switch_b(root.WithOpName("switch_b"), value, cond_b); Output add_a = ops::Add(root.WithOpName("add_a"), switch_a.output_true, switch_b.output_true); Output add_b = ops::Add(root.WithOpName("add_b"), switch_a.output_true, switch_b.output_true); Output add = ops::Add(root.WithOpName("add_c"), add_a, add_b); std::unique_ptr<Graph> graph(new Graph(OpRegistry::Global()));
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 14 10:11:10 UTC 2024 - 79.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/utils/export_utils.cc
for (mlir::Attribute a : attr.getValue()) { if (auto attr = mlir::dyn_cast<mlir::BoolAttr>(a)) { list->add_b(attr.getValue()); } else if (auto attr = mlir::dyn_cast<mlir::IntegerAttr>(a)) { list->add_i(attr.getInt()); } else if (auto attr = mlir::dyn_cast<mlir::FloatAttr>(a)) { list->add_f(attr.getValueAsDouble()); } else if (auto attr = mlir::dyn_cast<mlir::StringAttr>(a)) { AttrValue nested_value;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 17:58:54 UTC 2024 - 19.7K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_activity_listener_test.cc
->set_global_jit_level(OptimizerOptions::ON_2); std::unique_ptr<Session> session(NewSession(options)); TF_ASSERT_OK(session->Create(graph_def)); std::vector<std::string> output_names = {std::string("add_4:0")}; Tensor tensor_2x2(DT_FLOAT, TensorShape({2, 2})); for (int i = 0; i < 4; i++) { tensor_2x2.matrix<float>()(i / 2, i % 2) = 5 * i; } Tensor tensor_3x3(DT_FLOAT, TensorShape({3, 3}));
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 08:47:20 UTC 2024 - 5.9K bytes - Viewed (0) -
tensorflow/compiler/jit/deadness_analysis_test.cc
ops::Switch sw_3 = CreateSwitch(root, "3"); Output add0 = ops::Add(root.WithOpName("add0"), sw_0.output_false, sw_1.output_false); Output add1 = ops::Add(root.WithOpName("add1"), sw_2.output_false, sw_3.output_false); ops::Merge m0(root.WithOpName("m0"), {add0, add1}); ops::Merge m1(root.WithOpName("m1"), {add0, add1}); Output add2 = ops::Add(root.WithOpName("add2"), m0.output, m1.output);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 06:59:07 UTC 2024 - 51.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/flatbuffer2mlir/composite_op_round_trip.mlir
// CHECK-ROUNDTRIP: %0 = stablehlo.composite "stablehlo.add_n" %arg0 {composite_attributes = {test_bool = false, test_int = 2 : i64, test_string = "test"}, decomposition = @add_n.impl} : (tensor<i64>) -> tensor<i64> %0 = stablehlo.composite "stablehlo.add_n" %arg0 { composite_attributes = { test_int = 2 : i64, test_bool = 0 : i1, test_string = "test"}, decomposition = @add_n.impl } : (tensor<i64>) -> tensor<i64> return %0 : tensor<i64> }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Apr 24 03:40:50 UTC 2024 - 1.1K bytes - Viewed (0) -
src/runtime/mbitmap.go
return } tp0 := s.typePointersOfType(typ, addr) tp1 := s.typePointersOf(addr, size) failed := false for { var addr0, addr1 uintptr tp0, addr0 = tp0.next(addr + size) tp1, addr1 = tp1.next(addr + size) if addr0 != addr1 { failed = true break } if addr0 == 0 { break } } if failed { tp0 := s.typePointersOfType(typ, addr) tp1 := s.typePointersOf(addr, size)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 00:18:55 UTC 2024 - 60K bytes - Viewed (0)