- Sort Score
- Result 10 results
- Languages All
Results 41 - 50 of 109 for Selu (0.04 sec)
-
tensorflow/compiler/mlir/lite/transforms/optimize_patterns.td
(HasRankAtMost<4> $a), (HasRankAtMost<4> $b)]>; } // We can eliminate Relu from Relu(SquaredDifference(x, y)), // since the result of SquaredDifference is always non-negative. // TFLite interpreter doesn't support Relu+int32 for now. So the test cases // are failing without the following pattern to optimize Relu away fixes // the problem. def OptimizeReluSquaredDifference : Pat<
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 16 20:31:41 UTC 2024 - 66.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/layout_optimization_move_transposes_end.mlir
// CHECK: %[[TANH:[0-9]*]] = "tf.Tanh"(%arg0) {{.*}} tensor<1x4x4x8xf32> // CHECK: %[[RELU:[0-9]*]] = "tf.Relu"(%[[TANH]]) {{.*}} tensor<1x4x4x8xf32> // CHECK: %[[RES_TRANSPOSE:[0-9]*]] = "tf.Transpose"(%[[RELU]], %[[RES_PERM]]) // CHECK: return %[[RES_TRANSPOSE]] %0 = "tf.Const"() {value = dense<[0, 3, 1, 2]> : tensor<4xi32>} : () -> tensor<4xi32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 30 06:52:55 UTC 2023 - 9.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/mlir2flatbuffer/svdf_v2.mlir
// CHECK-NEXT: outputs: [ 5 ], // CHECK-NEXT: builtin_options_type: SVDFOptions, // CHECK-NEXT: builtin_options: { // CHECK-NEXT: rank: 2, // CHECK-NEXT: fused_activation_function: RELU // CHECK-NEXT: } // CHECK-NEXT: } ], // CHECK-NEXT: name: "main" // CHECK-NEXT: } ], // CHECK-NEXT: description: "MLIR Converted.", // CHECK-NEXT: buffers: [ { // CHECK-EMPTY:
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jul 14 16:41:28 UTC 2022 - 3.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/stablehlo/transforms/composite_lowering_patterns.td
(variadic $inputs), ConstantStrAttr<StrAttr, "aten.gelu.default">, $attrs, $_, $_), (TFL_GeluOp $inputs, ConstBoolAttrFalse), [(IsStrCompositeAttribute<"approximate", "none"> $attrs)]>; def LegalizeCompositeApproximateAtenGELU : Pat< (MHLO_CompositeOp:$composite (variadic $inputs), ConstantStrAttr<StrAttr, "aten.gelu.default">, $attrs, $_, $_),
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jun 06 18:45:51 UTC 2024 - 4.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfr/examples/mnist/mnist_train.py
# output shape: [-1, 28, 28, 32] conv1 = gen_mnist_ops.new_conv2d(x, self.weights['f1'], self.biases['b1'], 1, 1, 1, 1, 'SAME', 'RELU') # Max pooling. The kernel size spec {ksize} also follows the layout of # the data. Here we have a pooling window of 2, and a stride of 2. # output shape: [-1, 14, 14, 32]
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Oct 20 03:05:18 UTC 2021 - 6.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/mlir2flatbuffer/svdf.mlir
// CHECK-NEXT: outputs: [ 5 ], // CHECK-NEXT: builtin_options_type: SVDFOptions, // CHECK-NEXT: builtin_options: { // CHECK-NEXT: rank: 2, // CHECK-NEXT: fused_activation_function: RELU // CHECK-NEXT: } // CHECK-NEXT: } ], // CHECK-NEXT: name: "main" // CHECK-NEXT: } ], // CHECK-NEXT: description: "MLIR Converted.", // CHECK-NEXT: buffers: [ { // CHECK-EMPTY:
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jul 14 16:41:28 UTC 2022 - 3.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfr/examples/mnist/mnist_ops_test.py
'input_': input_, 'filter_': filter_, 'bias': bias, 'stride_w': 2, 'stride_h': 2, 'dilation_w': 1, 'dilation_h': 1, 'padding': 'SAME', 'act': 'RELU' } self._assertOpAndComposite([input_, filter_, bias], tf.function(gen_mnist_ops.new_conv2d), ops_defs._composite_conv_add_relu, kwargs)
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Sep 28 21:37:05 UTC 2021 - 4K bytes - Viewed (0) -
tensorflow/compiler/jit/tests/keras_imagenet_main_graph_mode.golden_summary
Conv2DBackpropInput 52 DivNoNan 1 Equal 1 FusedBatchNorm 53 FusedBatchNormGrad 53 Identity 2 MatMul 3 MaxPool 1 MaxPoolGrad 1 Mean 1 Mul 164 Pad 1 ReadVariableOp 646 Relu 49 ReluGrad 49 Reshape 2 ResourceApplyKerasMomentum 161 ShapeN 50 Softmax 1 SparseSoftmaxCrossEntropyWithLogits 1 Square 55 Squeeze 1 Sub 106 Sum 57 Tile 1
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Jan 06 10:38:14 UTC 2023 - 740 bytes - Viewed (0) -
tensorflow/compiler/mlir/tfr/README.md
attrs=['act: {"", "RELU", "RELU6", "TANH"} = ""'], derived_attrs=['T: {float, int8}'], outputs=['o: T']) def _composite_fully_connected(input_, filter_, bias, act): res = tf.raw_ops.MatMul( a=input_, b=filter_, transpose_a=False, transpose_b=True) res = tf.raw_ops.Add(x=res, y=bias) if act == 'RELU': return tf.raw_ops.Relu(features=res) elif act == 'RELU6':
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Mar 29 18:32:13 UTC 2022 - 6.2K bytes - Viewed (0) -
tensorflow/compiler/jit/tests/keras_imagenet_main.golden_summary
Conv2D 53 Conv2DBackpropFilter 53 Conv2DBackpropInput 52 Equal 1 FusedBatchNormGradV2 53 FusedBatchNormV2 53 MatMul 3 MaxPool 1 MaxPoolGrad 1 Mean 1 Mul 218 Pad 2 ReadVariableOp 538 Relu 49 ReluGrad 49 Reshape 2 ResourceApplyKerasMomentum 161 Slice 1 Softmax 1 SparseSoftmaxCrossEntropyWithLogits 1 Squeeze 1 Sum 1 Tile 1 Transpose 1 cluster 1 size 815 AddN 1
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Jan 06 10:38:14 UTC 2023 - 874 bytes - Viewed (0)