Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 20 of 339 for relu (0.36 sec)

  1. tensorflow/compiler/mlir/tensorflow/tests/layout_optimization_to_nhwc.mlir

      %14 = "tf.AddV2"(%10, %12) : (tensor<?x256x56x56xf32>, tensor<?x256x56x56xf32>) -> tensor<?x256x56x56xf32>
      %15 = "tf.Relu"(%14) : (tensor<?x256x56x56xf32>) -> tensor<?x256x56x56xf32>
    
      // CHECK: %[[ADD:[0-9]*]] = "tf.AddV2"(%[[BATCH_NORM1]], %[[BATCH_NORM2]])
      // CHECK: %[[RELU:[0-9]*]] = "tf.Relu"(%[[ADD]])
    
      // Reduce spatial dimensions
      %16 = "tf.Mean"(%15, %1) : (tensor<?x256x56x56xf32>, tensor<2xi32>) -> tensor<?x256xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 7.3K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/quantization/tensorflow/tests/lift_quantizable_spots_as_functions_drq.mlir

    // CHECK-SAME: f = @composite_conv3d_fn_1}>
    // CHECK-NOT: {_tfl_quant_trait = "fully_quantizable"
    // CHECK: %[[RELU:.*]] = "tf.Relu"(%[[PARTITIONEDCALL_0]])
    // CHECK: return %[[RELU]]
    
    // CHECK-LABEL: private @composite_conv3d_fn_1
    
    // WEIGHTONLY-DAG: %[[CST:.*]] = "tf.Const"() {{.*}} : () -> tensor<2x3x3x3x2xf32>
    // WEIGHTONLY: %[[PARTITIONEDCALL_0:.*]] = "tf.PartitionedCall"(%arg0, %[[CST]])
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 11.8K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/tensorflow/tests/layout_optimization_move_transposes_begin.mlir

      // CHECK: %[[TANH:[0-9]*]] = "tf.Tanh"(%[[ARG_TRANSPOSE]]) {{.*}} tensor<1x8x4x4xf32>
      // CHECK: %[[RELU:[0-9]*]] = "tf.Relu"(%[[TANH]]) {{.*}} tensor<1x8x4x4xf32>
      // CHECK: return %[[RELU]]
    
      %0 = "tf.Tanh"(%arg0) : (tensor<1x4x4x8xf32>) -> tensor<1x4x4x8xf32>
      %1 = "tf.Relu"(%0) : (tensor<1x4x4x8xf32>) -> tensor<1x4x4x8xf32>
    
      %2 = "tf.Const"() {value = dense<[0, 3, 1, 2]> : tensor<4xi32>} : () -> tensor<4xi32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 6.3K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/tensorflow/passes/quantized_function_library_uniform_quantized.mlir

          {"quantized_ops": ["${main_op}", "BiasAdd", "Relu"], "act_func": "internal_requantize_and_relu_fn", "output_type": "!tf_type.qint8"},
          {"quantized_ops": ["${main_op}", "BiasAdd", "Relu6"], "act_func": "internal_requantize_and_relu6_fn", "output_type": "!tf_type.qint8"},
        ]
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Aug 29 01:13:58 UTC 2023
    - 19.3K bytes
    - Viewed (0)
  5. tensorflow/c/experimental/gradients/nn_grad_test.cc

    using tensorflow::TF_StatusPtr;
    
    Status ReluModel(AbstractContext* ctx,
                     absl::Span<AbstractTensorHandle* const> inputs,
                     absl::Span<AbstractTensorHandle*> outputs) {
      return ops::Relu(ctx, inputs[0], &outputs[0], "Relu");
    }
    
    Status SparseSoftmaxCrossEntropyWithLogitsModel(
        AbstractContext* ctx, absl::Span<AbstractTensorHandle* const> inputs,
        absl::Span<AbstractTensorHandle*> outputs) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 28 13:53:47 UTC 2024
    - 8.3K bytes
    - Viewed (0)
  6. tensorflow/compiler/jit/mark_for_compilation_pass_test.cc

        Node* add1 =
            ops::BinaryOp("Add", unstage, b, builder.opts().WithName("add1"));
        Node* relu0 = ops::UnaryOp("Relu", add0, builder.opts().WithName("relu0"));
        ops::UnaryOp("Relu", add1, builder.opts().WithName("relu1"));
        MakeStageNode(builder, "stage", {DT_FLOAT}, {relu0});
    
        return GraphDefBuilderToGraph(builder, graph->get());
      };
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 14 10:11:10 UTC 2024
    - 79.6K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/tensorflow/tests/layout_optimization_move_transposes_end.mlir

      // CHECK: %[[TANH:[0-9]*]] = "tf.Tanh"(%arg0) {{.*}} tensor<1x4x4x8xf32>
      // CHECK: %[[RELU:[0-9]*]] = "tf.Relu"(%[[TANH]]) {{.*}} tensor<1x4x4x8xf32>
      // CHECK: %[[RES_TRANSPOSE:[0-9]*]] = "tf.Transpose"(%[[RELU]], %[[RES_PERM]])
      // CHECK: return %[[RES_TRANSPOSE]]
    
      %0 = "tf.Const"() {value = dense<[0, 3, 1, 2]> : tensor<4xi32>} : () -> tensor<4xi32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 9.5K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/lite/tests/mlir2flatbuffer/svdf_v2.mlir

    // CHECK-NEXT:         outputs: [ 5 ],
    // CHECK-NEXT:         builtin_options_type: SVDFOptions,
    // CHECK-NEXT:         builtin_options: {
    // CHECK-NEXT:           rank: 2,
    // CHECK-NEXT:           fused_activation_function: RELU
    // CHECK-NEXT:         }
    // CHECK-NEXT:       } ],
    // CHECK-NEXT:       name: "main"
    // CHECK-NEXT:     } ],
    // CHECK-NEXT:     description: "MLIR Converted.",
    // CHECK-NEXT:     buffers: [ {
    // CHECK-EMPTY:
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jul 14 16:41:28 UTC 2022
    - 3.7K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tensorflow/tests/fused_kernel_matcher.mlir

      // CHECK: %[[VAL_0:.*]] = "tf._FusedConv2D"(%arg2, %arg1, %arg0) <{data_format = "NHWC", dilations = [1, 1, 1, 1], epsilon = 0.000000e+00 : f32, explicit_paddings = [], fused_ops = ["BiasAdd", "Relu"], num_args = 1 : i64, operandSegmentSizes = array<i32: 1, 1, 1, 0>, padding = "SAME", strides = [1, 1, 1, 1], use_cudnn_on_gpu = true}> {TArgs = [f32]} : (tensor<8x32x32x3xf32>, tensor<1x1x3x128xf32>, tensor<128xf32>) -> tensor<*xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 13.2K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/lite/tests/mlir2flatbuffer/svdf.mlir

    // CHECK-NEXT:         outputs: [ 5 ],
    // CHECK-NEXT:         builtin_options_type: SVDFOptions,
    // CHECK-NEXT:         builtin_options: {
    // CHECK-NEXT:           rank: 2,
    // CHECK-NEXT:           fused_activation_function: RELU
    // CHECK-NEXT:         }
    // CHECK-NEXT:       } ],
    // CHECK-NEXT:       name: "main"
    // CHECK-NEXT:     } ],
    // CHECK-NEXT:     description: "MLIR Converted.",
    // CHECK-NEXT:     buffers: [ {
    // CHECK-EMPTY:
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jul 14 16:41:28 UTC 2022
    - 3.6K bytes
    - Viewed (0)
Back to top