Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 118 for conv4 (0.04 sec)

  1. tensorflow/compiler/mlir/tensorflow/tests/layout_optimization_to_nhwc.mlir

                                 %arg2: tensor<256xf32>,          // batch_norm args
                                 %arg3: tensor<7x7x3x64xf32>,    // conv filter #0
                                 %arg4: tensor<1x1x64x256xf32>   // conv filter #1
                                ) -> tensor<?x256xf32> {
    
      // This is a simplified ResNet layer that gets input in NHWC format, converts
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 7.3K bytes
    - Viewed (0)
  2. test/inline.go

    }
    
    // Ensure OCONVNOP is zero cost.
    func Conv(v uint64) uint64 { // ERROR "can inline Conv"
    	return conv2(conv2(conv2(v))) // ERROR "inlining call to (conv1|conv2)"
    }
    func conv2(v uint64) uint64 { // ERROR "can inline conv2"
    	return conv1(conv1(conv1(conv1(v)))) // ERROR "inlining call to conv1"
    }
    func conv1(v uint64) uint64 { // ERROR "can inline conv1"
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu Oct 19 23:33:25 UTC 2023
    - 11.7K bytes
    - Viewed (0)
  3. test/newinline.go

    }
    
    // Ensure OCONVNOP is zero cost.
    func Conv(v uint64) uint64 { // ERROR "can inline Conv"
    	return conv2(conv2(conv2(v))) // ERROR "inlining call to (conv1|conv2)"
    }
    func conv2(v uint64) uint64 { // ERROR "can inline conv2"
    	return conv1(conv1(conv1(conv1(v)))) // ERROR "inlining call to conv1"
    }
    func conv1(v uint64) uint64 { // ERROR "can inline conv1"
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu Nov 16 20:15:25 UTC 2023
    - 11.2K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/lite/tests/dilated-conv.mlir

      // CHECK-NEXT: [[CONV:%.*]] = "tf.Conv2D"([[INPUT]], [[FILTER]]) <{dilations = [1, 2, 2, 1], padding = "SAME", strides = [1, 1, 1, 1]}> : (tensor<1x128x128x3xf32>, tensor<5x5x3x8xf32>) -> tensor<1x128x128x8xf32>
      // CHECK-NEXT: [[RESULT:%.*]] = "tf.BiasAdd"([[CONV]], [[BIAS]]) : (tensor<1x128x128x8xf32>, tensor<8xf32>) -> tensor<1x128x128x8xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 44.7K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/lite/tests/prepare-quantize-signed.mlir

      %b2 = arith.constant dense<[1.0e-2, 2.1473647e1, -2.1473647e2]> : tensor<3xf32>
      %conv = "tfl.conv_2d"(%0, %w, %b) {
        dilation_h_factor = 1 : i32, dilation_w_factor = 1 : i32, fused_activation_function = "RELU",
        padding = "SAME", stride_h = 1 : i32, stride_w = 1 : i32
      } : (tensor<1x5x5x2xf32>, tensor<3x1x1x2xf32>, tensor<3xf32>) -> tensor<1x5x5x3xf32>
      %conv2 = "tfl.conv_2d"(%0, %w, %b2) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 18.4K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/lite/tests/prepare-quantize.mlir

      %conv1 = "tfl.conv_2d"(%1, %2, %cst) {dilation_h_factor = 1 : i32, dilation_w_factor = 1 : i32, fused_activation_function = "NONE", padding = "SAME", stride_h = 2 : i32, stride_w = 2 : i32} : (tensor<1x224x224x3xf32>, tensor<32x3x3x3xf32>, tensor<32xf32>) -> tensor<1x112x112x32xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 67.5K bytes
    - Viewed (0)
  7. src/cmd/compile/internal/walk/convert.go

    	init.Append(as)
    	return res
    }
    
    // Returns the data word (the second word) used to represent conv.X in
    // an interface.
    func dataWord(conv *ir.ConvExpr, init *ir.Nodes) ir.Node {
    	pos, n := conv.Pos(), conv.X
    	fromType := n.Type()
    
    	// If it's a pointer, it is its own representation.
    	if types.IsDirectIface(fromType) {
    		return n
    	}
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Oct 09 17:28:22 UTC 2023
    - 18.2K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/quantization/stablehlo/tests/bridge/optimize.mlir

        %zp_offset: tensor<?x2x2x1xi32>, %bias: tensor<1xi32>
      ) -> tensor<?x2x2x1xi32> {
      // CHECK-DAG: %[[conv:.*]] = mhlo.convolution
      // CHECK-DAG: %[[combined:.*]] = chlo.broadcast_add %[[zp_offset:.*]], %[[bias:.*]]
      // CHECK-DAG: %[[result:.*]] = chlo.broadcast_add %[[conv]], %[[combined]]
      // CHECK: return %[[result]]
      %0 = mhlo.convolution(%lhs, %rhs)
          dim_numbers = [b, 0, 1, f]x[0, 1, i, o]->[b, 0, 1, f],
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sat Feb 24 02:26:47 UTC 2024
    - 10.7K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/quantization/tensorflow/passes/replace_cast_hacks_with_tf_xla_ops.td

    // support non-constant weights.
    def ConvertTFConv2DToXLAConvOp : Pat<
      (TF_Conv2DOp:$conv
        (TF_SubOp (TF_CastOp $input, $truncate), $input_zp),
        (TF_CastOp (TF_IdentityOp $filter), $truncate1),
        $strides, $use_cudnn, $padding, $explicit_padding,
        IsDataFormatNHWC:$data_format, $dilations),
      (CreateXLAConvOpFromTFConv2DOp
        $input, $filter, $input_zp, $conv, $strides,
        $dilations, $padding, $explicit_padding),
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sun Dec 10 05:52:02 UTC 2023
    - 21.1K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/quantization/stablehlo/passes/bridge/optimize.td

    def IsNull : Constraint<CPred<"!$0">>;
    
    // This pattern optimizes:
    //   conv/dot_general + a + b -> conv/dot_general + (a + b)
    //   conv/dot_general - a - b -> conv/dot_general - (a + b)
    //   conv/dot_general + a - b -> conv/dot_general + (a - b)
    //   conv/dot_general - a + b -> conv/dot_general - (a - b)
    foreach OpsTuple = [
        [CHLO_BroadcastAddOp, CHLO_BroadcastAddOp, CHLO_BroadcastAddOp],
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sat Feb 24 02:26:47 UTC 2024
    - 2K bytes
    - Viewed (0)
Back to top