Search Options

Results per page
Sort
Preferred Languages
Advance

Results 71 - 80 of 480 for MUL (0.06 sec)

  1. tensorflow/compiler/jit/tests/keras_imagenet_main_graph_mode.pbtxt

      op: "Mul"
      input: "loss_1/fc1000/bias/Regularizer/mul/x"
      input: "loss_1/ArithmeticOptimizer/HoistCommonFactor_Add_AddN"
      device: "/job:localhost/replica:0/task:0/device:GPU:0"
      attr {
        key: "T"
        value {
          type: DT_FLOAT
        }
      }
    }
    node {
      name: "bn_conv1_1/AssignMovingAvg/mul"
      op: "Mul"
      input: "bn_conv1_1/AssignMovingAvg/sub_1"
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 30 02:52:54 UTC 2019
    - 1.1M bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/quantization/tensorflow/passes/quantized_function_library.mlir

        %cast = "tf.Cast"(%accumulation) {Truncate = false} : (tensor<*xi32>) -> tensor<*xf32>
        %mul = "tf.Mul"(%cast, %rescale_factor) : (tensor<*xf32>, tensor<*xf32>) -> tensor<*xf32>
        %add = "tf.AddV2"(%mul, %float_out_zp) : (tensor<*xf32>, tensor<*xf32>) -> tensor<*xf32>
        func.return %add : tensor<*xf32>
      }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Jan 08 01:16:10 UTC 2024
    - 30.6K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/quantization/tensorflow/tests/replace_cast_hacks_with_tf_xla_ops.mlir

        %5 = "tf.AddV2"(%4, %arg2) : (tensor<1x3x2x2xi32>, tensor<2xi32>) -> tensor<1x3x2x2xi32>
        %6 = "tf.Mul"(%arg3, %arg5) : (tensor<f32>, tensor<f32>) -> tensor<f32>
        %7 = "tf.Div"(%6, %arg9) : (tensor<f32>, tensor<f32>) -> tensor<f32>
        %8 = "tf.Cast"(%5) {Truncate = false} : (tensor<1x3x2x2xi32>) -> tensor<1x3x2x2xf32>
        %9 = "tf.Mul"(%7, %8) : (tensor<f32>, tensor<1x3x2x2xf32>) -> tensor<1x3x2x2xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 81K bytes
    - Viewed (0)
  4. src/crypto/internal/bigmod/nat_riscv64.s

    	MOV	3*8(X7), X17	// x[3]
    
    	MULHU	X8, X6, X9	// z_hi[0] = x[0] * y
    	MUL	X8, X6, X8	// z_lo[0] = x[0] * y
    	ADD	X8, X10, X21	// z_lo[0] = x[0] * y + z[0]
    	SLTU	X8, X21, X22
    	ADD	X9, X22, X9	// z_hi[0] = x[0] * y + z[0]
    	ADD	X21, X29, X10	// z_lo[0] = x[0] * y + z[0] + c
    	SLTU	X21, X10, X22
    	ADD	X9, X22, X29	// next c
    
    	MULHU	X11, X6, X12	// z_hi[1] = x[1] * y
    	MUL	X11, X6, X11	// z_lo[1] = x[1] * y
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu Nov 09 13:57:06 UTC 2023
    - 2.2K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/tensorflow/tests/while_licm.mlir

      %a = "tf.Neg"(%arg0) : (tensor<4xf32>) -> tensor<4xf32>
      %b = "tf.Abs"(%arg0) : (tensor<4xf32>) -> tensor<4xf32>
      // Verify that the Div and Mul are hoisted out of the body
      // CHECK: "tf.Div"
      // CHECK: constant dense<2.200000e+01>
      // CHECK: "tf.Mul"
      // CHECK: "tf.WhileRegion"
      // Verify that Add and Sub is not hoisted out
      // CHECK: "tf.Add"
      // CHECK: "tf.Sub"
      %0:2 = "tf.WhileRegion"(%arg0, %arg1) (
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Mar 30 03:28:59 UTC 2022
    - 3K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/quantization/tensorflow/tests/propagate_quantize_type.mlir

        %2 = "tf.MatMul"(%arg0, %1) {attr_map = "0:transpose_a,1:transpose_a", device = "", transpose_a = false, transpose_b = false} : (tensor<1x2x2x2xf32>, tensor<2x1024xf32>) -> tensor<*xf32>
        %3 = "tf.Mul"(%2, %cst_0) : (tensor<*xf32>, tensor<f32>) -> tensor<*xf32>
        return %3 : tensor<*xf32>
      }
    
    // CHECK-LABEL: func @not_propagate_matmul
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 6.6K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/lite/tests/flatbuffer2mlir/pruning_function_input_as_output.mlir

    // CHECK: (%[[ARG:.*]]: tensor<4xf32>) -> (tensor<4xf32>, tensor<4xf32>, tensor<4xf32>)
    func.func @main(%arg0: tensor<4xf32>) -> tensor<4xf32> attributes {tf.entry_function = {inputs = "mul"}} {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Mar 24 07:35:24 UTC 2022
    - 1.1K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/tensorflow/tests/compile_mlir_util/convert_mhlo_quant_to_int.mlir

    // CHECK:         %[[SUB:.*]] = s32[] subtract(s32[] %[[CONVERT_1]],
    // CHECK:         %[[CONVERT_2:.*]] = f32[] convert(s32[] %[[SUB]])
    // CHECK:         %[[MUL:.*]] = f32[] multiply(f32[] %[[CONVERT_2]],
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Sep 07 16:28:50 UTC 2023
    - 1.2K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/lite/tests/mlir2flatbuffer/math.mlir

      %1 = "tfl.squared_difference"(%arg0, %0) {fused_activation_function = "NONE"} : (tensor<4xf32>, tensor<4xf32>) -> tensor<4xf32> loc("squared_difference")
      %2 = "tfl.mul"(%arg0, %1) {fused_activation_function = "NONE"} : (tensor<4xf32>, tensor<4xf32>) -> tensor<4xf32> loc("mul")
      %3 = "tfl.div"(%2, %1) {fused_activation_function = "NONE"} : (tensor<4xf32>, tensor<4xf32>) -> tensor<4xf32> loc("div")
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jul 14 16:41:28 UTC 2022
    - 5.6K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/lite/tests/optimize_no_verify.mlir

      %1 = "tfl.mul"(%0, %cst_2) {fused_activation_function = "NONE"} : (tensor<1x256xbf16>, tensor<32x1x256xbf16>) -> tensor<32x1x256xbf16>
      func.return %1 : tensor<32x1x256xbf16>
    
    // CHECK:  %[[V0:.*]] = "tfl.fully_connected"(%arg0, {{.*}}) <{{{.*}}}> : (tensor<1x10368xbf16>, tensor<256x10368xbf16>, none) -> tensor<1x256xbf16>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 5.8K bytes
    - Viewed (0)
Back to top