Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 20 of 24 for fusedBatchNormV3 (0.29 sec)

  1. tensorflow/compiler/mlir/tensorflow/tests/compile_mlir_util/shape-inference-after-legalization.mlir

      func.func @main(%arg0: tensor<8x16x16x64xbf16>, %arg1: tensor<64xf32>) -> (tensor<8x16x16x64xbf16>, tensor<64xf32>, tensor<64xf32>, tensor<64xf32>, tensor<64xf32>, tensor<*xf32>) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Mar 23 18:56:13 UTC 2022
    - 1.2K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/tensorflow/tests/layout_optimization_layout_assignment_to_nchw.mlir

      // CHECK: %[[ARG_PERM:.*]] = "tf.Const"()
      // CHECK-SAME: <{value = dense<[0, 3, 1, 2]> : tensor<4xi64>}>
      // CHECK: %[[ARG_TRANSPOSE:[0-9]*]] = "tf.Transpose"(%arg0, %[[ARG_PERM]])
    
      // CHECK: "tf.FusedBatchNormV3"
      // CHECK-SAME: (%[[ARG_TRANSPOSE]], %arg1, %arg1, %arg1, %arg1)
      // CHECK-SAME: data_format = "NCHW"
      // CHECK-SAME: (tensor<1x64x28x28xf32>, tensor<64xf32>,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 9K bytes
    - Viewed (0)
  3. tensorflow/cc/gradients/nn_grad_test.cc

    using ops::AvgPool3D;
    using ops::BiasAdd;
    using ops::Conv2D;
    using ops::Conv2DBackpropInput;
    using ops::DepthwiseConv2dNative;
    using ops::Elu;
    using ops::FractionalAvgPool;
    using ops::FractionalMaxPool;
    using ops::FusedBatchNormV3;
    using ops::L2Loss;
    using ops::LogSoftmax;
    using ops::LRN;
    using ops::MaxPool;
    using ops::MaxPool3D;
    using ops::MaxPoolV2;
    using ops::Placeholder;
    using ops::Relu;
    using ops::Relu6;
    using ops::Selu;
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Mar 22 20:45:22 UTC 2022
    - 15K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/tensorflow/tests/prepare_lifting.mlir

      %cst_0 = "tf.Const"() {value = dense<0.500000e+00> : tensor<2xf32>} : () -> tensor<2xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 14 03:24:59 UTC 2024
    - 33.3K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/quantization/tensorflow/passes/prepare_lifting.td

      (TF_CheckNumericsOp $arg, $msg),
      (replaceWithValue $arg)>;
    
    // Remove StopGradient op
    def RemoveStopGradient : Pat<
      (TF_StopGradientOp $arg),
      (replaceWithValue $arg)>;
    
    // Converts tf.FusedBatchNormV3 into a sequence of more primitive arithmetic
    // operations. Specifically, performs the following calculation:
    //
    //   (x - mean) * scale / sqrt(variance + epsilon) + offset
    //
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 14 03:24:59 UTC 2024
    - 8.4K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/lite/transforms/prepare_tf.cc

    // the checks on intermediate values, ideally we should find equivalent
    // checks that guarantees the resultant ops are valid.
    // The extra conditions are the broadcasting conditions.
    //
    // The pattern lower FusedBatchNormV3 to arithmetic ops.
    // Specifically, performs the following calculation:
    //
    //   (x - mean) * scale / sqrt(variance + epsilon) + offset
    //
    // Let multiplier = scale / sqrt(variance + epsilon),
    // to compute
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 28 21:49:50 UTC 2024
    - 64.6K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/tf2xla/tests/legalize-tf-with-tf2xla-hlo-importer.mlir

        %0:6 = "tf.FusedBatchNormV3"(%arg0, %arg1, %arg2, %arg3, %arg4) {T = "tfdtype$DT_FLOAT", data_format = "NHWC", epsilon = 0.001 : f32, is_training = false} : (tensor<8x8x8x8xf32>, tensor<8xf32>, tensor<8xf32>, tensor<8xf32>, tensor<8xf32>) -> (tensor<8x8x8x8xf32>, tensor<8xf32>,...
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sat Apr 06 15:32:52 UTC 2024
    - 38.6K bytes
    - Viewed (0)
  8. tensorflow/cc/gradients/nn_grad.cc

            grad_outputs->push_back(NoGradient());
            grad_outputs->push_back(NoGradient());
            return scope.status();
          },
          grad_outputs);
    }
    
    REGISTER_GRADIENT_OP("FusedBatchNormV3", FusedBatchNormV3Grad);
    
    Status Conv2DBackpropInputGrad(const Scope& scope, const Operation& op,
                                   const std::vector<Output>& grad_inputs,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 27 23:34:33 UTC 2022
    - 24.5K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tf2xla/tests/legalize-tf.mlir

    func.func @fusedBatchNormV3_training_dynamic_unsupported1(%arg0: tensor<?x?x?x?xf32>, %arg1: tensor<?xf32>, %arg2: tensor<?xf32>, %arg3: tensor<?xf32>, %arg4: tensor<?xf32>) -> (tensor<?x?x?x?xf32>) {
      // CHECK: tf.FusedBatchNormV3
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon May 06 18:46:23 UTC 2024
    - 335.5K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/quantization/tensorflow/python/integration_test/quantize_model_test.py

          self.assertTrue(self._contains_op(output_graphdef, 'XlaCallModule'))
        else:
          self.assertTrue(self._contains_quantized_function_call(output_graphdef))
        self.assertFalse(self._contains_op(output_graphdef, 'FusedBatchNormV3'))
    
      @parameterized.named_parameters(
          ('to_tf_with_int32_input_type', dtypes.int32, quant_opts_pb2.TF),
          ('to_xla_with_int32_input_type', dtypes.int32, quant_opts_pb2.XLA),
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 03:36:50 UTC 2024
    - 235.6K bytes
    - Viewed (0)
Back to top