Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 14 for reduce_mean (0.18 sec)

  1. tensorflow/compiler/mlir/tfr/examples/mnist/mnist_train.py

        with tf.GradientTape() as tape:
          logits = model(inputs)
          loss_value = tf.reduce_mean(
              tf.nn.softmax_cross_entropy_with_logits(labels, logits))
    
        grads = tape.gradient(loss_value, model.trainable_variables)
        correct_prediction = tf.equal(tf.argmax(logits, 1), tf.argmax(labels, 1))
        accuracy = tf.reduce_mean(tf.cast(correct_prediction, tf.float32))
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Oct 20 03:05:18 UTC 2021
    - 6.5K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/experimental/tac/transforms/device_transform_patterns.h

      using OpRewritePattern<TFL::MeanOp>::OpRewritePattern;
    
      LogicalResult matchAndRewrite(TFL::MeanOp mean_op,
                                    PatternRewriter& rewriter) const override;
    };
    
    // Insert Requant ops for reduce_mean.
    struct InsertRequantForReduceMean : public OpRewritePattern<TFL::MeanOp> {
      using OpRewritePattern<TFL::MeanOp>::OpRewritePattern;
    
      LogicalResult matchAndRewrite(TFL::MeanOp mean_op,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Mar 03 16:37:16 UTC 2022
    - 4.3K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/stablehlo/tests/tfl_legalize_hlo.mlir

      // CHECK:  %3 = "mhlo.broadcast_in_dim"(%2) <{broadcast_dimensions = dense<2> : tensor<1xi64>}> : (tensor<256xi32>) -> tensor<4x32x256xi32>
      // CHECK:  %cst = arith.constant dense<2> : tensor<1xi32>
      // CHECK:  %4 = "tfl.reduce_max"(%arg0, %cst) <{keep_dims = false}> : (tensor<4x32x256xf32>, tensor<1xi32>) -> tensor<4x32xf32>
      // CHECK:  %5 = "tfl.arg_max"(%arg0, %cst) : (tensor<4x32x256xf32>, tensor<1xi32>) -> tensor<4x32xi32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 40.1K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/tensorflow/python/integration_test/quantize_model_test.py

        average_min_max_output = converted_model_average_min_max.signatures[
            'serving_default'
        ](input_tensor=sample_input)['output']
    
        def get_mean_square_error(x, y):
          ret = tensorflow.reduce_mean(tensorflow.square(tensorflow.subtract(x, y)))
          try:
            ret = ret.numpy()
          except AttributeError:
            ret = ret.eval()
          return ret
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 03:36:50 UTC 2024
    - 235.6K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/lite/tests/flatbuffer2mlir/test_schema.fbs

      SUM = 74,
      SQRT = 75,
      RSQRT = 76,
      SHAPE = 77,
      POW = 78,
      ARG_MIN = 79,
      FAKE_QUANT = 80,
      REDUCE_PROD = 81,
      REDUCE_MAX = 82,
      PACK = 83,
      LOGICAL_OR = 84,
      ONE_HOT = 85,
      LOGICAL_AND = 86,
      LOGICAL_NOT = 87,
      UNPACK = 88,
      REDUCE_MIN = 89,
      FLOOR_DIV = 90,
      REDUCE_ANY = 91,
      SQUARE = 92,
      ZEROS_LIKE = 93,
      FILL = 94,
      FLOOR_MOD = 95,
      RANGE = 96,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Apr 19 19:46:06 UTC 2021
    - 26.1K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/lite/tests/legalize-tf.mlir

    }
    
    func.func @reduce_min(%arg0: tensor<8x16x16xf32>, %arg1: tensor<2xi32>) -> tensor<?xf32> {
      %0 = "tf.Min"(%arg0, %arg1) {keep_dims = false} : (tensor<8x16x16xf32>, tensor<2xi32>) -> tensor<?xf32>
      func.return %0 : tensor<?xf32>
    
      // CHECK-LABEL: reduce_min
      // CHECK: "tfl.reduce_min"(%arg0, %arg1) <{keep_dims = false}> : (tensor<8x16x16xf32>, tensor<2xi32>) -> tensor<?xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 05 01:54:33 UTC 2024
    - 153.4K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/lite/schema/schema_v3b.fbs

      SUM = 74,
      SQRT = 75,
      RSQRT = 76,
      SHAPE = 77,
      POW = 78,
      ARG_MIN = 79,
      FAKE_QUANT = 80,
      REDUCE_PROD = 81,
      REDUCE_MAX = 82,
      PACK = 83,
      LOGICAL_OR = 84,
      ONE_HOT = 85,
      LOGICAL_AND = 86,
      LOGICAL_NOT = 87,
      UNPACK = 88,
      REDUCE_MIN = 89,
      FLOOR_DIV = 90,
      REDUCE_ANY = 91,
      SQUARE = 92,
      ZEROS_LIKE = 93,
      FILL = 94,
      FLOOR_MOD = 95,
      RANGE = 96,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 28 14:28:27 UTC 2024
    - 30K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/lite/tests/optimize.mlir

      %cst_1 = arith.constant dense<[1, 128]> : tensor<2xi32>
      %0 = "tfl.reduce_min"(%arg0, %cst) {keep_dims = false} : (tensor<8x128xf32>, tensor<1xi32>) -> tensor<128xf32>
      %1 = "tfl.reshape"(%0, %cst_1) : (tensor<128xf32>, tensor<2xi32>) -> tensor<1x128xf32>
      func.return %1 : tensor<1x128xf32>
    
    // CHECK-LABEL: FoldReduceMinKeepDim
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 16 20:31:41 UTC 2024
    - 284.1K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/lite/schema/schema.fbs

      SUM = 74,
      SQRT = 75,
      RSQRT = 76,
      SHAPE = 77,
      POW = 78,
      ARG_MIN = 79,
      FAKE_QUANT = 80,
      REDUCE_PROD = 81,
      REDUCE_MAX = 82,
      PACK = 83,
      LOGICAL_OR = 84,
      ONE_HOT = 85,
      LOGICAL_AND = 86,
      LOGICAL_NOT = 87,
      UNPACK = 88,
      REDUCE_MIN = 89,
      FLOOR_DIV = 90,
      REDUCE_ANY = 91,
      SQUARE = 92,
      ZEROS_LIKE = 93,
      FILL = 94,
      FLOOR_MOD = 95,
      RANGE = 96,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 03 18:01:23 UTC 2024
    - 41.7K bytes
    - Viewed (0)
  10. tensorflow/cc/gradients/math_grad_test.cc

      // gradients when perturbing each entry in the Tensor (which then
      // changes how many minima exist.)
      // Instead, we use a single input that broadcast-multiplies a larger
      // tensor with equal values, and apply reduce_min to the multiplied
      // result.
      TensorShape x_shape({1});
      auto x = Placeholder(scope_, DT_FLOAT, Placeholder::Shape(x_shape));
      auto all_same = Mul(scope_, Const(scope_, {1.f, 1.f, 1.f}), x);
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Aug 25 18:20:20 UTC 2023
    - 36K bytes
    - Viewed (0)
Back to top