Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 9 of 9 for _input_shapes (0.26 sec)

  1. tensorflow/compiler/mlir/quantization/stablehlo/python/integration_test/quantize_model_test.py

          merge_fusion_with_dequantize: bool,
      ):
        lhs_dim_size, rhs_dim_size = dim_sizes
        input_shape = (*lhs_dim_size,)
        filter_shape = (*rhs_dim_size,)
        static_input_shape = [dim if dim is not None else 2 for dim in input_shape]
        model = self._create_matmul_model(
            input_shape,
            filter_shape,
            self._input_saved_model_path,
            bias_fn,
            activation_fn,
        )
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 14 06:31:57 UTC 2024
    - 51.4K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/tensorflow/transforms/lower_tf.cc

        // 2. Permute dimensions of `reshaped` to produce `permuted` of shape
        //      [batch / prod(block_shape),
        //
        //       input_shape[1], block_shape[0],
        //       ...,
        //       input_shape[M], block_shape[M-1],
        //
        //       input_shape[M+1], ..., input_shape[N-1]]
        SmallVector<int64_t> permutation(reshaped_shape.size());
        permutation[0] = block_rank;
        for (int i = 0; i < block_rank; ++i) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 74.9K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/quantization/tensorflow/python/integration_test/quantize_model_test_base.py

            """
            out = math_ops.matmul(input_tensor, self.filters, name='sample/matmul')
    
            if self.has_reshape():
              input_shape = input_tensor.shape
              if len(input_shape) == 3:
                reshape_shape = (input_shape[0], -1, self.bias_size)
              else:
                reshape_shape = (-1, self.bias_size)
    
              out = array_ops.reshape(out, reshape_shape)
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Mar 21 08:51:46 UTC 2024
    - 51.2K bytes
    - Viewed (0)
  4. tensorflow/cc/gradients/math_grad.cc

      // [[g1, g1, g1],
      //  [g2, g2, g2]]
    
      // input_shape = [2, 3]
      auto input_shape = Shape(scope, op.input(0));
    
      // output_shape_kept_dims = [2, 1]
      auto output_shape_kept_dims =
          ReducedShapeHelper(scope, input_shape, op.input(1));
    
      // This step "flips" any 1s with values from the input_shape, and
      // replaces remaining entries with 1. This creates a shape that
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Aug 25 18:20:20 UTC 2023
    - 50.7K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/lite/transforms/prepare_tf.cc

        int num_input_dims = ranked_input_type.getRank();
        SmallVector<int32_t, 4> padding_begin(num_input_dims, 0);
        auto input_shape = ranked_input_type.getShape();
        SmallVector<int32_t, 4> padding_end(input_shape.begin(), input_shape.end());
        SmallVector<int32_t, 4> padding_strides(num_input_dims, 1);
    
        int begin_mask = strided_slice_op.getBeginMask();
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 28 21:49:50 UTC 2024
    - 64.6K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/lite/transforms/lower_static_tensor_list.cc

            size_diff, scalar_zero);
    
        // Build the argument/result types for if branch function.
        auto input_shape = rewriter.create<TF::ShapeOp>(
            loc, tensorflow::GetTypeFromTFTensorShape({-1}, shape_dtype),
            input_handle);
    
        Type branch_args_type[] = {input_handle.getType(), input_shape.getType(),
                                   size_diff.getType(), size.getType()};
        Type branch_result_type[] = {result_type};
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 20:00:43 UTC 2024
    - 70.7K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/quantization/tensorflow/passes/quantize_composite_functions.cc

        llvm::ArrayRef<int64_t> input_shape = input_type.getShape();
    
        // If weight_shape[2] != 1, it means weight shape was already restored.
        if (weight_shape[2] != 1) return failure();
    
        // Weight was reshaped into [H, W, 1, InxMul].
        // Since we know in_channels from input_shape, we can derive multiplier.
        int64_t in_channels = input_shape[3];
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 54.5K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/quantization/tensorflow/tests/replace_cast_hacks_with_tf_xla_ops.mlir

    // CHECK-DAG: %[[filter:.*]] = "tf.Const"() <{value = dense<2> : tensor<2x3x3x2xi8>}> {device = ""} : () -> tensor<2x3x3x2xi8>
    // CHECK-DAG: %[[input_shape:.*]] = "tf.Shape"({{.*}}) : (tensor<?x?x?x3xi8>) -> tensor<4xi32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 81K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tensorflow/tests/lower_tf.mlir

      // CHECK-DAG: [[PADDINGS_SUM:%.+]] = "tf.AddV2"([[PADDINGS]]#0, [[PADDINGS]]#1)
      // CHECK-DAG: [[INPUT_SHAPE:%.+]] = "tf.Const"() <{value = dense<[3, 5, 7, 10]> : tensor<4xi64>}>
      // CHECK-DAG: [[PADDED_SHAPE:%.+]] = "tf.AddV2"([[PADDINGS_SUM]], [[INPUT_SHAPE]])
      // CHECK-DAG: [[PADDED_SHAPE_SPLITS:%.+]]:4 = "tf.Split"([[ZERO_I32]], [[PADDED_SHAPE]])
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Jan 05 18:35:42 UTC 2024
    - 92K bytes
    - Viewed (0)
Back to top