Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 20 of 27 for hasTask (0.14 sec)

  1. tensorflow/compiler/mlir/quantization/tensorflow/passes/replace_cast_hacks_with_tf_xla_ops.cc

      auto input_shape = mlir::cast<ShapedType>(input.getType());
      auto filter_shape = mlir::cast<ShapedType>(filter.getType());
      if (!input_shape.hasRank() || input_shape.getRank() != 4 ||
          !filter_shape.hasRank() || filter_shape.getRank() != 4) {
        emitError(loc, "input and filter are expected to be 4D tensors");
        return {};
      }
    
      const int feature_group_cnt =
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 47.1K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/quantization/tensorflow/passes/prepare_lifting.cc

                             ArrayRef<int> val2_indices) {
      ShapedType val1_shape = mlir::cast<ShapedType>(val1.getType());
      ShapedType val2_shape = mlir::cast<ShapedType>(val2.getType());
      if (!val1_shape.hasRank() || !val2_shape.hasRank()) return false;
    
      int val1_result = 1;
      int val2_result = 1;
      for (auto idx : val1_indices) {
        if (idx < 0) idx = idx + val1_shape.getRank();
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 17:58:54 UTC 2024
    - 13.3K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/transforms/legalize_tf.cc

        // Matches fail when lhs|rhs|cond is unranked tensor.
        // TODO(b/176202543): Support unranked tensor.
        if (!mlir::cast<ShapedType>(lhs.getType()).hasRank() ||
            !mlir::cast<ShapedType>(rhs.getType()).hasRank() ||
            !mlir::cast<ShapedType>(cond.getType()).hasRank()) {
          return failure();
        }
    
        // Calculates symbolic broadcast shape that is only used in types.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon May 20 20:06:54 UTC 2024
    - 45.2K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/tensorflow/ir/tf_traits.h

              "requires compatible element types for all operands and results");
        }
      }
      return success();
    }
    
    inline ShapedType MergeType(ShapedType a, ShapedType b) {
      if (!a.hasRank()) {
        return b;
      }
      if (!b.hasRank()) {
        return a;
      }
      int64_t rank = a.getRank();
      SmallVector<int64_t, 4> dims;
      dims.resize(rank);
      for (int i = 0, e = rank; i != e; i++) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 12.7K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/tensorflow/transforms/layout_optimization.cc

              auto operand_type =
                  mlir::dyn_cast_or_null<ShapedType>(operand.getType());
              return result_type && operand_type && result_type.hasRank() &&
                     operand_type.hasRank() &&
                     result_type.getRank() == operand_type.getRank();
            });
        if (!is_valid_move) return;
      }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 19.3K bytes
    - Viewed (0)
  6. src/compress/flate/deflate.go

    	// stop things from getting too large.
    	maxFlateBlockTokens = 1 << 14
    	maxStoreBlockSize   = 65535
    	hashBits            = 17 // After 17 performance degrades
    	hashSize            = 1 << hashBits
    	hashMask            = (1 << hashBits) - 1
    	maxHashOffset       = 1 << 24
    
    	skipNever = math.MaxInt32
    )
    
    type compressionLevel struct {
    	level, good, lazy, nice, chain, fastSkipHashing int
    }
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri Apr 26 13:32:40 UTC 2024
    - 20.3K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/tensorflow/utils/xla_sharding_util.cc

      // Correctly set output shapes of split op output if input shape is statically
      // known.
      mlir::Type output_type;
      auto input_type = mlir::cast<mlir::TensorType>(src_input.getType());
    
      if (input_type.hasRank()) {
        if (input_type.getShape()[split_dimension] == mlir::ShapedType::kDynamic) {
          output_type = input_type;
        } else {
          auto shape = llvm::to_vector<4>(input_type.getShape());
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 22 21:28:13 UTC 2024
    - 34K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/lite/stablehlo/transforms/optimize.cc

                                         PatternRewriter &rewriter) {
      auto lhs_type = mlir::cast<ShapedType>(op.getLhs().getType());
      auto rhs_type = mlir::cast<ShapedType>(op.getRhs().getType());
      if (!lhs_type.hasRank() || !rhs_type.hasRank()) {
        return rewriter.notifyMatchFailure(op, "unsupported unranked input type");
      }
      if (lhs_type.getRank() < 1 || 2 < lhs_type.getRank() ||
          rhs_type.getRank() < 1 || 2 < rhs_type.getRank()) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 26.9K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/quantization/common/attrs_and_constraints.h

    // `ShapedType` or its rank is unknown.
    inline bool HasRankOf(Value value, const int64_t rank) {
      auto shaped_type = mlir::dyn_cast_or_null<ShapedType>(value.getType());
      return shaped_type && shaped_type.hasRank() && shaped_type.getRank() == rank;
    }
    
    // Creates a new type that has the shape from the `old_type` and the element
    // type from the `element_type`.
    Type CloneTypeWithNewElementType(Type old_type, Type element_type);
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 9.9K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/quantization/tensorflow/utils/tf_to_uniform_attribute_utils.cc

        auto output_scale_type =
            mlir::dyn_cast<ShapedType>(op->getOperand(3).getType());
        if (!output_scale_type) {
          return failure();
        }
        if (output_scale_type.hasRank() && 0 < output_scale_type.getRank()) {
          output_quantization_axis = activation_quantization_axis;
        }
      }
      // For per-axis -> per-axis requantization, input and output quantization
      // axis must be equal.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 18.7K bytes
    - Viewed (0)
Back to top