Search Options

Results per page
Sort
Preferred Languages
Advance

Results 31 - 40 of 56 for hasTask (0.14 sec)

  1. tensorflow/compiler/mlir/tensorflow/transforms/layout_optimization.cc

              auto operand_type =
                  mlir::dyn_cast_or_null<ShapedType>(operand.getType());
              return result_type && operand_type && result_type.hasRank() &&
                     operand_type.hasRank() &&
                     result_type.getRank() == operand_type.getRank();
            });
        if (!is_valid_move) return;
      }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 19.3K bytes
    - Viewed (0)
  2. src/compress/flate/deflate.go

    	// stop things from getting too large.
    	maxFlateBlockTokens = 1 << 14
    	maxStoreBlockSize   = 65535
    	hashBits            = 17 // After 17 performance degrades
    	hashSize            = 1 << hashBits
    	hashMask            = (1 << hashBits) - 1
    	maxHashOffset       = 1 << 24
    
    	skipNever = math.MaxInt32
    )
    
    type compressionLevel struct {
    	level, good, lazy, nice, chain, fastSkipHashing int
    }
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri Apr 26 13:32:40 UTC 2024
    - 20.3K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/tensorflow/utils/xla_sharding_util.cc

      // Correctly set output shapes of split op output if input shape is statically
      // known.
      mlir::Type output_type;
      auto input_type = mlir::cast<mlir::TensorType>(src_input.getType());
    
      if (input_type.hasRank()) {
        if (input_type.getShape()[split_dimension] == mlir::ShapedType::kDynamic) {
          output_type = input_type;
        } else {
          auto shape = llvm::to_vector<4>(input_type.getShape());
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 22 21:28:13 UTC 2024
    - 34K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/lite/utils/nms_utils.cc

      // TODO(b/158709815): Add support for batches with padded NMS.
      auto boxes_type =
          mlir::dyn_cast<RankedTensorType>(func_.getFunctionType().getInput(0));
      if (boxes_type == nullptr || !boxes_type.hasRank() ||
          boxes_type.getRank() != 2) {
        return func_.emitWarning() << "TFLite does not support batched input for "
                                      "non_max_suppression_padded";
      }
      return success();
    }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 8.1K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/quantization/common/attrs_and_constraints.cc

      return shaped_type.hasStaticShape();
    }
    
    bool HasStaticShapeAtDims(Value value, const ArrayRef<int> dims) {
      auto shaped_type = mlir::dyn_cast<ShapedType>(value.getType());
      if (!shaped_type || !shaped_type.hasRank()) return false;
    
      for (auto dim : dims) {
        if (shaped_type.isDynamicDim(dim)) return false;
      }
      return true;
    }
    
    Type CloneTypeWithNewElementType(Type old_type, Type element_type) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 6.8K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/quantization/tensorflow/utils/fake_quant_utils.h

          return failure();
        }
    
        Value input = tf_op.getInputs();
        int quant_dim = -1;
        auto input_type = mlir::cast<ShapedType>(input.getType());
        if (PerAxis) {
          if (!input_type.hasRank()) {
            tf_op.emitError("The input should have known rank for per-channel op.");
            return failure();
          }
          // This is a special case that the quant_dim is the last dimensions.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 6.3K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/lite/transforms/optimize.cc

    bool IsTailOfShape(Type type1, Type type2) {
      auto tail_type = mlir::dyn_cast<ShapedType>(type1);
      auto full_type = mlir::dyn_cast<ShapedType>(type2);
      if (!tail_type || !full_type || !tail_type.hasRank() ||
          !full_type.hasRank() || tail_type.getRank() > full_type.getRank())
        return false;
      auto i1 = tail_type.getShape().rbegin(), e1 = tail_type.getShape().rend();
      auto i2 = full_type.getShape().rbegin();
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Apr 30 00:40:15 UTC 2024
    - 102.3K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/tensorflow/transforms/lower_tf.cc

        auto out_size_element_ty = out_size_ty.getElementType();
    
        // Input should be rank 4.
        if (!input_ty.hasRank() || input_ty.getRank() != 4) {
          return failure();
        }
    
        // Check that out_size is rank-1, length-2. Otherwise the size is not legal.
        if (!out_size_ty.hasRank() || out_size_ty.getRank() != 1 ||
            out_size_ty.getShape()[0] != 2) {
          return failure();
        }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 74.9K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tensorflow/ir/tf_ops_a_m.cc

      // Output batch and channel dimension can be obtained using utilities from
      // tensorflow/core/util/tensor_format.h.
      if (input_ty.hasRank()) {
        return_shape[GetTensorBatchDimIndex(num_dims, format)] =
            input_ty.getDimSize(GetTensorBatchDimIndex(num_dims, format));
      }
      if (filter_ty.hasRank()) {
        return_shape[GetTensorFeatureDimIndex(num_dims, format)] =
            filter_ty.getDimSize(GetFilterTensorOutputChannelsDimIndex(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 146.7K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/lite/stablehlo/transforms/optimize.cc

                                         PatternRewriter &rewriter) {
      auto lhs_type = mlir::cast<ShapedType>(op.getLhs().getType());
      auto rhs_type = mlir::cast<ShapedType>(op.getRhs().getType());
      if (!lhs_type.hasRank() || !rhs_type.hasRank()) {
        return rewriter.notifyMatchFailure(op, "unsupported unranked input type");
      }
      if (lhs_type.getRank() < 1 || 2 < lhs_type.getRank() ||
          rhs_type.getRank() < 1 || 2 < rhs_type.getRank()) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 26.9K bytes
    - Viewed (0)
Back to top