Search Options

Results per page
Sort
Preferred Languages
Advance

Results 21 - 24 of 24 for hasTask (0.36 sec)

  1. tensorflow/compiler/mlir/tfrt/saved_model/saved_model.cc

    ProcessTensorSpec(mlir::TensorType type) {
      tensorflow::DataType dtype;
      TF_RETURN_IF_ERROR(
          ConvertScalarTypeToDataType(type.getElementType(), &dtype));
    
      if (!type.hasRank())
        return std::make_pair(dtype, tensorflow::PartialTensorShape());
    
      auto shape = type.getShape();
      llvm::SmallVector<int64_t, 4> dims;
      dims.assign(shape.begin(), shape.end());
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 5.9K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/utils/nms_utils.cc

      // TODO(b/158709815): Add support for batches with padded NMS.
      auto boxes_type =
          mlir::dyn_cast<RankedTensorType>(func_.getFunctionType().getInput(0));
      if (boxes_type == nullptr || !boxes_type.hasRank() ||
          boxes_type.getRank() != 2) {
        return func_.emitWarning() << "TFLite does not support batched input for "
                                      "non_max_suppression_padded";
      }
      return success();
    }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 8.1K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/quantization/common/attrs_and_constraints.cc

      return shaped_type.hasStaticShape();
    }
    
    bool HasStaticShapeAtDims(Value value, const ArrayRef<int> dims) {
      auto shaped_type = mlir::dyn_cast<ShapedType>(value.getType());
      if (!shaped_type || !shaped_type.hasRank()) return false;
    
      for (auto dim : dims) {
        if (shaped_type.isDynamicDim(dim)) return false;
      }
      return true;
    }
    
    Type CloneTypeWithNewElementType(Type old_type, Type element_type) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 6.8K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/tensorflow/utils/fake_quant_utils.h

          return failure();
        }
    
        Value input = tf_op.getInputs();
        int quant_dim = -1;
        auto input_type = mlir::cast<ShapedType>(input.getType());
        if (PerAxis) {
          if (!input_type.hasRank()) {
            tf_op.emitError("The input should have known rank for per-channel op.");
            return failure();
          }
          // This is a special case that the quant_dim is the last dimensions.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 6.3K bytes
    - Viewed (0)
Back to top