- Sort Score
- Result 10 results
- Languages All
Results 41 - 50 of 65 for hasTask (0.13 sec)
-
src/compress/flate/deflate.go
// stop things from getting too large. maxFlateBlockTokens = 1 << 14 maxStoreBlockSize = 65535 hashBits = 17 // After 17 performance degrades hashSize = 1 << hashBits hashMask = (1 << hashBits) - 1 maxHashOffset = 1 << 24 skipNever = math.MaxInt32 ) type compressionLevel struct { level, good, lazy, nice, chain, fastSkipHashing int }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Apr 26 13:32:40 UTC 2024 - 20.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/utils/xla_sharding_util.cc
// Correctly set output shapes of split op output if input shape is statically // known. mlir::Type output_type; auto input_type = mlir::cast<mlir::TensorType>(src_input.getType()); if (input_type.hasRank()) { if (input_type.getShape()[split_dimension] == mlir::ShapedType::kDynamic) { output_type = input_type; } else { auto shape = llvm::to_vector<4>(input_type.getShape());
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 22 21:28:13 UTC 2024 - 34K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/utils/nms_utils.cc
// TODO(b/158709815): Add support for batches with padded NMS. auto boxes_type = mlir::dyn_cast<RankedTensorType>(func_.getFunctionType().getInput(0)); if (boxes_type == nullptr || !boxes_type.hasRank() || boxes_type.getRank() != 2) { return func_.emitWarning() << "TFLite does not support batched input for " "non_max_suppression_padded"; } return success(); }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 8.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/common/attrs_and_constraints.cc
return shaped_type.hasStaticShape(); } bool HasStaticShapeAtDims(Value value, const ArrayRef<int> dims) { auto shaped_type = mlir::dyn_cast<ShapedType>(value.getType()); if (!shaped_type || !shaped_type.hasRank()) return false; for (auto dim : dims) { if (shaped_type.isDynamicDim(dim)) return false; } return true; } Type CloneTypeWithNewElementType(Type old_type, Type element_type) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 6.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/utils/fake_quant_utils.h
return failure(); } Value input = tf_op.getInputs(); int quant_dim = -1; auto input_type = mlir::cast<ShapedType>(input.getType()); if (PerAxis) { if (!input_type.hasRank()) { tf_op.emitError("The input should have known rank for per-channel op."); return failure(); } // This is a special case that the quant_dim is the last dimensions.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 6.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/optimize.cc
bool IsTailOfShape(Type type1, Type type2) { auto tail_type = mlir::dyn_cast<ShapedType>(type1); auto full_type = mlir::dyn_cast<ShapedType>(type2); if (!tail_type || !full_type || !tail_type.hasRank() || !full_type.hasRank() || tail_type.getRank() > full_type.getRank()) return false; auto i1 = tail_type.getShape().rbegin(), e1 = tail_type.getShape().rend(); auto i2 = full_type.getShape().rbegin();
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Apr 30 00:40:15 UTC 2024 - 102.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/lower_tf.cc
auto out_size_element_ty = out_size_ty.getElementType(); // Input should be rank 4. if (!input_ty.hasRank() || input_ty.getRank() != 4) { return failure(); } // Check that out_size is rank-1, length-2. Otherwise the size is not legal. if (!out_size_ty.hasRank() || out_size_ty.getRank() != 1 || out_size_ty.getShape()[0] != 2) { return failure(); }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 74.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/ir/tf_ops_a_m.cc
// Output batch and channel dimension can be obtained using utilities from // tensorflow/core/util/tensor_format.h. if (input_ty.hasRank()) { return_shape[GetTensorBatchDimIndex(num_dims, format)] = input_ty.getDimSize(GetTensorBatchDimIndex(num_dims, format)); } if (filter_ty.hasRank()) { return_shape[GetTensorFeatureDimIndex(num_dims, format)] = filter_ty.getDimSize(GetFilterTensorOutputChannelsDimIndex(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 146.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/stablehlo/transforms/optimize.cc
PatternRewriter &rewriter) { auto lhs_type = mlir::cast<ShapedType>(op.getLhs().getType()); auto rhs_type = mlir::cast<ShapedType>(op.getRhs().getType()); if (!lhs_type.hasRank() || !rhs_type.hasRank()) { return rewriter.notifyMatchFailure(op, "unsupported unranked input type"); } if (lhs_type.getRank() < 1 || 2 < lhs_type.getRank() || rhs_type.getRank() < 1 || 2 < rhs_type.getRank()) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 26.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/common/attrs_and_constraints.h
// `ShapedType` or its rank is unknown. inline bool HasRankOf(Value value, const int64_t rank) { auto shaped_type = mlir::dyn_cast_or_null<ShapedType>(value.getType()); return shaped_type && shaped_type.hasRank() && shaped_type.getRank() == rank; } // Creates a new type that has the shape from the `old_type` and the element // type from the `element_type`. Type CloneTypeWithNewElementType(Type old_type, Type element_type);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 9.9K bytes - Viewed (0)