- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 21 for Hinsu (0.09 sec)
-
tensorflow/compiler/mlir/tensorflow/ir/tf_executor.cc
mlir::isa<tf_type::TensorFlowRefType>(output_tensor_ty.getElementType()); for (Type operand_type : merge.getOperandTypes()) { if (mlir::isa<ControlType>(operand_type)) break; // TODO(hinsu): Update ControlOperandsAfterAllData trait to verify this // constraint. TensorType operand_tensor_ty = mlir::dyn_cast<TensorType>(operand_type); if (!operand_tensor_ty) return merge.emitOpError()
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 42.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/ir/tf_ops_a_m.cc
template <typename OpT, typename std::enable_if<llvm::is_one_of< OpT, ConcatOp, ConcatV2Op>::value>::type* = nullptr> static LogicalResult Verify(OpT op) { // TODO(hinsu): Convert variadic length attributes to derived attributes. Operation::operand_range values = op.getValues(); int axis_idx = std::is_same<OpT, ConcatOp>() ? 0 : 1; Value axis = *op.getODSOperands(axis_idx).begin();
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 146.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/transforms/legalize_tf.cc
} // Broadcasts `input` to the shape of `broadcast_to` value following // TF::BroadcastTo semantics. // // Requires that input is a ranked tensor. // // TODO(hinsu): Utilize TF::ShapeOp followed by TF::BroadcastTo once ShapeOp // supports unranked inputs in the lowering. static Value BroadcastToShapeOf(Location loc, Value input, Value broadcast_to,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 20:00:43 UTC 2024 - 291.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/prepare_tf.cc
return builder->create<TF::CastOp>(loc, type, x, truncate); } } // namespace //===----------------------------------------------------------------------===// // The actual PrepareTF Pass. // // TODO(hinsu): Add and use TensorFlow dialect ops for the ops created in this // pass. namespace { #define GEN_PASS_DEF_PREPARETFPASS #include "tensorflow/compiler/mlir/lite/transforms/passes.h.inc"
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 28 21:49:50 UTC 2024 - 64.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/ir/tf_ops_n_z.cc
// PackOp //===----------------------------------------------------------------------===// LogicalResult PackOp::verify() { PackOp op = *this; // TODO(hinsu): Convert variadic length attributes to derived attributes. Operation::operand_range values = op.getValues(); if (failed(VerifyTypesCompatibility(values, /*mask_one_dim=*/false,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 09 22:07:10 UTC 2024 - 170.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/optimize_patterns.td
$filter_width, $filter_height, TFL_AF_None)), (TFL_MaxPool2DOp $input, $padding, $stride_w, $stride_h, $filter_width, $filter_height, ActFnAttr), [(HasOneUse $pool_out)]>; } // TODO(hinsu): Also fuse ops corresponding to SIGN_BIT fused // activation functions. // Currently we're not fusing tanh, sigmoid, hard_swish and other activations // those cannot be simply translated into clamping.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 16 20:31:41 UTC 2024 - 66.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/lower_static_tensor_list.cc
TF::TensorListPushBackOp, TF::TensorListReserveOp, TF::TensorListStackOp, TF::TensorListResizeOp, TF::TensorListConcatV2Op>(); // TODO(hinsu): Use TFLite constant op for constants. target.addLegalOp<arith::ConstantOp>(); target.addLegalOp<func::FuncOp>(); target.addDynamicallyLegalOp<func::ReturnOp>(is_legal);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 20:00:43 UTC 2024 - 70.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/lower_tf.cc
LogicalResult matchAndRewrite(Operation *op, PatternRewriter &rewriter) const override { auto addn_op = cast<AddNOp>(op); // TODO(hinsu): Support variant with TensorList type. tf.AddV2 doesn't // support variant type so variant types require special handling. if (mlir::isa<VariantType>(getElementTypeOrSelf(addn_op.getType()))) return failure();
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 74.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/canonicalize.mlir
// Here, custom folder doesn't handle complex dtypes and it is folded through // the constant folding hook. // TODO(hinsu): Handle complex dtypes in the custom folder for FillOp. // CHECK-DAG: "tf.Const"() <{value = dense<(0.000000e+00,1.000000e+00)> : tensor<3x2x1xcomplex<f32>>}> : () -> tensor<*xcomplex<f32>>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 09 22:07:10 UTC 2024 - 132.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/flatbuffer_export.cc
element_type, status.status().message())), false; } return true; } // Returns true if the module holds all the invariants expected by the // Translator class. // TODO(hinsu): Now that translation is done by making a single pass over the // MLIR module, consider inlining these validation checks at the place where // these invariants are assumed instead of checking upfront.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 21:41:49 UTC 2024 - 164.5K bytes - Viewed (0)