- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 23 for func_20 (0.26 sec)
-
tensorflow/compiler/mlir/lite/experimental/tac/transforms/fold_constants_to_subgraph.cc
// See the example below: // // In main: // %0 = tfl.const... // %1 = tfl.const... // %2 = call func_1(..., %0,...) // %3 = call func_2(..., %0, ..., %1...) // ... // // Then those consts will be copied into each function and replace their usage. // func_1: // %0 = tfl.const... // func_2: // %0 = tfl.const... // %1 = tfl.const... class FoldConstantsToSubgraphPass
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 6.1K bytes - Viewed (0) -
src/internal/trace/testdata/testprog/stacks.go
go func() { // func9 cv.L.Lock() cv.Wait() cv.L.Unlock() }() ln, err := net.Listen("tcp", "127.0.0.1:0") if err != nil { log.Fatalf("failed to listen: %v", err) } go func() { // func10 c, err := ln.Accept() if err != nil { log.Printf("failed to accept: %v", err) return } c.Close() }() rp, wp, err := os.Pipe() if err != nil {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 17 18:48:18 UTC 2024 - 2.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/passes/merge_fusion_with_dequantize.cc
SymbolTable symbol_table(call_op->getParentOfType<ModuleOp>()); auto func_op = dyn_cast_or_null<func::FuncOp>(symbol_table.lookup(func_name)); if (!func_op) return failure(); // The quantized fusion should have requantize and return ops at the end. auto return_op = dyn_cast_or_null<func::ReturnOp>( func_op.getRegion().getBlocks().front().getTerminator()); if (!return_op) return failure();
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 5.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/tensor_device_copy_conversion.cc
}; // Folds tf.IdentityOp and tf.IdentityNOp if op device and the argument devices // from the defining ops match. void TensorDeviceCopyConversionPass::runOnOperation() { func::FuncOp func_op = getOperation(); auto should_fold_op_func = [&func_op](const Value &arg, const StringAttr &op_device) { // In TFRT TPU, tensor transfer is handled specifically by D2H and
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 4.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/utils/nms_utils.cc
void ConvertNMSPaddedFunc::RewriteFunc() { func_->setAttr(kTFImplements, StringAttr::get(func_.getContext(), kTfNMSPadded)); Value boxes = func_.getArgument(0); Value scores = func_.getArgument(1); Value max_output_size = func_.getArgument(2); Value iou_threshold = func_.getArgument(3); Value score_threshold = func_.getArgument(4); auto output_type0 = func_.getFunctionType().getResult(0);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 8.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/optimize.cc
OpPassManager &func_pm = pm.nest<func::FuncOp>(); // First operates on the executor dialect: // - remove dead islands. // - fuse islands as much as possible. // - materialize the eventual "pass-through" ops by inlining their content. func_pm.addPass(tf_executor::CreateTFExecutorGraphPruningPass()); func_pm.addPass(tf_executor::CreateTFExecutorIslandCoarseningPass());
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 8.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/experimental/common/outline_operations.cc
return_operands.push_back(cloned_output); } function_builder.create<mlir::func::ReturnOp>(new_func.getLoc(), return_operands); ops_added.func_op = new_func; module.push_back(new_func); return new_func; } void ExtractSubgraphToFunc(const Subgraph& subgraph, OpBuilder& builder, ModuleOp& module, OpsAdded& ops_added) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 8.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfr/integration/tfr_decompose_ctx.cc
: tfr_module_(tfr_module), pm_(tfr_module_.getContext()) { mlir::OpPassManager& func_pm = pm_.nest<mlir::func::FuncOp>(); // Prepare the imported graph. func_pm.addPass(mlir::CreateExecutorDialectToFunctionalConversionPass()); // Run TFR lowering, inlining and raising to tf. func_pm.addPass(mlir::TFR::CreateDecomposeTFOpsPass(tfr_module_)); func_pm.addPass(mlir::TFR::CreateRaiseToTFOpsPass(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Apr 29 02:34:43 UTC 2024 - 9.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/passes/insert_calibration_statistics_saver.cc
// Insert CalibrationStatisticsSaverOp to the end of each region. for (auto func_op : module_op.getOps<func::FuncOp>()) { int32_t output_file_idx = 0; StringRef func_name = func_op.getSymName(); func_op.walk([&output_file_idx, &ctx, &func_name, &aggregator_ops_to_ignore, this](Operation* op) { for (Region& region : op->getRegions()) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 14 06:31:57 UTC 2024 - 7.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/internal/passes/input_lowering_metrics_pass.cc
public: void runOnOperation() override; }; void InputMetricsLoweringPass::runOnOperation() { bool has_dynamic_op = false; Operation* func_op = getOperation(); func_op->walk([&](Operation* op) { auto abstractOp = op->getRegisteredInfo(); if (!abstractOp) return WalkResult::advance(); if (mlir::mhlo::IsDynamicPadderOp(abstractOp->getTypeID())) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 01 08:55:35 UTC 2024 - 2.9K bytes - Viewed (0)