Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 23 for func_20 (0.26 sec)

  1. tensorflow/compiler/mlir/lite/experimental/tac/transforms/fold_constants_to_subgraph.cc

    // See the example below:
    //
    // In main:
    // %0 = tfl.const...
    // %1 = tfl.const...
    // %2 = call func_1(..., %0,...)
    // %3 = call func_2(..., %0, ..., %1...)
    // ...
    //
    // Then those consts will be copied into each function and replace their usage.
    // func_1:
    //   %0 = tfl.const...
    // func_2:
    //   %0 = tfl.const...
    //   %1 = tfl.const...
    class FoldConstantsToSubgraphPass
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 6.1K bytes
    - Viewed (0)
  2. src/internal/trace/testdata/testprog/stacks.go

    	go func() { // func9
    		cv.L.Lock()
    		cv.Wait()
    		cv.L.Unlock()
    	}()
    	ln, err := net.Listen("tcp", "127.0.0.1:0")
    	if err != nil {
    		log.Fatalf("failed to listen: %v", err)
    	}
    	go func() { // func10
    		c, err := ln.Accept()
    		if err != nil {
    			log.Printf("failed to accept: %v", err)
    			return
    		}
    		c.Close()
    	}()
    	rp, wp, err := os.Pipe()
    	if err != nil {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri May 17 18:48:18 UTC 2024
    - 2.4K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/quantization/stablehlo/passes/merge_fusion_with_dequantize.cc

        SymbolTable symbol_table(call_op->getParentOfType<ModuleOp>());
        auto func_op =
            dyn_cast_or_null<func::FuncOp>(symbol_table.lookup(func_name));
        if (!func_op) return failure();
        // The quantized fusion should have requantize and return ops at the end.
        auto return_op = dyn_cast_or_null<func::ReturnOp>(
            func_op.getRegion().getBlocks().front().getTerminator());
        if (!return_op) return failure();
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 5.9K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/tensorflow/transforms/tensor_device_copy_conversion.cc

    };
    
    // Folds tf.IdentityOp and tf.IdentityNOp if op device and the argument devices
    // from the defining ops match.
    void TensorDeviceCopyConversionPass::runOnOperation() {
      func::FuncOp func_op = getOperation();
    
      auto should_fold_op_func = [&func_op](const Value &arg,
                                            const StringAttr &op_device) {
        // In TFRT TPU, tensor transfer is handled specifically by D2H and
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 4.3K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/lite/utils/nms_utils.cc

    void ConvertNMSPaddedFunc::RewriteFunc() {
      func_->setAttr(kTFImplements,
                     StringAttr::get(func_.getContext(), kTfNMSPadded));
      Value boxes = func_.getArgument(0);
      Value scores = func_.getArgument(1);
      Value max_output_size = func_.getArgument(2);
      Value iou_threshold = func_.getArgument(3);
      Value score_threshold = func_.getArgument(4);
      auto output_type0 = func_.getFunctionType().getResult(0);
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 8.1K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/tensorflow/transforms/optimize.cc

      OpPassManager &func_pm = pm.nest<func::FuncOp>();
    
      // First operates on the executor dialect:
      // - remove dead islands.
      // - fuse islands as much as possible.
      // - materialize the eventual "pass-through" ops by inlining their content.
      func_pm.addPass(tf_executor::CreateTFExecutorGraphPruningPass());
      func_pm.addPass(tf_executor::CreateTFExecutorIslandCoarseningPass());
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 8.1K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/lite/experimental/common/outline_operations.cc

        return_operands.push_back(cloned_output);
      }
      function_builder.create<mlir::func::ReturnOp>(new_func.getLoc(),
                                                    return_operands);
      ops_added.func_op = new_func;
      module.push_back(new_func);
      return new_func;
    }
    
    void ExtractSubgraphToFunc(const Subgraph& subgraph, OpBuilder& builder,
                               ModuleOp& module, OpsAdded& ops_added) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 8.5K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/tfr/integration/tfr_decompose_ctx.cc

        : tfr_module_(tfr_module), pm_(tfr_module_.getContext()) {
      mlir::OpPassManager& func_pm = pm_.nest<mlir::func::FuncOp>();
    
      // Prepare the imported graph.
      func_pm.addPass(mlir::CreateExecutorDialectToFunctionalConversionPass());
    
      // Run TFR lowering, inlining and raising to tf.
      func_pm.addPass(mlir::TFR::CreateDecomposeTFOpsPass(tfr_module_));
      func_pm.addPass(mlir::TFR::CreateRaiseToTFOpsPass(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Apr 29 02:34:43 UTC 2024
    - 9.1K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/quantization/stablehlo/passes/insert_calibration_statistics_saver.cc

      // Insert CalibrationStatisticsSaverOp to the end of each region.
      for (auto func_op : module_op.getOps<func::FuncOp>()) {
        int32_t output_file_idx = 0;
        StringRef func_name = func_op.getSymName();
    
        func_op.walk([&output_file_idx, &ctx, &func_name, &aggregator_ops_to_ignore,
                      this](Operation* op) {
          for (Region& region : op->getRegions()) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 14 06:31:57 UTC 2024
    - 7.2K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/tf2xla/internal/passes/input_lowering_metrics_pass.cc

     public:
      void runOnOperation() override;
    };
    
    void InputMetricsLoweringPass::runOnOperation() {
      bool has_dynamic_op = false;
      Operation* func_op = getOperation();
    
      func_op->walk([&](Operation* op) {
        auto abstractOp = op->getRegisteredInfo();
        if (!abstractOp) return WalkResult::advance();
    
        if (mlir::mhlo::IsDynamicPadderOp(abstractOp->getTypeID())) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 01 08:55:35 UTC 2024
    - 2.9K bytes
    - Viewed (0)
Back to top