Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 10 for PartitionedCall (0.16 sec)

  1. tensorflow/compiler/mlir/lite/stablehlo/tests/legalize-skip-partitioned-calls.mlir

        %1 = "tf.PartitionedCall"(%0) <{
          config = "", config_proto = "", executor_type = "", f = @some_other_func
        }> {
          _collective_manager_ids = [], device = ""
        } : (tensor<1x2x2x3xf32>) -> tensor<1x2x2x3xf32>
        // CHECK-SKIP: tf.PartitionedCall
        // CHECK-NOSKIP: call @some_other_func
        // CHECK-NOSKIP-NOT: tf.PartitionedCall
        func.return %1: tensor<1x2x2x3xf32>
      }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Apr 08 20:05:12 UTC 2024
    - 1.5K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/quantization/tensorflow/tests/quantize_drq.mlir

    // -----
    
    module {
      func.func @matmul(%arg0: tensor<1x2x2x3xf32>) -> (tensor<*xf32>) {
        %cst_0 = "tf.Const"() {value = dense<0.000000e+00> : tensor<2x1024xf32>} : () -> tensor<2x1024xf32>
        %1 = "tf.PartitionedCall"(%arg0, %cst_0) {_tfl_quant_trait = "fully_quantizable", config = "", config_proto = "", executor_type = "", f = @composite_matmul_fn} : (tensor<1x2x2x3xf32>, tensor<2x1024xf32>) -> tensor<*xf32>
        func.return %1: tensor<*xf32>
      }
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 08 19:32:28 UTC 2024
    - 1.6K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/quantization/stablehlo/tests/passes/insert_calibration_statistics_saver_with_skipping.mlir

    func.func @serving_default(%arg0: tensor<1x3x4x3xf32>) -> (tensor<1x2x2x2xf32>) attributes {tf.entry_function = {control_outputs = "", inputs = "serving_default_input_tensor:0", outputs = "PartitionedCall:0"}} {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 14 06:31:57 UTC 2024
    - 6.3K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/lite/stablehlo/transforms/tf_stablehlo_pass.cc

      Option<bool> skip_partitioned_calls_{
          *this, "skip-partitioned-calls",
          ::llvm::cl::desc(
              "Skip tf.StatefulPartitionedCall and tf.PartitionedCall")};
    };
    
    void TFToMhloPass::runOnOperation() {
      auto func = getOperation();
      MLIRContext *context = func->getContext();
    
      RewritePatternSet patterns(context);
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 28 21:49:50 UTC 2024
    - 7.5K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/tensorflow/utils/attribute_utils.cc

      auto value = device_attr.getValue();
      // TODO(b/229028654): Remove string conversion once we have C++17.
      absl::string_view device_type(value.data(), value.size());
      // Device type may be empty for some ops, e.g. tf.PartitionedCall.
      auto it = std::find(kValidDeviceTypes.begin(), kValidDeviceTypes.end(),
                          device_type);
      if (it == kValidDeviceTypes.end()) return failure();
      return success();
    }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 22:03:30 UTC 2024
    - 3.8K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/tensorflow/tests/decompose_optionals.mlir

      // CHECK-NOT: Optional
      %0 = "tf.OptionalFromValue"(%arg0) : (tensor<f32>) -> tensor<!tf_type.variant<tensor<f32>>>
      %1 = "tf.PartitionedCall"(%0) {
          config = "",
          config_proto = "",
          executor_type = "",
          f = @identity} : (tensor<!tf_type.variant<tensor<f32>>>) -> tensor<!tf_type.variant<tensor<f32>>>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 21:18:05 UTC 2024
    - 4.5K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/quantization/tensorflow/passes/convert_tpu_model_to_cpu.cc

          return failure();
        }
        rewriter.setInsertionPointAfter(call_op);
    
        // The TPUPartitionedCall has a TPUOrdinalSelectorOp for its last argument
        // which should be removed. So the replaced PartitionedCall op should keep
        // its original arguments except for the last element.
        SmallVector<Value> args = call_op.getOperands().drop_back();
    
        rewriter.replaceOpWithNewOp<TF::PartitionedCallOp>(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 5.5K bytes
    - Viewed (0)
  8. tensorflow/compiler/jit/ops/xla_ops.cc

        .SetShapeFn([](InferenceContext* c) {
          c->set_output(0, c->input(0));
          return absl::OkStatus();
        })
        .Doc(R"(XLA Merge Op. For use by the XLA JIT only.
    
    Merges the outputs from the PartitionedCall node and the _XlaRun node.
    Unlike the TensorFlow Merge op, which requires inputs of some types to be
    placed on the host, the _XlaMerge op can merge inputs of all types when
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sat Apr 06 09:08:06 UTC 2024
    - 4.5K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/quantization/tensorflow/tests/quantize.mlir

    // CHECK-NEXT: [[q_bias:%.+]] = "quantfork.qcast"([[bias]]) : (tensor<2xf32>) -> tensor<2x!quant.uniform<i32:f32, 0.044022349891595126>>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 08 19:32:28 UTC 2024
    - 6.4K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/quantization/tensorflow/passes/lift_hashtable_ops_as_args.cc

          return true;
        }
      }
      return false;
    }
    
    // Checks if the function is only used by supported ops. Returns false when the
    // function has no uses. Currently, only PartitionedCall is supported.
    // TODO(b/284222309): Support lifting for functions called by control flow.
    bool UsedBySupportedOps(ModuleOp module, func::FuncOp func) {
      auto function_uses =
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 17:58:54 UTC 2024
    - 8.2K bytes
    - Viewed (0)
Back to top