- Sort Score
- Result 10 results
- Languages All
Results 61 - 70 of 100 for PartitionedCall (0.47 sec)
-
tensorflow/compiler/mlir/quantization/tensorflow/tests/merge_initializer_function_ops_to_main.mlir
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Oct 31 08:49:35 UTC 2023 - 29.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/convert_tpu_model_to_cpu.cc
return failure(); } rewriter.setInsertionPointAfter(call_op); // The TPUPartitionedCall has a TPUOrdinalSelectorOp for its last argument // which should be removed. So the replaced PartitionedCall op should keep // its original arguments except for the last element. SmallVector<Value> args = call_op.getOperands().drop_back(); rewriter.replaceOpWithNewOp<TF::PartitionedCallOp>(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 5.5K bytes - Viewed (0) -
tensorflow/compiler/jit/ops/xla_ops.cc
.SetShapeFn([](InferenceContext* c) { c->set_output(0, c->input(0)); return absl::OkStatus(); }) .Doc(R"(XLA Merge Op. For use by the XLA JIT only. Merges the outputs from the PartitionedCall node and the _XlaRun node. Unlike the TensorFlow Merge op, which requires inputs of some types to be placed on the host, the _XlaMerge op can merge inputs of all types when
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Apr 06 09:08:06 UTC 2024 - 4.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/cc/saved_model_export_test.cc
tf_executor.graph { %control_0 = tf_executor.island wraps "tf.PartitionedCall"() <{config = "", config_proto = "", executor_type = "", f = @function_1}> : () -> () %control_1 = tf_executor.island wraps "tf.PartitionedCall"() <{config = "", config_proto = "", executor_type = "", f = @function_2}> : () -> ()
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Mar 20 11:11:25 UTC 2024 - 19.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/tests/unfreeze_constants.mlir
attributes {tf.entry_function = {control_outputs = "", inputs = "serving_default_input_tensor:0", outputs = "PartitionedCall:0"}, tf_saved_model.exported_names = ["serving_default"]} { %0 = "tf.PartitionedCall"(%arg0) {f = @__inference_main} : (tensor<1x5x5x1024xf32>) -> tensor<1x5x5x1024xf32> return %0 : tensor<1x5x5x1024xf32> }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 30 06:52:55 UTC 2023 - 17.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/tests/quantize.mlir
// CHECK-NEXT: [[q_bias:%.+]] = "quantfork.qcast"([[bias]]) : (tensor<2xf32>) -> tensor<2x!quant.uniform<i32:f32, 0.044022349891595126>>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 08 19:32:28 UTC 2024 - 6.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/lift_hashtable_ops_as_args.cc
return true; } } return false; } // Checks if the function is only used by supported ops. Returns false when the // function has no uses. Currently, only PartitionedCall is supported. // TODO(b/284222309): Support lifting for functions called by control flow. bool UsedBySupportedOps(ModuleOp module, func::FuncOp func) { auto function_uses =
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 17:58:54 UTC 2024 - 8.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/prepare_tpu_computation_for_tf_export.cc
// Collect all the ops that needs to have token input names attributes. These // ops are communication ops and all their parent ops via nesting or function // calls. For example, IfRegion op and PartitionedCall op. std::vector<Operation*> worklist; absl::flat_hash_set<Operation*> ops_with_tokens; module.walk([&](Operation* op) { if (IsCommunicationOp(op)) { ops_with_tokens.insert(op);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 11.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/internal/clustering_bridge_passes.cc
// Guarantee all functions have one use, which enables more exact shape // inference. pm.addPass(mlir::TF::CreateGuaranteeAllFuncsOneUsePass()); pm.addPass(mlir::TF::CreateTFShapeInferencePass()); // Encapsulate PartitionedCall ops within a cluster so that the composite // resource ops can be decomposed. pm.addPass(tensorflow::tf2xla::internal::CreateXlaClusterFormationPass());
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Apr 30 16:09:14 UTC 2024 - 11.2K bytes - Viewed (0) -
tensorflow/compiler/jit/mark_for_compilation_pass.cc
using jit::DeviceId; using jit::DeviceSet; // The clusters we create here are eventually lowered into an // _XlaCompile/_XlaRun pair with a TF executor "fallback" that uses the // PartitionedCall op to execute the cluster in the regular graph executor if // need be. PartitionedCall, however, reruns the entire TF graph optimization // pipeline over the cluster which includes this mark for compilation pass. To
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 21 12:19:41 UTC 2024 - 85.3K bytes - Viewed (0)