- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 49 for func_10 (0.26 sec)
-
tensorflow/compiler/mlir/lite/experimental/tac/tests/get-alternative-subgraph.mlir
%0 = func.call @func_0_GPU_FLOAT(%arg0, %arg1, %arg2) {tac.interface_name = "func_0"} : (tensor<1xf32>, tensor<1xf32>, tensor<1xf32>) -> tensor<1xf32> %1 = func.call @func_1_GPU_FLOAT(%arg0, %arg3) {tac.interface_name = "func_1"} : (tensor<1xf32>, tensor<1xf32>) -> tensor<1xf32> %2 = func.call @func_2_CPU_FLOAT(%0, %1) {tac.interface_name = "func_2"} : (tensor<1xf32>, tensor<1xf32>) -> tensor<2x1xf32> func.return %2 : tensor<2x1xf32> }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 20.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/experimental/tac/tests/pick-subgraphs.mlir
%0 = func.call @func_0_GPU_FLOAT(%arg0, %arg1, %arg2) {tac.device = "GPU", tac.inference_type = "FLOAT", tac.interface_name = "func_0"} : (tensor<100xf32>, tensor<100xf32>, tensor<100xf32>) -> tensor<100xf32> %1 = func.call @func_1_GPU_FLOAT(%arg0, %arg3) {tac.device = "GPU", tac.inference_type = "FLOAT", tac.interface_name = "func_1"} : (tensor<100xf32>, tensor<100xf32>) -> tensor<100xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 24.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/experimental/tac/README.md
%1 = call @func_1_GPU_FLOAT(%arg0, %arg3) {tac.device = "GPU", tac.inference_type = "FLOAT", tac.interface_name = "func_1"} : (tensor<1xf32>, tensor<1xf32>) -> tensor<1xf32> %2 = call @func_2_CPU_FLOAT(%0, %1) {tac.device = "CPU", tac.inference_type = "FLOAT", tac.interface_name = "func_2"} : (tensor<1xf32>, tensor<1xf32>) -> tensor<2x1xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Mar 29 18:32:13 UTC 2022 - 11.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/common/uniform_quantized_types_test.cc
ASSERT_TRUE(module_op); auto func_op = module_op->lookupSymbol<func::FuncOp>("fully_quantized_add"); ASSERT_THAT(func_op, NotNull()); auto add_op_itr = func_op.getBody().op_begin<mlir::stablehlo::AddOp>(); ASSERT_THAT(add_op_itr, Ne(func_op.getBody().op_end<mlir::stablehlo::AddOp>())); EXPECT_TRUE(IsOpFullyQuantized(*add_op_itr)); }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 28.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/cluster_tf_ops_pass.cc
func::FuncOp func_op) { llvm::StringMap<FunctionMetadata> metadatas; WalkResult result = func_op.getBody().walk([&](Operation *op) { std::string op_host = GetHost(op); FunctionMetadata &func_metadata = metadatas[op_host]; func_metadata.original_name = func_op.getName(); func_metadata.insertion_point = ++Block::iterator(func_op); func_metadata.ops.push_back(op);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 13.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/stablehlo/transforms/legalize_tf_xla_call_module_to_stablehlo_pass.cc
} return failure(); } }; namespace { bool IsShloMainFuncOp(func::FuncOp func_op) { if (func_op == nullptr) { return false; } if (!func_op.getSymName().contains(kStablehloModuleDefaultEntryFuncName)) { return false; } if (func_op.getSymVisibility() == "nested" || func_op.getSymVisibility() == "private") { return false; } return true; }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jan 25 09:43:18 UTC 2024 - 10.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/insert_main_function.cc
llvm::SmallVector<Type> arg_types, result_types; for (auto func_op : module_op.getOps<func::FuncOp>()) { if (!ShouldIncludeInMainFunction(func_op)) continue; arg_types.append(func_op.getArgumentTypes().begin(), func_op.getArgumentTypes().end()); auto& return_op = func_op.getBody().getBlocks().front().back(); result_types.append(return_op.getOperandTypes().begin(),
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 16.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/merge_save_function_ops_to_main.cc
}; // Returns true iff func_op has either no Region or the body has no Blocks. bool IsFuncOpEmpty(func::FuncOp func_op) { return func_op->getNumRegions() == 0 || func_op.getBody().empty(); } // Gets the GraphOp from the function op. Returns an empty op iff it doesn't // exist. GraphOp GetGraphOpFromFuncOp(func::FuncOp func_op) { if (IsFuncOpEmpty(func_op)) return {};
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 10.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/resource_device_inference.cc
llvm::SmallDenseMap<func::FuncOp, PerFunctionResult, 4> per_function_results; llvm::SetVector<func::FuncOp> worklist; for (auto func_op : module.getOps<func::FuncOp>()) { worklist.insert(func_op); per_function_results.try_emplace( func_op, func_op, resource_alias_analysis.GetAnalysisForFunc(func_op)); } // Helper that propagates an op's recorded operand device assignments to its // called function's arguments.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Mar 03 03:47:00 UTC 2023 - 13.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/sparsecore/embedding_program_key.cc
// tf_device.replicate. Operation* FindCompilePredecessor(Operation* func_op, Operation* preprocess_op) { llvm::DenseSet<Block*> blocks = GetAllBlocksBetween(preprocess_op, func_op); llvm::DenseMap<Block*, Operation*> scope; Operation* o = preprocess_op; while (o && o != func_op) { scope[o->getBlock()] = o; o = o->getParentOp(); }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 17.4K bytes - Viewed (0)