- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 69 for func_20 (0.33 sec)
-
test/closure3.dir/main.go
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Aug 17 19:36:29 UTC 2023 - 8.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/experimental/tac/tests/get-op-cost.mlir
func.func @func_0_CPU(%arg0: tensor<256x32x32x3xf32>, %arg1: tensor<256x32x32x3xf32>) -> tensor<256x32x32x3xf32> attributes {tac.device = "CPU", tac.interface_name = "func_0"} { // CHECK: tac.cost = 7.864320e+05 %0 = "tfl.add"(%arg0, %arg1) {fused_activation_function = "RELU", tac.device = "CPU"} : (tensor<256x32x32x3xf32>, tensor<256x32x32x3xf32>) -> tensor<256x32x32x3xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Mar 24 05:29:10 UTC 2022 - 5.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/experimental/tac/transforms/fold_constants_to_subgraph.cc
// See the example below: // // In main: // %0 = tfl.const... // %1 = tfl.const... // %2 = call func_1(..., %0,...) // %3 = call func_2(..., %0, ..., %1...) // ... // // Then those consts will be copied into each function and replace their usage. // func_1: // %0 = tfl.const... // func_2: // %0 = tfl.const... // %1 = tfl.const... class FoldConstantsToSubgraphPass
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 6.1K bytes - Viewed (0) -
test/func2.go
Rob Pike <******@****.***> 1329622133 +1100
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Sun Feb 19 03:28:53 UTC 2012 - 578 bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/experimental/tac/tests/compute-cost.mlir
// CHECK: tac.cost = 7.864320e+05 func.func @func_0_CPU(%arg0: tensor<256x32x32x3xf32>, %arg1: tensor<256x32x32x3xf32>) -> tensor<256x32x32x3xf32> attributes {tac.device = "CPU", tac.interface_name = "func_0"} { %0 = "tfl.add"(%arg0, %arg1) {fused_activation_function = "RELU", tac.device = "CPU"} : (tensor<256x32x32x3xf32>, tensor<256x32x32x3xf32>) -> tensor<256x32x32x3xf32> func.return %0 : tensor<256x32x32x3xf32> }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Mar 24 05:29:10 UTC 2022 - 4.1K bytes - Viewed (0) -
src/internal/trace/testdata/testprog/stacks.go
go func() { // func9 cv.L.Lock() cv.Wait() cv.L.Unlock() }() ln, err := net.Listen("tcp", "127.0.0.1:0") if err != nil { log.Fatalf("failed to listen: %v", err) } go func() { // func10 c, err := ln.Accept() if err != nil { log.Printf("failed to accept: %v", err) return } c.Close() }() rp, wp, err := os.Pipe() if err != nil {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 17 18:48:18 UTC 2024 - 2.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/merge_duplicate_resource_ops.cc
// exist. // TODO(b/284222084): Move executor dialect utilities to a new library. GraphOp GetGraphOpFromFuncOp(func::FuncOp func_op) { if (func_op->getNumRegions() == 0 || func_op.getBody().empty()) return {}; auto graph_op_range = func_op.front().without_terminator(); if (llvm::hasSingleElement(graph_op_range)) { // The pass runs on a valid tf_executor dialect, so the op should be the // GraphOp.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 26 04:26:16 UTC 2023 - 4.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfrt/analysis/tensor_array_side_effect_analysis.cc
static bool FunctionContainsOnlyNoSideEffectOpOrTensorArrayOp( mlir::func::FuncOp func_op) { for (mlir::Operation& op : func_op.front()) { if (!mlir::isMemoryEffectFree(&op) && !IsTensorArrayOp(&op)) return false; } return true; } TensorArraySideEffectAnalysis::TensorArraySideEffectAnalysis( mlir::ModuleOp module) { for (auto func_op : module.getOps<mlir::func::FuncOp>()) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Nov 16 01:49:07 UTC 2022 - 1.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/passes/merge_fusion_with_dequantize.cc
SymbolTable symbol_table(call_op->getParentOfType<ModuleOp>()); auto func_op = dyn_cast_or_null<func::FuncOp>(symbol_table.lookup(func_name)); if (!func_op) return failure(); // The quantized fusion should have requantize and return ops at the end. auto return_op = dyn_cast_or_null<func::ReturnOp>( func_op.getRegion().getBlocks().front().getTerminator()); if (!return_op) return failure();
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 5.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/passes/unwrap_xla_call_module_op.cc
if (call_op_has_platform_index_arg) { arg_mapper.map(func_op.getArgument(0), builder.create<mhlo::ConstantOp>( func_op.getLoc(), builder.getI16IntegerAttr(0))); } for (auto [func_arg, operand] : llvm::zip_equal( func_op.getArguments().take_back(call_op.getNumOperands()), call_op.getOperands())) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Mar 05 07:39:40 UTC 2024 - 4.8K bytes - Viewed (0)