- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 46 for Dadd (0.03 sec)
-
tensorflow/compiler/mlir/tf2xla/internal/passes/xla_broadcast.cc
<< replicate.getN() << ")"; } // add block arg to region. This is in $body. unsigned block_arg_idx = replicate.GetNumReplicatedBlockArguments(); Block& block = replicate.GetBody(); block_arg = block.insertArgument(block_arg_idx, type, replicate.getLoc()); // add to $replicated_inputs. This also updates OperandSegmentSizeAttr.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jun 13 18:52:07 UTC 2024 - 13.9K bytes - Viewed (0) -
tensorflow/compiler/jit/extract_outside_compilation_pass_test.cc
// Build the graph: // "add" = "arg0" + "arg1" // "ret0" = "add" // "ret1" = "arg1" tensorflow::Scope s = tensorflow::Scope::NewRootScope(); Output arg0 = ops::_Arg(s.WithOpName("arg0"), DT_INT32, 0); Output arg1 = ops::_Arg(s.WithOpName("arg1"), DT_FLOAT, 1); Output arg2 = ops::_Arg(s.WithOpName("arg2"), DT_INT32, 2); Output add = ops::Add(s.WithOpName("add"), arg0, arg0);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Sep 06 19:12:29 UTC 2023 - 41K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/stablehlo/transforms/fold_broadcast_pass.cc
} void getDependentDialects(::mlir::DialectRegistry ®istry) const override {} void runOnOperation() override { RewritePatternSet patterns(&getContext()); patterns.add<FoldBroadcastInDimBeforeMulOp>(&getContext()); patterns.add(ConstantFoldMul); if (failed(applyPatternsAndFoldGreedily(getOperation(), std::move(patterns)))) { return signalPassFailure(); }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 10.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/prepare_tpu_computation_for_tf_export.cc
// If the parent is not a FuncOp, then add the parent op containing a region // to worklist. Operation* parent = region->getParentOp(); if (!isa<func::FuncOp>(parent)) { if (ops_with_tokens.insert(parent).second) { worklist.push_back(parent); } continue; } // For functions, get all the users and add them to the worklist.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 11.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/promote_resources_to_args.cc
function.setType(builder.getFunctionType(argument_types, result_types)); // Add resource_name attribute to the output for the resources. for (auto& resource : output_only_resources) function.setResultAttr(resource.first, kResourceNameArgAttr, builder.getStringAttr(resource.second)); // Add aliasing_output attribute to the input argument for the resources that // are updated by the function.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 17.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/prepare_lifting.cc
MLIRContext* ctx = &getContext(); auto func = getOperation(); // The pattern includes decomposing batch normalization ops, fusing add/mul // with a constant operand to a preceding affine operation. RewritePatternSet patterns(ctx); populateWithGenerated(patterns); patterns.add<RemoveIdentity, ConstantFoldQuantizableOperands>(ctx); if (op_set_ != OpSet::XLA) { // Convert Einsum into BatchMatMul for non-XLA opsets.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 17:58:54 UTC 2024 - 13.3K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_launch_util_gpu_test.cc
rollout_config.enabled_for_gpu_ = true; // Set flag to enable using XLA devices. PJRT currently is only supported // for XLA devices. GetXlaDeviceFlags()->tf_xla_enable_xla_devices = true; // Add and setup the GPU device. auto device_type = DeviceType(DEVICE_GPU); auto jit_device_type = DeviceType(DEVICE_GPU); rollout_config.AllowForDeviceInXlaLaunch(device_type);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Sep 06 19:12:29 UTC 2023 - 10K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/legalize_tf.cc
}; void addPatterns(MLIRContext* context, RewritePatternSet& patterns, bool preserve_assert_op) { // Add TF->TF lowering patterns. TF::PopulateLoweringTFPatterns(context, &patterns); // Add the generated patterns to the list. populateWithGenerated(patterns); patterns.add<ConvertTFConcatV2Op, ConvertTFBatchMatMulOp, ConvertTFBatchMatMulV2Op, ConvertTFBatchMatMulV3Op,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon May 20 20:06:54 UTC 2024 - 45.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/insert_custom_aggregation_ops.cc
std::optional<StringRef> composite_function_name = GetCompsiteFunctionName(defining_op); if (!composite_function_name.has_value()) continue; // Do not add CustomAggregatorOp after Gather since it is a weight-only // quantizable op. if (auto call_op = dyn_cast_or_null<TF::PartitionedCallOp>(defining_op)) { StringRef function_name =
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 17:58:54 UTC 2024 - 14.8K bytes - Viewed (0) -
tensorflow/cc/tools/freeze_saved_model.cc
// names. void GetTensorNamesFromTensorInfo(const TensorInfo& tensor_info, std::unordered_set<string>* tensor_names) { if (tensor_info.has_coo_sparse()) { // If the tensor is sparse we have to add all three tensors of the sparse // representations. const TensorInfo_CooSparse& coo_sparse = tensor_info.coo_sparse(); tensor_names->insert(coo_sparse.values_tensor_name());
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 11 08:05:36 UTC 2023 - 11.9K bytes - Viewed (0)