- Sort Score
- Result 10 results
- Languages All
Results 21 - 30 of 65 for Dadd (0.48 sec)
-
tensorflow/compiler/mlir/tensorflow/transforms/prepare_tpu_computation_for_tf_export.cc
// If the parent is not a FuncOp, then add the parent op containing a region // to worklist. Operation* parent = region->getParentOp(); if (!isa<func::FuncOp>(parent)) { if (ops_with_tokens.insert(parent).second) { worklist.push_back(parent); } continue; } // For functions, get all the users and add them to the worklist.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 11.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/promote_resources_to_args.cc
function.setType(builder.getFunctionType(argument_types, result_types)); // Add resource_name attribute to the output for the resources. for (auto& resource : output_only_resources) function.setResultAttr(resource.first, kResourceNameArgAttr, builder.getStringAttr(resource.second)); // Add aliasing_output attribute to the input argument for the resources that // are updated by the function.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 17.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/prepare_lifting.cc
MLIRContext* ctx = &getContext(); auto func = getOperation(); // The pattern includes decomposing batch normalization ops, fusing add/mul // with a constant operand to a preceding affine operation. RewritePatternSet patterns(ctx); populateWithGenerated(patterns); patterns.add<RemoveIdentity, ConstantFoldQuantizableOperands>(ctx); if (op_set_ != OpSet::XLA) { // Convert Einsum into BatchMatMul for non-XLA opsets.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 17:58:54 UTC 2024 - 13.3K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_launch_util_gpu_test.cc
rollout_config.enabled_for_gpu_ = true; // Set flag to enable using XLA devices. PJRT currently is only supported // for XLA devices. GetXlaDeviceFlags()->tf_xla_enable_xla_devices = true; // Add and setup the GPU device. auto device_type = DeviceType(DEVICE_GPU); auto jit_device_type = DeviceType(DEVICE_GPU); rollout_config.AllowForDeviceInXlaLaunch(device_type);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Sep 06 19:12:29 UTC 2023 - 10K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfrt/tests/mlrt/tf_to_mlrt.mlir
func.func @branch0(%arg0: tensor<f32>, %arg1: tensor<f32>) -> tensor<f32> { %0 = "tf.Add" (%arg0, %arg1) {__op_key = 1, device = "/device:CPU:0"} : (tensor<f32>, tensor<f32>) -> tensor<f32> func.return %0 : tensor<f32> } // CHECK-LABEL: @branch1 func.func @branch1(%arg0: tensor<f32>, %arg1: tensor<f32>) -> tensor<f32> { %0 = "tf.Add" (%arg0, %arg1) {__op_key = 2, device = "/device:CPU:0"} : (tensor<f32>, tensor<f32>) -> tensor<f32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 31 20:44:15 UTC 2024 - 24.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/legalize_tf.cc
}; void addPatterns(MLIRContext* context, RewritePatternSet& patterns, bool preserve_assert_op) { // Add TF->TF lowering patterns. TF::PopulateLoweringTFPatterns(context, &patterns); // Add the generated patterns to the list. populateWithGenerated(patterns); patterns.add<ConvertTFConcatV2Op, ConvertTFBatchMatMulOp, ConvertTFBatchMatMulV2Op, ConvertTFBatchMatMulV3Op,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon May 20 20:06:54 UTC 2024 - 45.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/common/quantization_lib/quantization_config.h
case tensorflow::DT_QINT32: return 32; default: return 0; } } // Whether to add the NumericVerify ops to verify numbers before and after // quantization. bool verify_numeric = false; // Whether to add verification for layer by layer, or on whole model. When // disabled (per-layer) float and quantized ops will be run from same input
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Mar 13 10:16:19 UTC 2024 - 10.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/insert_custom_aggregation_ops.cc
std::optional<StringRef> composite_function_name = GetCompsiteFunctionName(defining_op); if (!composite_function_name.has_value()) continue; // Do not add CustomAggregatorOp after Gather since it is a weight-only // quantizable op. if (auto call_op = dyn_cast_or_null<TF::PartitionedCallOp>(defining_op)) { StringRef function_name =
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 17:58:54 UTC 2024 - 14.8K bytes - Viewed (0) -
tensorflow/cc/tools/freeze_saved_model.cc
// names. void GetTensorNamesFromTensorInfo(const TensorInfo& tensor_info, std::unordered_set<string>* tensor_names) { if (tensor_info.has_coo_sparse()) { // If the tensor is sparse we have to add all three tensors of the sparse // representations. const TensorInfo_CooSparse& coo_sparse = tensor_info.coo_sparse(); tensor_names->insert(coo_sparse.values_tensor_name());
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 11 08:05:36 UTC 2023 - 11.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/ir/tf_executor.cc
return success(); } }; // TODO(lyandy): Add canonicalization for empty IslandOps with more than one // control operand and no data results. } // anonymous namespace void IslandOp::getCanonicalizationPatterns(RewritePatternSet &results, MLIRContext *context) { results.add<DropEmptyIslandNoOperandNoDataResult,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 42.7K bytes - Viewed (0)