- Sort Score
- Result 10 results
- Languages All
Results 81 - 90 of 263 for created (0.34 sec)
-
tensorflow/compiler/mlir/lite/transforms/reduce_type_precision.cc
dyn_cast<arith::ConstantOp>(op.getOperand(0).getDefiningOp()); if (!input_op) { return failure(); } Builder builder(op.getContext()); auto new_gather_op = rewriter.create<TFL::GatherOp>( op.getLoc(), /*result=*/ mlir::cast<TensorType>(op.getResult().getType()) .clone(builder.getI4Type()), /*operand=*/op.getOperands(), op->getAttrs());
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 5.4K bytes - Viewed (0) -
tensorflow/cc/framework/gradients.cc
std::vector<Output>* grad_outputs); // Returns a list mapping whether each node in the graph is reachable // from outputs_. Keyed by node id. std::vector<bool> GetReachableNodes(); // Creates the gradient subgraph for a while loop (or just stores // `summed_grads` if not all incoming gradients are available yet). All exit // nodes (which are the first nodes of a loop encountered in the backwards
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Apr 13 05:57:22 UTC 2024 - 22K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_platform_info.cc
const std::string& profiler_name = GetPjRtDeviceCompilationProfilerResourceName(device_type); bool deleted_old_device_compiler = false; // Lookup the DeviceCompiler, create one if not found. Status s = rm->Lookup<PjRtDeviceCompiler>( rm->default_container(), compiler_name, pjrt_device_compiler); if (s.ok() && device_type == DEVICE_TPU) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 17:23:27 UTC 2024 - 17.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/set_tpu_infeed_layout.cc
} else { /* If we're not running on a TPU node, we might not be able to * actually call the part of the TPU API that gives us layout. * This happens e.g. for unit tests. Below we just create a reasonable * layout. We sort by dimension size, which makes the layout agree with * the "correct" TPU layout in surprisingly many cases. * Note that the corresponding InfeedEnqueue op will be generated
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 6.1K bytes - Viewed (0) -
tensorflow/c/eager/c_api_unified_experimental.cc
using tensorflow::tracing::TracingTensorHandle; void TF_SetTracingImplementation(const char* name, TF_Status* s) { tsl::Set_TF_Status_from_Status(s, SetDefaultTracingEngine(name)); } // Creates a new TensorFlow function, it is an execution context attached to a // given tracing context. TF_ExecutionContext* TF_CreateFunction(const char* fn_name, TF_Status* s) { return wrap(CreateTracingExecutionContext(fn_name, s));
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 09 10:15:17 UTC 2024 - 9K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/ir/tf_ops_a_m.cc
// op's type is deduced from `value`; if `value` is of scalar type, // wraps it up with a tensor type of empty shape. // TODO(jpienaar): This one differs from the autogenerated one as it takes an // attribute but always creates an ElementsAttr internally. void ConstOp::build(OpBuilder& builder, OperationState& result, Attribute value) { ShapedType type; if (auto elem_attr = mlir::dyn_cast<ElementsAttr>(value)) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 146.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/quantize_composite_functions.cc
TensorType zero_point_type = scale_type.clone(rewriter.getI32Type()); scale = rewriter.create<TF::ConstOp>( loc, scale_type, DenseFPElementsAttr::get(scale_type, {static_cast<float>(qtype.getScale())})); zero_point = rewriter.create<TF::ConstOp>( loc, zero_point_type, DenseIntElementsAttr::get(zero_point_type,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 54.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/utils/tpu_rewrite_device_util.cc
for (const auto& device : devices) if (DeviceNameUtils::IsCompleteSpecification(spec, device)) matching_devices.push_back(device); return matching_devices; } // Create error message for a conflicting attribute of a device. template <typename T> absl::Status MismatchedTPUSystemAttributeErr(absl::string_view attribute, T a, T b) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Jun 10 20:10:40 UTC 2024 - 32.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/host_runtime/tpu_metadata_utils.cc
"bad '{0}' attribute at index {1} with value '{2}': failed to parse to {3}"; constexpr char kBadArrayAttrLengthMsg[] = "bad '{0}' attribute, expected array attribute of size {1}, got size {2}"; // Creates a missing attribute error message. std::string CreateMissingAttributeMsg(llvm::StringRef attribute) { return llvm::formatv("requires attribute '{0}'", attribute).str(); }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 10.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/decompose_reduce_dataset.cc
} auto dataset_while = builder.create<TF::WhileRegionOp>( reduce_dataset.getLoc(), while_input_types, /*input=*/while_input_values, /*parallel_iterations=*/10, false, /*shape_invariant=*/false); // `_lower_using_switch_merge` is the default for While ops created // in TensorFlow and allows lowering to V1 control flow for loop // parallelization.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 14K bytes - Viewed (0)