- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 17 for Computation (0.35 sec)
-
tensorflow/compiler/mlir/tf2xla/api/v1/compile_tf_graph.cc
} if (output_to_input_alias.empty()) return absl::OkStatus(); xla::HloModuleProto* module_proto = compilation_result->computation->mutable_proto(); absl::StatusOr<xla::ProgramShape> program_shape_or_status = compilation_result->computation->GetProgramShape(); TF_RET_CHECK(program_shape_or_status.ok()); xla::ProgramShape& program_shape = program_shape_or_status.value();
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 22:19:26 UTC 2024 - 14K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/passes.h
// Creates a pass that lifts operations on external resource variables from // device computation nested in `tf_device::LaunchOp` out so that resource // variable load operations are all before device computation while resource // variable store operations are all after device computation. After this pass, // device computation no longer interacts with external resource variables.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 21:18:05 UTC 2024 - 31.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/tf_passes.td
%computation = "tf_device.cluster_func"(%read) {func = @computation, use_spmd_for_xla_partitioning = true} : (tensor<i32>) -> tensor<i32> "tf.AssignVariableOp"(%partitioned_variable, %computation) : (tensor<!tf_type.resource<tensor<i32>>>, tensor<i32>) -> () return } func @computation(%arg0: tensor<i32>) -> tensor<i32> { return %arg0: tensor<i32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 21:18:05 UTC 2024 - 99.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/ir/tf_generated_ops.td
let description = [{ This operation holds a replicated output from a `tpu.replicate()` computation subgraph. Each replicated output has the same shape and type alongside the input. For example: ``` %computation = "tf.Computation"() %replicated_output:2 = "tf.TPUReplicatedOutput"(%computation) ```
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 23:24:08 UTC 2024 - 793K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/transforms/legalize_tf.cc
output_dims.insert(output_dims.begin() + axis, depth); Location loc = op.getLoc(); // The iota result is the effective output shape of the computation, // and indices must be broadcast into it. At this point, this computation // would need to be reworked quite a bit to support dynamic shapes, so // just using static broadcasting. auto index_type =
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 20:00:43 UTC 2024 - 291.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/utils/tpu_rewrite_device_util.cc
} return DeviceNameUtils::ParsedNameToString(tpu_device); } // Determine execution devices when topology and device assignment are not // defined. This is a special case where a single core computation is replicated // to every core in the mesh. TPU devices are simply added to // `execution_devices` of one replica. `num_replicas` must be 1 or the total // number of TPU devices available, and `num_cores_per_replica` must be 1.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Jun 10 20:10:40 UTC 2024 - 32.8K bytes - Viewed (0) -
android/guava-tests/test/com/google/common/util/concurrent/JSR166TestCase.java
* SMALL_DELAY_MS}, {@code MEDIUM_DELAY_MS}, {@code LONG_DELAY_MS}. The idea here is that a * SHORT is always discriminable from zero time, and always allows enough time for the small * amounts of computation (creating a thread, calling a few methods, etc) needed to reach a * timeout point. Similarly, a SMALL is always discriminable as larger than SHORT and smaller
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Mon Jun 10 19:21:11 UTC 2024 - 37.7K bytes - Viewed (0) -
RELEASE.md
* oneDNN optimizations can yield slightly different numerical results compared to when oneDNN optimizations are disabled due to floating-point round-off errors from different computation approaches and orders. * To verify if oneDNN optimizations are on, look for a message with *"oneDNN custom operations are on"* in the log. If the exact phrase is not there, it means they are off.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 23:24:08 UTC 2024 - 730.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/shape_inference.cc
// Returns true if a return type was changed. bool InferShapeForXlaCallModule(XlaCallModuleOp op); // Infers the shape of _XlaHostComputeMlir based on the host computation // module. Returns true if a return type was changed. bool InferShapeForXlaHostComputeMlir(_XlaHostComputeMlirOp op); // Infers the shape of function attached to XlaHostCompute.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Jun 08 07:28:49 UTC 2024 - 134.1K bytes - Viewed (0) -
pilot/pkg/config/kube/gateway/conversion.go
} gw, gwMap, nsReferences := convertGateways(ctx) ctx.GatewayReferences = gwMap result.Gateway = gw result.VirtualService = convertVirtualService(ctx) // Once we have gone through all route computation, we will know how many routes bound to each gateway. // Report this in the status. for _, dm := range gwMap { for _, pri := range dm { if pri.ReportAttachedRoutes != nil { pri.ReportAttachedRoutes() }
Registered: Fri Jun 14 15:00:06 UTC 2024 - Last Modified: Fri Jun 14 04:34:37 UTC 2024 - 84.7K bytes - Viewed (0)