- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 19 for computations (0.32 sec)
-
src/cmd/compile/internal/syntax/positions.go
// Copyright 2020 The Go Authors. All rights reserved. // Use of this source code is governed by a BSD-style // license that can be found in the LICENSE file. // This file implements helper functions for scope position computations. package syntax // StartPos returns the start position of n. func StartPos(n Node) Pos { // Cases for nodes which don't need a correction are commented out. for m := n; ; { switch n := m.(type) { case nil:
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Jun 10 17:49:19 UTC 2024 - 6.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/passes.h
// Creates a pass that lifts operations on external resource variables from // device computation nested in `tf_device::LaunchOp` out so that resource // variable load operations are all before device computation while resource // variable store operations are all after device computation. After this pass, // device computation no longer interacts with external resource variables.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 21:18:05 UTC 2024 - 31.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/tf_passes.td
%computation = "tf_device.cluster_func"(%read) {func = @computation, use_spmd_for_xla_partitioning = true} : (tensor<i32>) -> tensor<i32> "tf.AssignVariableOp"(%partitioned_variable, %computation) : (tensor<!tf_type.resource<tensor<i32>>>, tensor<i32>) -> () return } func @computation(%arg0: tensor<i32>) -> tensor<i32> { return %arg0: tensor<i32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 21:18:05 UTC 2024 - 99.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/transforms/legalize_tf.cc
output_dims.insert(output_dims.begin() + axis, depth); Location loc = op.getLoc(); // The iota result is the effective output shape of the computation, // and indices must be broadcast into it. At this point, this computation // would need to be reworked quite a bit to support dynamic shapes, so // just using static broadcasting. auto index_type =
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 20:00:43 UTC 2024 - 291.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/ir/tf_generated_ops.td
let description = [{ This operation holds a replicated output from a `tpu.replicate()` computation subgraph. Each replicated output has the same shape and type alongside the input. For example: ``` %computation = "tf.Computation"() %replicated_output:2 = "tf.TPUReplicatedOutput"(%computation) ```
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 23:24:08 UTC 2024 - 793K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/api/v1/compile_tf_graph.cc
} if (output_to_input_alias.empty()) return absl::OkStatus(); xla::HloModuleProto* module_proto = compilation_result->computation->mutable_proto(); absl::StatusOr<xla::ProgramShape> program_shape_or_status = compilation_result->computation->GetProgramShape(); TF_RET_CHECK(program_shape_or_status.ok()); xla::ProgramShape& program_shape = program_shape_or_status.value();
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 22:19:26 UTC 2024 - 14K bytes - Viewed (0) -
RELEASE.md
* Introducing `tf.types.experimental.AtomicFunction` as the fastest way to perform TF computations in Python. * Can be accessed through `inference_fn` property of `ConcreteFunction`s * Does not support gradients. * See `tf.types.experimental.AtomicFunction` documentation for how to call and use it.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 23:24:08 UTC 2024 - 730.3K bytes - Viewed (0) -
CHANGELOG/CHANGELOG-1.27.md
- `metric_computation_duration_seconds`: The time(seconds) that the HPA controller takes to calculate one metric. - `metric_computation_total`: Number of metric computations. ([#116326](https://github.com/kubernetes/kubernetes/pull/116326), [@sanposhiho](https://github.com/sanposhiho)) [SIG Apps, Autoscaling and Instrumentation]
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Tue Jun 11 23:01:06 UTC 2024 - 455.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/utils/tpu_rewrite_device_util.cc
} return DeviceNameUtils::ParsedNameToString(tpu_device); } // Determine execution devices when topology and device assignment are not // defined. This is a special case where a single core computation is replicated // to every core in the mesh. TPU devices are simply added to // `execution_devices` of one replica. `num_replicas` must be 1 or the total // number of TPU devices available, and `num_cores_per_replica` must be 1.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Jun 10 20:10:40 UTC 2024 - 32.8K bytes - Viewed (0) -
android/guava-tests/test/com/google/common/util/concurrent/JSR166TestCase.java
* SMALL_DELAY_MS}, {@code MEDIUM_DELAY_MS}, {@code LONG_DELAY_MS}. The idea here is that a * SHORT is always discriminable from zero time, and always allows enough time for the small * amounts of computation (creating a thread, calling a few methods, etc) needed to reach a * timeout point. Similarly, a SMALL is always discriminable as larger than SHORT and smaller
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Mon Jun 10 19:21:11 UTC 2024 - 37.7K bytes - Viewed (0)