Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 19 for computations (0.32 sec)

  1. src/cmd/compile/internal/syntax/positions.go

    // Copyright 2020 The Go Authors. All rights reserved.
    // Use of this source code is governed by a BSD-style
    // license that can be found in the LICENSE file.
    
    // This file implements helper functions for scope position computations.
    
    package syntax
    
    // StartPos returns the start position of n.
    func StartPos(n Node) Pos {
    	// Cases for nodes which don't need a correction are commented out.
    	for m := n; ; {
    		switch n := m.(type) {
    		case nil:
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Jun 10 17:49:19 UTC 2024
    - 6.5K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/tensorflow/transforms/passes.h

    // Creates a pass that lifts operations on external resource variables from
    // device computation nested in `tf_device::LaunchOp` out so that resource
    // variable load operations are all before device computation while resource
    // variable store operations are all after device computation. After this pass,
    // device computation no longer interacts with external resource variables.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 21:18:05 UTC 2024
    - 31.8K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/tensorflow/transforms/tf_passes.td

          %computation = "tf_device.cluster_func"(%read) {func = @computation, use_spmd_for_xla_partitioning = true} : (tensor<i32>) -> tensor<i32>
          "tf.AssignVariableOp"(%partitioned_variable, %computation) : (tensor<!tf_type.resource<tensor<i32>>>, tensor<i32>) -> ()
          return
        }
    
        func @computation(%arg0: tensor<i32>) -> tensor<i32> {
          return %arg0: tensor<i32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 21:18:05 UTC 2024
    - 99.6K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/tf2xla/transforms/legalize_tf.cc

        output_dims.insert(output_dims.begin() + axis, depth);
    
        Location loc = op.getLoc();
    
        // The iota result is the effective output shape of the computation,
        // and indices must be broadcast into it. At this point, this computation
        // would need to be reworked quite a bit to support dynamic shapes, so
        // just using static broadcasting.
        auto index_type =
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 20:00:43 UTC 2024
    - 291.8K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/tensorflow/ir/tf_generated_ops.td

      let description = [{
    This operation holds a replicated output from a `tpu.replicate()` computation subgraph.
    Each replicated output has the same shape and type alongside the input.
    
    For example:
    ```
    %computation = "tf.Computation"()
    %replicated_output:2 = "tf.TPUReplicatedOutput"(%computation)
    ```
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 23:24:08 UTC 2024
    - 793K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/tf2xla/api/v1/compile_tf_graph.cc

      }
    
      if (output_to_input_alias.empty()) return absl::OkStatus();
    
      xla::HloModuleProto* module_proto =
          compilation_result->computation->mutable_proto();
      absl::StatusOr<xla::ProgramShape> program_shape_or_status =
          compilation_result->computation->GetProgramShape();
      TF_RET_CHECK(program_shape_or_status.ok());
    
      xla::ProgramShape& program_shape = program_shape_or_status.value();
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 22:19:26 UTC 2024
    - 14K bytes
    - Viewed (0)
  7. RELEASE.md

    * Introducing `tf.types.experimental.AtomicFunction` as the fastest way to perform TF computations in Python.
    
        * Can be accessed through `inference_fn` property of `ConcreteFunction`s
        * Does not support gradients.
        * See `tf.types.experimental.AtomicFunction` documentation for how to call and use it.
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 23:24:08 UTC 2024
    - 730.3K bytes
    - Viewed (0)
  8. CHANGELOG/CHANGELOG-1.27.md

      - `metric_computation_duration_seconds`: The time(seconds) that the HPA controller takes to calculate one metric.
      - `metric_computation_total`: Number of metric computations. ([#116326](https://github.com/kubernetes/kubernetes/pull/116326), [@sanposhiho](https://github.com/sanposhiho)) [SIG Apps, Autoscaling and Instrumentation]
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Tue Jun 11 23:01:06 UTC 2024
    - 455.3K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tensorflow/utils/tpu_rewrite_device_util.cc

      }
      return DeviceNameUtils::ParsedNameToString(tpu_device);
    }
    
    // Determine execution devices when topology and device assignment are not
    // defined. This is a special case where a single core computation is replicated
    // to every core in the mesh. TPU devices are simply added to
    // `execution_devices` of one replica. `num_replicas` must be 1 or the total
    // number of TPU devices available, and `num_cores_per_replica` must be 1.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Jun 10 20:10:40 UTC 2024
    - 32.8K bytes
    - Viewed (0)
  10. android/guava-tests/test/com/google/common/util/concurrent/JSR166TestCase.java

     *       SMALL_DELAY_MS}, {@code MEDIUM_DELAY_MS}, {@code LONG_DELAY_MS}. The idea here is that a
     *       SHORT is always discriminable from zero time, and always allows enough time for the small
     *       amounts of computation (creating a thread, calling a few methods, etc) needed to reach a
     *       timeout point. Similarly, a SMALL is always discriminable as larger than SHORT and smaller
    Registered: Wed Jun 12 16:38:11 UTC 2024
    - Last Modified: Mon Jun 10 19:21:11 UTC 2024
    - 37.7K bytes
    - Viewed (0)
Back to top