- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 33 for computation (0.42 sec)
-
tensorflow/compiler/mlir/tf2xla/internal/passes/extract_outside_compilation.cc
} } } // Since we have the outputs from host and device computation after moving // outside compiled ops, we can create the actual parallel_execute regions. // Still, one region is for the host computation for outside compilation and // the other one is for the original Device cluster computation. mlir::tf_device::ParallelExecuteOp CreateFinalParallelExecuteOp(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Apr 30 21:25:12 UTC 2024 - 68.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/ir/tf_ops.td
let summary = "Op that compiles a computation in MLIR into a TPU program, and loads and executes it on a TPU device."; let description = [{ For the internal use of the TPU compiler. 'static_shapes' are tensors specifying the maximum dimension sizes for the tensors specified in `dynamic_operands`. 'args' are inputs to the TPU computation.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Apr 24 04:08:35 UTC 2024 - 90.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/resource_op_lifting.cc
// 2) Any resource stores except the last one. // TODO(ycao): Store-load forwarding implemented here is only correct when // computation is purely sequential (no concurrency). Need to support concurrent // computation as well. void ForwardStoreToLoad(Block* block) { // resource_handle_to_last_store_op keeps track of the most recent (last)
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 55.1K bytes - Viewed (0) -
src/time/time.go
// suggest a representation, namely using 1-1-1 00:00:00 UTC as the // epoch, and that's what we do. // // The Add and Sub computations are oblivious to the choice of epoch. // // The presentation computations - year, month, minute, and so on - all // rely heavily on division and modulus by positive constants. For // calendrical calculations we want these divisions to round down, even
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 50.7K bytes - Viewed (0) -
guava/src/com/google/common/util/concurrent/AbstractFuture.java
* java.util.concurrent.Executor) Futures.catching}. * * <p>This class implements all methods in {@code ListenableFuture}. Subclasses should provide a way * to set the result of the computation through the protected methods {@link #set(Object)}, {@link * #setFuture(ListenableFuture)} and {@link #setException(Throwable)}. Subclasses may also override
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Jun 07 22:25:23 UTC 2024 - 62.8K bytes - Viewed (1) -
android/guava/src/com/google/common/util/concurrent/AbstractFuture.java
* java.util.concurrent.Executor) Futures.catching}. * * <p>This class implements all methods in {@code ListenableFuture}. Subclasses should provide a way * to set the result of the computation through the protected methods {@link #set(Object)}, {@link * #setFuture(ListenableFuture)} and {@link #setException(Throwable)}. Subclasses may also override
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Jun 07 22:25:23 UTC 2024 - 63.1K bytes - Viewed (1) -
src/runtime/mgc.go
// 966-975. // For journal quality proofs that these steps are complete, correct, and terminate see // Hudson, R., and Moss, J.E.B. Copying Garbage Collection without stopping the world. // Concurrency and Computation: Practice and Experience 15(3-5), 2003. // // 1. GC performs sweep termination. // // a. Stop the world. This causes all Ps to reach a GC safe-point. //
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 16:25:21 UTC 2024 - 62K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/tpu_cluster_formation.mlir
// ----- // TF produces Identity ops between TPUReplicatedOutput and // TPUPartitionedOutputV2 ops. This test ensures that they are erased // and not considered within the clustered computation. It also ensures that // the expected interleaving pattern is present in the output. func.func @partitioned_outputs(%arg0: tensor<?xi32>) -> (tensor<?xi32>, tensor<?xi32>, tensor<?xi32>, tensor<?xi32>) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 22:03:30 UTC 2024 - 53.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/prepare_tf.cc
} }; // StridedSlice can have complicated attributes like begin_axis_mask, // end_axis_mask, ellipsis_axis_mask, new_axis_mask, shrink_axis_mask. These // masks will complicate the strided_slice computation logic, we can simplify // the logic by inserting a reshape op to pad the inputs so strided_slice can // be easier to handle. // // So the graph may looks like below: // original_input -> strided_slice -> output
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 28 21:49:50 UTC 2024 - 64.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/lower_tf.cc
Value input = op.getX(); TensorType original_tensor_type = mlir::cast<TensorType>(op.getX().getType()); // The approximation is not precise enough for float16. Do the computation // in float32 for that case. TensorType tensor_type = original_tensor_type; FloatType float_type = mlir::cast<FloatType>(tensor_type.getElementType()); bool needs_cast = float_type.getWidth() < 32;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 74.9K bytes - Viewed (0)