- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 412 for Computation (0.2 sec)
-
tensorflow/compiler/mlir/tf2xla/api/v2/legalize_tf.cc
const std::variant<MlirToHloArgs, FunctionToHloArgs>& computation) { if (computation.index() == 1) return true; return std::get<0>(computation).rollout_state == ConfigProto::Experimental::MLIR_BRIDGE_ROLLOUT_DISABLED; } void DumpComputationInput( const std::variant<tpu::MlirToHloArgs, tpu::FunctionToHloArgs> computation) { if (!VLOG_IS_ON(2)) { return; }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 29 00:40:46 UTC 2024 - 6.8K bytes - Viewed (0) -
tensorflow/compiler/aot/test_google.cc
// clang-format on namespace tensorflow { namespace tfcompile { namespace { void zero_buffers(XlaCompiledCpuFunction* computation) { for (int i = 0; i < computation->num_args(); ++i) { memset(computation->arg_data(i), 0, computation->arg_size(i)); } } // Trivial test that runs the generated function to ensure it doesn't crash. TEST(TEST_NAME, NoCrash) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Sep 06 19:12:29 UTC 2023 - 3K bytes - Viewed (0) -
platforms/core-runtime/functional/src/main/java/org/gradle/internal/serialization/Cached.java
Registered: Wed Jun 12 18:38:38 UTC 2024 - Last Modified: Fri Sep 22 08:22:02 UTC 2023 - 2K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/tpu_resource_partitioning.mlir
%1 = "tf.ReadVariableOp"(%0) : (tensor<!tf_type.resource<tensor<i32>>>) -> tensor<i32> // CHECK: [[COMPUTATION:%.+]] = "tf_device.cluster_func"([[INPUT]]) %2 = "tf_device.cluster_func"(%1) {func = @computation, use_spmd_for_xla_partitioning = true} : (tensor<i32>) -> tensor<i32> // CHECK: [[OUTPUT:%.+]]:2 = "tf.TPUPartitionedOutputV2"([[COMPUTATION]]) // CHECK-SAME: _XlaSharding = "" // CHECK-SAME: partition_dims = []
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jan 23 23:53:20 UTC 2024 - 15.7K bytes - Viewed (0) -
tensorflow/compiler/aot/test.cc
// clang-format on namespace tensorflow { namespace tfcompile { namespace { void zero_buffers(XlaCompiledCpuFunction* computation) { for (int i = 0; i < computation->num_args(); ++i) { memset(computation->arg_data(i), 0, computation->arg_size(i)); } } // Trivial test that runs the generated function to ensure it doesn't crash. TEST(TEST_NAME, NoCrash) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Sep 06 19:12:29 UTC 2023 - 3K bytes - Viewed (0) -
tensorflow/compiler/aot/compile.cc
return true; } namespace { // Compiles the XLA computation into executable code. Status CompileXla(xla::CompileOnlyClient* client, const xla::XlaComputation& computation, const xla::cpu::CpuAotCompilationOptions& aot_opts, CompileResult* compile_result) { // Retrieves arg and result layouts from the computation. // TODO(toddw): Should we let the user choose the major/minor ordering?
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Apr 15 08:28:57 UTC 2024 - 11.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/transforms/tf2xla_rewriter_test.cc
xla::Tuple(&builder, tuple_values); TF_ASSERT_OK_AND_ASSIGN(XlaComputation computation, builder.Build()); EXPECT_EQ(computation.proto().computations_size(), 2); TF_ASSERT_OK(CreateMlirModule()); TF_ASSERT_OK_AND_ASSIGN(TupleOp root_tuple, ImportXlaComputationIntoModule(computation)); EXPECT_TRUE(root_tuple); int num_func_ops = 0;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:16:07 UTC 2024 - 11.7K bytes - Viewed (0) -
platforms/core-configuration/kotlin-dsl/src/main/kotlin/org/gradle/kotlin/dsl/concurrent/future.kt
/** * Starts and exposes the given suspending [computation] as a [Future] value. * * The [computation] executes synchronously until its first suspension point. */ internal fun <T> future(context: CoroutineContext = EmptyCoroutineContext, computation: suspend () -> T): Future<T> = FutureContinuation<T>(context).also { k -> computation.startCoroutine(completion = k) } private
Registered: Wed Jun 12 18:38:38 UTC 2024 - Last Modified: Wed Aug 02 08:06:49 UTC 2023 - 2.2K bytes - Viewed (0) -
tensorflow/compiler/jit/encapsulate_util.h
limitations under the License. ==============================================================================*/ // This file contains some utility functions for encapsulating XLA computation // in host graph and encapsulating outside compilation in XLA computation. #ifndef TENSORFLOW_COMPILER_JIT_ENCAPSULATE_UTIL_H_ #define TENSORFLOW_COMPILER_JIT_ENCAPSULATE_UTIL_H_ #include "absl/container/flat_hash_map.h"
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 06:59:07 UTC 2024 - 7.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/internal/legalize_tf_mlir.cc
/*lower_to_xla_hlo=*/true, computation, metadata, device_type, shape_determination_fns, use_tuple_args, compilation_result, custom_legalization_passes, arg_shapes, arg_core_mapping, per_core_arg_shapes); if (mlir_bridge_status.ok()) { VLOG(1) << "Successfully compiled MLIR computation to XLA HLO using MLIR " "tf2xla bridge";
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sun Apr 14 20:29:34 UTC 2024 - 6.1K bytes - Viewed (0)