- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 15 for _xla (0.07 sec)
-
tensorflow/compiler/jit/BUILD
"@local_xla//xla:status_macros", "@local_xla//xla/client:local_client", "@local_xla//xla/pjrt:pjrt_client", "@local_xla//xla/pjrt:pjrt_future", "@local_xla//xla/pjrt:pjrt_stream_executor_client", "@local_xla//xla/pjrt:tracked_device_buffer", "@local_xla//xla/service:shaped_buffer", "@local_xla//xla/stream_executor:device_memory_allocator",
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 31 00:41:19 UTC 2024 - 61.5K bytes - Viewed (0) -
tensorflow/BUILD
[], otherwise = [ "@local_xla//xla/stream_executor/cuda:all_runtime", "@local_xla//xla/stream_executor/cuda:cuda_driver", "@local_xla//xla/stream_executor/cuda:cuda_platform", "@local_xla//xla/stream_executor/cuda:cudnn_plugin", "@local_xla//xla/stream_executor/cuda:cufft_plugin", "@local_xla//xla/stream_executor:cuda_platform", ], ),
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 31 16:51:59 UTC 2024 - 53.5K bytes - Viewed (0) -
.bazelrc
# --config=dbg --per_file_copt=+tensorflow/core/kernels/identity_op.*@-g # Since this .bazelrc file is synced between the tensorflow/tensorflow repo and # the openxla/xla repo, also include debug info for files under xla/. build:dbg --per_file_copt=+.*,-tensorflow.*,-xla.*@-g0 build:dbg --per_file_copt=+tensorflow/core/kernels.*@-g0 # for now, disable arm_neon. see: https://github.com/tensorflow/tensorflow/issues/33360
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 17:12:54 UTC 2024 - 52.9K bytes - Viewed (0) -
tensorflow/compiler/jit/mark_for_compilation_pass.cc
#include "tensorflow/compiler/tf2xla/const_analysis.h" #include "tensorflow/compiler/tf2xla/resource_operation_table.h" #include "tensorflow/compiler/tf2xla/xla_op_registry.h" #include "xla/service/graphcycles/graphcycles.h" #include "xla/statusor.h" #include "xla/union_find.h" #include "xla/util.h" #include "tensorflow/core/common_runtime/function.h" #include "tensorflow/core/common_runtime/graph_constructor.h" #include "tensorflow/core/framework/bounds_check.h"
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 21 12:19:41 UTC 2024 - 85.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/quantize_composite_functions.cc
llvm::cl::desc("Choose target opset."), llvm::cl::values( clEnumValN(OpSet::TF, "TF", "Uses TF ops that mimic quantization behavior"), clEnumValN(OpSet::XLA, "XLA", "Uses TF XLA ops"), clEnumValN(OpSet::UNIFORM_QUANTIZED, "UNIFORM_QUANTIZED", "Uses TF Uniform Quantized ops"))}; Option<bool> enable_per_channel_quantization_{
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 54.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/ir/tf_ops.td
} def TF_TPUPartitionedOutputOp : TF_Op<"TPUPartitionedOutput", [Pure]> { let summary = [{ An op that demultiplexes a tensor to be sharded by XLA to a list of partitioned }]; let description = [{ outputs outside the XLA computation. }]; let arguments = (ins TF_Tensor:$inputs, DefaultValuedOptionalAttr<I64Attr, "0">:$partition_dim,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Apr 24 04:08:35 UTC 2024 - 90.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/BUILD
"@llvm-project//mlir:Support", "@llvm-project//mlir:TransformUtils", "@llvm-project//mlir:Transforms", "@local_xla//xla:status", "@local_xla//xla:statusor", "@local_xla//xla/mlir_hlo", "@stablehlo//:stablehlo_ops", ], ) cc_library( name = "tensorflow_lite_optimize", srcs = [ "transforms/generated_optimize.inc", "transforms/optimize.cc",
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 21:41:49 UTC 2024 - 49.9K bytes - Viewed (0) -
tensorflow/compiler/jit/encapsulate_subgraphs_pass.cc
#include "tensorflow/compiler/jit/shape_inference_helpers.h" #include "tensorflow/compiler/jit/xla_cluster_util.h" #include "tensorflow/compiler/tf2xla/const_analysis.h" #include "xla/service/graphcycles/graphcycles.h" #include "xla/status_macros.h" #include "tensorflow/core/common_runtime/device_factory.h" #include "tensorflow/core/common_runtime/function.h" #include "tensorflow/core/common_runtime/optimization_registry.h"
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 08:47:20 UTC 2024 - 51K bytes - Viewed (0) -
configure.py
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Jun 10 04:32:44 UTC 2024 - 53.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/internal/passes/extract_outside_compilation.cc
// have a valid XLA type. LogicalResult CheckClusterResults(mlir::tf_device::ClusterOp cluster) { for (OpResult result : cluster.getResults()) { if (!tensorflow::TypeValidForXLA(result.getType())) { return cluster.emitError() << "The ExtractHeadTailOutsideCompilation pass produced a Device " "cluster with a result with a non-XLA type: " << result.getType();
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Apr 30 21:25:12 UTC 2024 - 68.3K bytes - Viewed (0)