- Sort Score
- Result 10 results
- Languages All
Results 21 - 30 of 50 for input_tensor (0.85 sec)
-
tensorflow/compiler/jit/pjrt_device_context.cc
result_tensor->SetBuffer(std::move(*buffer_or)); } pjrt_buffer->GetReadyFuture().OnReady(std::move(done)); } void PjRtDeviceContext::CopyTensorInSameDevice(const Tensor* input_tensor, Device* device, Tensor* output_tensor, StatusCallback done) const {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Apr 13 08:49:31 UTC 2024 - 11.6K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_host_send_device_context.h
Tensor* cpu_tensor, StatusCallback done) override { done(errors::Internal("host->device copy not implemented.")); } void CopyTensorInSameDevice(const Tensor* input_tensor, Device* device, Tensor* output_tensor, StatusCallback done) const override { done(errors::Internal("device->device copy not implemented.")); }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 22:46:36 UTC 2024 - 3.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfrt/runtime_fallback/runtime_fallback_executor.cc
exec_arguments.reserve(compute->num_arguments()); exec_arguments.push_back(tfrt::GetReadyChain().release()); for (const Tensor& input_tensor : arguments) { auto av = MakeAvailableAsyncValueRef<FallbackTensor>(input_tensor); exec_arguments.push_back(av.release()); } // Space for returned values. llvm::SmallVector<RCReference<AsyncValue>> results(compute->num_results());
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 08 00:18:59 UTC 2024 - 9.1K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_host_recv_device_context.h
StringPiece tensor_name, Device* device, Tensor* cpu_tensor, StatusCallback done) override; void CopyTensorInSameDevice(const Tensor* input_tensor, Device* device, Tensor* output_tensor, StatusCallback done) const override { done(errors::Internal("device->device copy not implemented.")); }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 22:46:36 UTC 2024 - 3.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/tests/quantize_xla.mlir
// RUN: tf-quant-opt %s -split-input-file -quant-lift-quantizable-spots-as-functions -quant-quantize='target-opset=XLA' -verify-each=false | FileCheck %s func.func private @conv(%input: tensor<1x3x4x3xf32> {tf._user_specified_name = "input_tensor"}) -> tensor<*xf32> attributes {tf._construction_context = "kEagerRuntime", tf._input_shapes = [#tf_type.shape<1x3x4x3>]} { %weight = arith.constant dense_resource<__elided__> : tensor<2x3x3x2xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 08 19:32:28 UTC 2024 - 11.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/ir/tfl_ops.cc
if (input_tensor.isSplat()) { return input_tensor.reshape(input_tensor.getType().cloneWith( output_shape, input_tensor.getElementType())); } // MLIR implementation pads elements < 8 bits to 8 bits and pads non byte // aligned to the nearest byte. So this is allowed. const char* raw_input = input_tensor.getRawData().data(); const int element_byte_size =
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 169.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/tests/passes/insert_calibration_statistics_saver_with_skipping.mlir
// ----- module attributes {tf.versions = {bad_consumers = [], min_consumer = 12 : i32, producer = 1836 : i32}, tf_saved_model.semantics} {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 14 06:31:57 UTC 2024 - 6.3K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_device_context.h
absl::string_view tensor_name, Device* device, Tensor* cpu_tensor, StatusCallback done) override; void CopyTensorInSameDevice(const Tensor* input_tensor, Device* device, Tensor* output_tensor, StatusCallback done) const override; xla::LocalClient* client() const { return client_; }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Sep 06 19:12:29 UTC 2023 - 5.1K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_launch_util.cc
input_mapping[alias->parameter_number] - missing_ctx_input_prefix; const Tensor input_tensor = ctx->input(tf_param).dtype() != DT_RESOURCE ? ctx->input(tf_param) : *resource_vars_snapshots.at(missing_ctx_input_prefix + tf_param); se::DeviceMemoryBase input_buffer = XlaTensor::DeviceMemoryFromTensor(input_tensor); se::DeviceMemoryBase output_buffer = output.buffer({output_num});
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 16 00:36:08 UTC 2024 - 40.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/tests/add_dump_tensor_op_stablehlo.mlir
// RUN: tf-quant-opt %s -split-input-file -quant-add-dump-tensor-op='debugger_type=float_per_layer' | FileCheck --check-prefix=FloatPerLayer %s module { func.func @matmul2(%arg0: tensor<?x2xf32> {tf_saved_model.index_path = ["input_tensor"]}) -> (tensor<?x2xf32>) { %0 = stablehlo.constant dense<[-0.211145893, -0.708605706]> : tensor<2xf32> %1 = stablehlo.constant dense<[[-0.630731344, 0.54962182], [0.180364341, -0.764542698]]> : tensor<2x2xf32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Mar 22 22:55:22 UTC 2024 - 18K bytes - Viewed (0)