- Sort Score
- Result 10 results
- Languages All
Results 101 - 110 of 185 for Axis (0.05 sec)
-
tensorflow/c/eager/immediate_execution_tensor_handle.h
// Returns number of elements across all dimensions. virtual Status NumElements(int64_t* num_elements) const = 0; // Returns size of specified dimension // // -1 indicates an unknown axis length; this is unreachable for most standard // ImmediateExecutionTensorHandles, but comes up for example when computing // the shape of a parallel tensor with component shapes differing across // devices.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Mar 10 21:56:24 UTC 2023 - 4.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/tests/bridge/convert-tf-quant-types.mlir
} // ----- // CHECK-LABEL: func @concat_uniform_quantize func.func @concat_uniform_quantize(%arg0: tensor<3x3xf32>, %arg1: tensor<3x3xf32>) -> tensor<6x3x!tf_type.qint8> { %axis = "tf.Const"() { value = dense<0> : tensor<i64> } : () -> tensor<i64> %scales = "tf.Const"() { value = dense<1.0> : tensor<f32> } : () -> tensor<f32> %zps = "tf.Const"() { value = dense<3> : tensor<i32> } : () -> tensor<i32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 30 06:52:55 UTC 2023 - 25.9K bytes - Viewed (0) -
tensorflow/c/experimental/ops/math_ops.cc
} // Op: Sum() // Summary: Computes the sum of elements across dimensions of a tensor. // // Description: // Reduces `input` along the dimensions given in `axis`. Unless // `keep_dims` is true, the rank of the tensor is reduced by 1 for each entry // in `axis`. If `keep_dims` is true, the reduced dimensions are retained with // length 1. Status Sum(AbstractContext* ctx, AbstractTensorHandle* const input,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 10 19:11:36 UTC 2022 - 12.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/legalize-tf.mlir
// CHECK: "tfl.unpack"(%arg0) <{axis = 1 : i32, num = 3 : i32}> : (tensor<2x3xi32>) -> (tensor<2xi32>, tensor<2xi32>, tensor<2xi32>) } func.func @unpackNegAxis(%arg0: tensor<2x3xi32>) -> tensor<2xi32> { %0:3 = "tf.Unpack"(%arg0) {axis = -1 : i64} : (tensor<2x3xi32>) -> (tensor<2xi32>, tensor<2xi32>, tensor<2xi32>) func.return %0#0 : tensor<2xi32> // CHECK-LABEL: unpackNegAxis
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 05 01:54:33 UTC 2024 - 153.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfrt/tests/analysis/cost_analysis.mlir
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Aug 14 15:35:49 UTC 2023 - 12.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/quantized_function_library_tf_drq.mlir
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Mar 03 15:43:38 UTC 2023 - 12.2K bytes - Viewed (0) -
tensorflow/cc/gradients/math_grad.cc
bool reverse; TF_RETURN_IF_ERROR(GetNodeAttr(op.node()->attrs(), "reverse", &reverse)); attrs.reverse_ = !reverse; auto axis = op.input(1); auto sum = Cumsum(scope, grad_inputs[0], axis, attrs); grad_outputs->push_back(sum.out); grad_outputs->push_back(NoGradient()); return scope.status(); } REGISTER_GRADIENT_OP("Cumsum", CumsumGrad);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Aug 25 18:20:20 UTC 2023 - 50.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/const-fold.mlir
%2 = arith.constant dense<[]> : tensor<0xi32> %3 = "tfl.concatenation"(%0, %1, %2) {axis = 0 : i32, fused_activation_function = "NONE"} : (tensor<2xi32>, tensor<2xi32>, tensor<0xi32>) -> tensor<?xi32> func.return %3 : tensor<?xi32> // CHECK: %0 = "tfl.concatenation"(%[[CST]], %[[CST]]) <{axis = 0 : i32, fused_activation_function = "NONE"}> // CHECK: return %0 : tensor<?xi32> }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 45.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/experimental/tac/execution_metadata_exporter_test.cc
%2 = "tfl.add"(%arg0, %arg3) {fused_activation_function = "RELU6", per_device_costs = {CPU = 5.0 : f32, GPU = 1.0 : f32}, tac.device = "GPU"} : (tensor<1xf32>, tensor<1xf32>) -> tensor<1xf32> %3 = "tfl.pack"(%1, %2) {axis = 0 : i32, per_device_costs = {CPU = 2.0 : f32, GPU = -1.0 : f32}, values_count = 2 : i32, tac.device = "CPU"} : (tensor<1xf32>, tensor<1xf32>) -> tensor<2x1xf32> func.return %3 : tensor<2x1xf32> })";
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 06:11:34 UTC 2024 - 6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfr/python/tfr_gen.py
return type_ def _pack_tensor_list(self, value): # This is packing a list of tensors, then the axis is 0. axis = self._ssa_name('zero') self._emit_with_loc('\n{} = arith.constant 0 : i64'.format(axis)) casted = self._ssa_name('pack') self.emit('\n{} = tfr.call @tf__pack({}, {})'.format(casted, value, axis)) self._emit_with_loc(' : (!tfr.tensor_list, i64) -> !tfr.tensor') # load the op def of tf.Pack
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Apr 27 15:27:03 UTC 2022 - 55.8K bytes - Viewed (0)