- Sort Score
- Result 10 results
- Languages All
Results 1 - 4 of 4 for zipsum (0.12 sec)
-
tensorflow/compiler/mlir/tensorflow/transforms/batchmatmul_to_einsum.cc
#include "tensorflow/compiler/mlir/tensorflow/ir/tf_ops.h" #include "tensorflow/core/util/matmul_bcast.h" namespace mlir { namespace TF { namespace { // Replace TF BatchMatMul by TF Einsum op template <typename BatchMatMulOpType> class ConvertTFBatchMatMulToEinsumOp : public OpRewritePattern<BatchMatMulOpType> { using OpRewritePattern<BatchMatMulOpType>::OpRewritePattern;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 3.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/common/lift_as_function_call.h
bool IsInLiftedFunc(Operation* op); // Checks if the op is inside a StableHLO op with region. // If the given op pointer is a nullptr, returns false. bool IsInStableHloOpRegion(Operation* op); // Checks if a given einsum op is supported for XlaDotV2 quantization. bool IsEinsumSupportedByXlaDotV2(StringAttr equation_attr); // Gets the quantization method from `op`. It is retrieved from the // `kQuantizationMethodAttr` string attribute. Returns
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 10 04:07:09 UTC 2024 - 5.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/ops/tf_op_quant_spec.cc
function_name.contains("and_bias")) { spec->biases_params[2] = {{0, 1}, quant::GetUniformQuantizedTypeForBias}; } } else if (function_name.contains("einsum")) { spec->coeff_op_quant_dim[1] = -1; if (function_name.contains("with_bias")) { spec->biases_params[2] = {{0, 1}, quant::GetUniformQuantizedTypeForBias}; }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 6.3K bytes - Viewed (0) -
build-logic/integration-testing/src/main/kotlin/gradlebuild/integrationtests/ide/AndroidStudioProvisioningPlugin.kt
Registered: Wed Jun 12 18:38:38 UTC 2024 - Last Modified: Tue May 14 15:43:39 UTC 2024 - 5.4K bytes - Viewed (0)