Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 47 for _einsum (0.13 sec)

  1. tensorflow/compiler/mlir/quantization/stablehlo/python/integration_test/quantize_model_test_base.py

              ]
          )
          def einsum_without_kernel(
              self, x: core.Tensor, y: core.Tensor
          ) -> Mapping[str, core.Tensor]:
            return self._einsum(x, y)
    
          def _einsum(self, x, y):
    
            out = tensorflow.einsum(equation, x, y)
            if self._bias is not None:
              out = nn_ops.bias_add(out, self._bias)
            return {'output': out}
    
        model = EinsumModel()
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 14 06:31:57 UTC 2024
    - 18.2K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/quantization/tensorflow/python/integration_test/quantize_model_test_base.py

                  ),
              ]
          )
          def einsum_without_kernel(
              self, x: core.Tensor, y: core.Tensor
          ) -> Mapping[str, core.Tensor]:
            return self._einsum(x, y)
    
          def _einsum(self, x, y):
            if is_qat_model:
              x = array_ops.fake_quant_with_min_max_vars(
                  x,
                  min=ops.convert_to_tensor(self._min[0]),
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Mar 21 08:51:46 UTC 2024
    - 51.2K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/tensorflow/tests/einsum.mlir

    // RUN: tf-opt -split-input-file -verify-diagnostics -tf-einsum %s | FileCheck %s
    
    func.func @unary_einsum_reduce_sum_transpose(%arg0: tensor<3x4x5x6xf32>) -> tensor<3x5x4xf32> {
      %0 = "tf.Einsum"(%arg0) {T = "tfdtype$DT_FLOAT", equation = "...gse->...sg"}: (tensor<3x4x5x6xf32>) -> tensor<3x5x4xf32>
      func.return %0 : tensor<3x5x4xf32>
      // CHECK-LABEL: unary_einsum_reduce_sum_transpose
      // CHECK-DAG: %[[cst:.*]] = arith.constant dense<3> : tensor<1xi32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Jan 05 18:35:42 UTC 2024
    - 25.9K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/tensorflow/transforms/einsum.cc

    }
    
    struct EinsumDimensionNumbers {
      // Each field contains the list of dimensions appearing only in the specifed
      // arguments of the einsum op with natural ordering. For example `rhs_out`
      // contains the dimensions appearing in the RHS and the OUTPUT of the einsum
      // but not in the LHS.
      std::vector<int64_t> lhs;
      std::vector<int64_t> rhs;
      std::vector<std::tuple<int64_t, int64_t>> lhs_rhs;
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 33.3K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/tensorflow/transforms/einsum.h

    See the License for the specific language governing permissions and
    limitations under the License.
    ==============================================================================*/
    // This pass identifies patterns for certain Einsum Ops and replaces them
    // with other equivalent TF Ops.
    
    #ifndef TENSORFLOW_COMPILER_MLIR_TENSORFLOW_TRANSFORMS_EINSUM_H_
    #define TENSORFLOW_COMPILER_MLIR_TENSORFLOW_TRANSFORMS_EINSUM_H_
    
    #include <cstdint>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sat Dec 12 02:01:03 UTC 2020
    - 2.1K bytes
    - Viewed (0)
  6. tensorflow/cc/gradients/linalg_grad.cc

      // Claim: For the einsum operation z = einsum("{eq_x},{eq_y}->{eq_z}", x, y),
      //   where the equation involves only Tensor contractions, generalized traces
      //   and transposes, the input gradients are given by the vector-jacobian
      //   products (VJPs):
      //
      //     grad_wrt_x = einsum("{eq_y},{eq_z}->{eq_x}", y, grad_wrt_z)
      //     grad_wrt_y = einsum("{eq_x},{eq_z}->{eq_y}", x, grad_wrt_z}
      //
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Mar 07 23:11:54 UTC 2022
    - 20.4K bytes
    - Viewed (0)
  7. tensorflow/cc/gradients/linalg_grad_test.cc

      auto z = Einsum(scope_, {x}, "ii->");
      TensorShape z_shape({});
      RunTest({x}, {x_shape}, {z}, {z_shape});
    }
    
    TEST_F(LinalgGradTest, Einsum_TraceBroadcast) {
      TensorShape x_shape({4, 3, 3});
      Output x = Placeholder(scope_, DT_FLOAT, Placeholder::Shape(x_shape));
      // Note: In Python this could just be "ii" becuase tf.einsum normalizes the
      // equation, but c++ doesn't do that.
      auto z = Einsum(scope_, {x}, "...ii->...");
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Mar 07 23:11:54 UTC 2022
    - 5.8K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/tensorflow/tests/batchmatmul_to_einsum.mlir

    // RUN: tf-opt %s -tf-batch-matmul-to-tf-einsum | FileCheck %s
    
    func.func @test_batch_matmul_to_einsum(%arg0: tensor<1x2x3xf32>, %arg1: tensor<3x4xf32>) -> tensor<1x2x4xf32> {
      // CHECK-LABEL: test_batch_matmul_to_einsum
      // CHECK: "tf.Einsum"(%arg0, %arg1) <{equation = "...mk,...kn->...mn"}> : (tensor<1x2x3xf32>, tensor<3x4xf32>) -> tensor<1x2x4xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 3K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/quantization/tensorflow/tests/convert_tf_xla_op_to_tf_op.mlir

      func.return %0 : tensor<?x2x4x5xf32>
    }
    
    // CHECK: func @xla_dot_v2
    // CHECK: %[[einsum:.*]] = "tf.Einsum"(%arg0, %arg1) <{equation = "abc,cde->abde"}> : (tensor<?x2x3xf32>, tensor<3x4x5xf32>) -> tensor<?x2x4x5xf32>
    // CHECK: return %[[einsum]] : tensor<?x2x4x5xf32>
    
    // -----
    
    // dimension_numbers: {
    //   offset_dims: 0
    //   collapsed_slice_dims: 1
    //   start_index_map: 1
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 3.7K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/quantization/common/lift_as_function_call.td

                        "\""# func_name #"\", $0...)", returns>;
    
    // Add the second argument to the first argument, which is expected to be an
    // argument list.
    // bias(einsum(inputs), bias) --> einsum_with_bias(AppendToVector(inputs, bias))
    // Since inputs is a vector in case of einsum, we cannot use ArgumentList here.
    def AppendToVector : NativeCodeCall<"AppendToVector($0, $1)">;
    
    // The list of arguments of the composite function.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Mar 25 00:32:20 UTC 2024
    - 3.4K bytes
    - Viewed (0)
Back to top