Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 163 for input_ (0.21 sec)

  1. tensorflow/compiler/mlir/tfr/examples/mnist/mnist_ops_test.py

        }
    
        self._assertOpAndComposite([input_, filter_, bias],
                                   tf.function(gen_mnist_ops.new_conv2d),
                                   ops_defs._composite_conv_add_relu, kwargs)
    
      def test_new_conv2d_relu6(self):
        input_ = tf.random.uniform([1, 4, 4, 1])
        filter_ = tf.random.uniform([2, 2, 1, 8])
        bias = tf.zeros([8])
        kwargs = {
            'input_': input_,
            'filter_': filter_,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Sep 28 21:37:05 UTC 2021
    - 4K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/tfr/examples/mnist/ops_defs.py

      ]
    
    
    @Composite(
        'NewFullyConnected',
        inputs=['input_: T', 'filter_: T', 'bias: T'],
        attrs=['act: {"", "RELU", "RELU6", "TANH"} = ""'],
        derived_attrs=['T: {float, int8}'],
        outputs=['o: T'])
    def _composite_fully_connected(input_, filter_, bias, act):
      res = tf.raw_ops.MatMul(
          a=input_, b=filter_, transpose_a=False, transpose_b=True)
      res = tf.raw_ops.Add(x=res, y=bias)
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Aug 31 20:23:51 UTC 2023
    - 6.8K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/tfr/examples/pad/pad_ops_test.py

      def test_mirror_pad(self, mode):
        input_ = tf.constant([[1, 2, 3], [4, 5, 6]], dtype=tf.float32)
        paddings = tf.constant([[
            1,
            1,
        ], [2, 2]])
        kwargs = {
            'input': input_,
            'paddings': paddings,
            'mode': mode,
        }
        kwargs_ = {
            'input_': input_,
            'paddings': paddings,
            'mode': mode,
        }
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Sep 28 21:37:05 UTC 2021
    - 3.4K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/tfr/examples/pad/ops_defs.py

              num_split=2)
    
        input_ = tf.raw_ops.Concat(
            concat_dim=i, values=[left_padding, input_, right_padding])
      return input_
    
    
    @tf.RegisterGradient('NewMirrorPad')
    def _mirror_pad_grad(op, grad):
      mode = op.get_attr('mode')
      return [gen_array_ops.mirror_pad_grad(grad, op.inputs[1], mode=mode), None]
    
    
    @Composite(
        'NewMirrorPadGrad',
        inputs=['input_: T', 'paddings: Tpaddings'],
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Oct 01 05:00:29 UTC 2021
    - 5.6K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/tfr/README.md

    ```python
    import tensorflow as tf
    
    @Composite(
        'FusedFullyConnected',
        inputs=['input_: T', 'filter_: T', 'bias: T'],
        attrs=['act: {"", "RELU", "RELU6", "TANH"} = ""'],
        derived_attrs=['T: {float, int8}'],
        outputs=['o: T'])
    def _composite_fully_connected(input_, filter_, bias, act):
      res = tf.raw_ops.MatMul(
          a=input_, b=filter_, transpose_a=False, transpose_b=True)
      res = tf.raw_ops.Add(x=res, y=bias)
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Mar 29 18:32:13 UTC 2022
    - 6.2K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/tfr/tests/decompose.mlir

      %data_format = tfr.constant "NHWC" -> !tfr.attr
      %MaxPool = tfr.call @tf__max_pool(%input_, %stride, %filter, %padding, %explicit_paddings, %data_format) : (!tfr.tensor, !tfr.attr, !tfr.attr, !tfr.attr, !tfr.attr, !tfr.attr) -> (!tfr.tensor)
      tfr.return %MaxPool : !tfr.tensor
    // CHECK: tf__max_pool
    }
    
    // CHECK-LABEL: @tf__cast_float
    tfr.func @tf__cast_float(%input_: !tfr.tensor, %out_type: !tfr.attr{tfr.name="out_type"}) -> (!tfr.tensor) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Oct 30 06:52:55 UTC 2023
    - 16.7K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/tfr/python/op_reg_gen.py

              .format(op_name, expected_args, all_func_args))
    
        cxx_reg_code = ['\nREGISTER_OP("{}")'.format(op_name)]
        for input_ in inputs:
          cxx_reg_code.append('.Input("{}")'.format(input_))
        for attr in attrs:
          py_str = attr.replace('"', "'")
          cxx_reg_code.append('.Attr("{}")'.format(py_str))
        for attr in all_dec_args.get('derived_attrs', []):
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Sep 28 21:37:05 UTC 2021
    - 5K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/lite/utils/lstm_utils.cc

      // TFL lstm only supports time-majored inputs, so if it's not time-majored,
      // we will transpose the inputs and outputs.
      auto time_major_attr = func_op->getAttrOfType<BoolAttr>("tf.time_major");
      if (time_major_attr == nullptr) return failure();
    
      bool time_majored = time_major_attr.getValue();
      auto input_type = mlir::dyn_cast_or_null<RankedTensorType>(input.getType());
      if (!input_type) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 36.2K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/lite/utils/lstm_utils.h

      func::FuncOp fused_func_op_;
      Value input_;
      Value weight_;
      Value bias_;
      Value projection_;
      bool couple_input_forget_gates_;
    
      // internal state
      Value weight_transposed_;
      Value projection_transposed_;
      RankedTensorType weight_type_;
      RankedTensorType projection_type_;
      int num_gates_;
      int n_cell_;
      int n_output_;
      int n_input_;
      int num_cols_weight_transposed_;
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sat Jun 03 00:14:05 UTC 2023
    - 7.3K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/tensorflow/translate/import_model.cc

        }
      }
    
      if (!inputs.empty() || !outputs.empty()) {
        arg_nodes->resize(inputs.size());
        ret_nodes->resize(outputs.size());
    
        for (Node* n : GetOrderedNodes()) {
          // Handle inputs/arguments.
          auto input_it = inputs.find(n->name());
          if (input_it != inputs.end()) {
            (*arg_nodes)[input_it->second] = {n, 0};
          }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 01 11:17:36 UTC 2024
    - 183.2K bytes
    - Viewed (0)
Back to top