Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 16 for grad_ops (0.27 sec)

  1. tensorflow/cc/BUILD

        ],
    )
    
    tf_cc_test(
        name = "framework_gradients_test",
        srcs = ["framework/gradients_test.cc"],
        deps = [
            ":cc_ops",
            ":client_session",
            ":grad_op_registry",
            ":grad_ops",
            ":gradients",
            ":testutil",
            "//tensorflow/core:all_kernels",
            "//tensorflow/core:framework",
            "//tensorflow/core:framework_internal",
            "//tensorflow/core:protos_all_cc",
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Mar 27 18:00:18 UTC 2024
    - 23.5K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/tensorflow/transforms/remove_vars_in_session_initializer.cc

                         llvm::SmallVectorImpl<Operation*>& erase_list,
                         llvm::SmallPtrSetImpl<Operation*>& dead_ops) {
      for (mlir::Value res : op->getResults()) {
        for (Operation* user : res.getUsers()) {
          if (!dead_ops.insert(user).second) continue;
          RecursiveRemove(user, erase_list, dead_ops);
        }
      }
    
      erase_list.push_back(op);
    
      for (auto& use : op->getOpOperands()) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 3.6K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/tfr/examples/mnist/ops_defs.py

    def _composite_fully_connected(input_, filter_, bias, act):
      res = tf.raw_ops.MatMul(
          a=input_, b=filter_, transpose_a=False, transpose_b=True)
      res = tf.raw_ops.Add(x=res, y=bias)
      if act == 'RELU':
        return tf.raw_ops.Relu(features=res)
      elif act == 'RELU6':
        return tf.raw_ops.Relu6(features=res)
      elif act == 'TANH':
        return tf.raw_ops.Tanh(x=res)
      else:
        return res
    
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Aug 31 20:23:51 UTC 2023
    - 6.8K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/tfr/examples/pad/ops_defs.py

      shape = input_.shape.as_list()
      for i in range(len(shape)):
        rdims = tf.raw_ops.OneHot(
            indices=i, depth=len(shape), on_value=True, off_value=False, axis=-1)
        rarray = tf.raw_ops.Reverse(tensor=input_, dims=rdims)
    
        left_padding_size = tf.raw_ops.GatherNd(params=paddings, indices=[i, 0])
        right_padding_size = tf.raw_ops.GatherNd(params=paddings, indices=[i, 1])
    
        if mode == 'REFLECT':
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Oct 01 05:00:29 UTC 2021
    - 5.6K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/tfr/examples/pad/BUILD

        ],
        licenses = ["notice"],
    )
    
    package_group(
        name = "friends",
        packages = [
            "//tensorflow/compiler/mlir/tfr/...",
        ],
    )
    
    gen_op_libraries(
        name = "pad_ops",
        src = "ops_defs.py",
        deps = [
            "//tensorflow:tensorflow_py",
            "//tensorflow/python/ops:array_ops_gen",
            "//tensorflow/python/platform:flags",
            "@absl_py//absl:app",
        ],
    )
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Oct 19 00:51:34 UTC 2023
    - 1.2K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/tfr/README.md

    def _composite_fully_connected(input_, filter_, bias, act):
      res = tf.raw_ops.MatMul(
          a=input_, b=filter_, transpose_a=False, transpose_b=True)
      res = tf.raw_ops.Add(x=res, y=bias)
      if act == 'RELU':
        return tf.raw_ops.Relu(features=res)
      elif act == 'RELU6':
        return tf.raw_ops.Relu6(features=res)
      elif act == 'TANH':
        return tf.raw_ops.Tanh(x=res)
      else:
        return res
    
    ```
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Mar 29 18:32:13 UTC 2022
    - 6.2K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/tfr/examples/pad/pad_ops_test.py

        }
        kwargs_ = {
            'input_': input_,
            'paddings': paddings,
            'mode': mode,
        }
        # Make sure the composition python function is correct
        self._assertOpAndComposite([input_], tf.raw_ops.MirrorPad,
                                   ops_defs._composite_mirror_pad, kwargs_, kwargs)
        # Make sure the translation and decomposition is correct
        self._assertOpAndComposite([input_],
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Sep 28 21:37:05 UTC 2021
    - 3.4K bytes
    - Viewed (0)
  8. tensorflow/cc/gradients/nn_grad.cc

      // The "truth" labels are at index 1.
      auto softmax_grad = op.output(1);
    
      // The loss is the output at index 0, and backprop is the output at index 1.
      auto grad_loss = grad_inputs[0];
      auto grad_grad = grad_inputs[1];
    
      auto grad = BroadcastMul(scope, grad_loss, softmax_grad);
      if (!IsZero(scope, grad_grad)) {
        std::vector<int> axis;
        auto logits_softmax = Softmax(scope, logits);
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 27 23:34:33 UTC 2022
    - 24.5K bytes
    - Viewed (0)
  9. RELEASE.md

    *   Fixes a security vulnerability caused by lack of validation in
        `tf.raw_ops.DataFormatVecPermute` and `tf.raw_ops.DataFormatDimMap`
        ([CVE-2020-26267](https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-26267))
    *   Fixes a vulnerability caused by attempting to write to immutable memory
        region in `tf.raw_ops.ImmutableConst`
        ([CVE-2020-26268](https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-26268)
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 23:24:08 UTC 2024
    - 730.3K bytes
    - Viewed (0)
  10. tensorflow/cc/gradients/array_grad.cc

      int axis;
      TF_RETURN_IF_ERROR(GetNodeAttr(op.node()->attrs(), "axis", &axis));
    
      grad_outputs->reserve(N);
      auto grad_op = Unstack(scope, grad_inputs[0], N, Unstack::Axis(axis));
      for (const Output& o : grad_op.output) {
        grad_outputs->emplace_back(o);
      }
      return scope.status();
    }
    REGISTER_GRADIENT_OP("Pack", PackGrad);
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Oct 10 23:33:32 UTC 2023
    - 31.7K bytes
    - Viewed (0)
Back to top