Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 65 for Brad (0.05 sec)

  1. src/database/sql/sql_test.go

    		wantErr string
    	}
    	execTests := []execTest{
    		// Okay:
    		{[]any{"Brad", 31}, ""},
    		{[]any{"Brad", int64(31)}, ""},
    		{[]any{"Bob", "32"}, ""},
    		{[]any{7, 9}, ""},
    
    		// Invalid conversions:
    		{[]any{"Brad", int64(0xFFFFFFFF)}, "sql: converting argument $2 type: sql/driver: value 4294967295 overflows int32"},
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 23 18:42:28 UTC 2024
    - 111.6K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/tfr/examples/mnist/ops_defs.py

    
    @tf.RegisterGradient('NewConv2D')
    def _conv_add_relu_grad(op: ops.Operation, grad):
      act = op.get_attr('act')
      y = op.outputs[0]
      if act == 'RELU':
        grad = gen_nn_ops.relu_grad(grad, y)
      elif act == 'RELU6':
        grad = gen_nn_ops.relu6_grad(grad, y)
      elif act == 'TANH':
        y = math_ops.conj(y)
        grad = gen_math_ops.tanh_grad(y, grad)
    
      broadcast_shape = tf.shape(y)
      input_value_shape = tf.shape(op.inputs[2])
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Aug 31 20:23:51 UTC 2023
    - 6.8K bytes
    - Viewed (0)
  3. tensorflow/cc/gradients/math_grad.cc

                   std::vector<Output>* grad_outputs) {
      auto grad = grad_inputs[0];
      auto two_over_root_pi =
          Cast(scope, Const(scope, 2 / std::sqrt(M_PI)), grad.type());
      Scope grad_scope = scope.WithControlDependencies(grad);
      auto x = ConjugateHelper(grad_scope, op.input(0));
      // grad * 2/sqrt(pi) * exp(-x**2)
      auto dx = Mul(grad_scope, Mul(grad_scope, grad, two_over_root_pi),
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Aug 25 18:20:20 UTC 2023
    - 50.7K bytes
    - Viewed (0)
  4. tensorflow/cc/gradients/nn_grad.cc

        auto multiply_result = Multiply(scope, subtraction_result, logits_softmax);
        grad = Add(scope, grad, multiply_result);
      }
      auto minus_log_softmax = Multiply(scope, LogSoftmax(scope, logits), -1.0f);
      grad_outputs->push_back(grad);
      grad_outputs->push_back(BroadcastMul(scope, grad_loss, minus_log_softmax));
      return scope.status();
    }
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 27 23:34:33 UTC 2022
    - 24.5K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/tensorflow/tests/decompose_resource_ops.mlir

        // CHECK: [[GRAD_SQUARE:%.*]] = "tf.Mul"([[GRAD]], [[GRAD]]) : (tensor<f32>, tensor<f32>) -> tensor<f32>
        // CHECK: [[NEW_ACC:%.*]] = "tf.AddV2"([[OLD_ACC]], [[GRAD_SQUARE]]) : (tensor<*xf32>, tensor<f32>) -> tensor<*xf32>
        // CHECK: [[LR_MULTIPLY:%.*]] = "tf.Mul"([[LR]], [[GRAD]]) : (tensor<f32>, tensor<f32>) -> tensor<f32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 22 19:47:48 UTC 2024
    - 51.3K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/tensorflow/transforms/decompose_resource_ops.td

         (CreateTFReadVariableOp $src_op, $grad, $ms_resource),
         (TF_AddV2Op:$ms_new
           (TF_MulOp
             (TF_MulOp $grad, $grad),
             (TF_SubOp $one, $rho)
           ),
           (TF_MulOp
              (CreateTFReadVariableOp $src_op, $grad, $ms_resource),
              $rho
           )
         ),
         (TF_AssignVariableOp $ms_resource, $ms_new, (CreateConstBoolAttrFalse)),
         // mg = grad * (one - rho) + mg * rho;
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 22 19:47:48 UTC 2024
    - 20.7K bytes
    - Viewed (0)
  7. tensorflow/c/experimental/gradients/math_grad.cc

        /* Given upstream grad U and a Sub op A-B, the gradients are:
         *
         *    dA =  U
         *    dB = -U
         *
         */
    
        // Grad for A
        DCHECK(grad_outputs[0]);
        grad_inputs[0] = grad_outputs[0];
        grad_inputs[0]->Ref();
    
        // Grad for B
        // negate the upstream grad
        std::string name = "Neg_Sub_Grad_B";
        TF_RETURN_IF_ERROR(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 28 13:53:47 UTC 2024
    - 15.2K bytes
    - Viewed (0)
  8. tensorflow/c/experimental/gradients/nn_grad.cc

                     absl::Span<AbstractTensorHandle*> grad_inputs) override {
        // Grad for Softmax Input
        TF_RETURN_IF_ERROR(BroadcastMul(
            ctx, grad_outputs[0], forward_outputs_[1],
            grad_inputs.subspan(0, 1)));  // upstream_grad * local softmax grad
    
        // Grad for labels is null
        grad_inputs[1] = nullptr;
        return absl::OkStatus();
      }
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Apr 09 06:38:45 UTC 2024
    - 5.7K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tfr/examples/pad/ops_defs.py

        input_ = tf.raw_ops.Concat(
            concat_dim=i, values=[left_padding, input_, right_padding])
      return input_
    
    
    @tf.RegisterGradient('NewMirrorPad')
    def _mirror_pad_grad(op, grad):
      mode = op.get_attr('mode')
      return [gen_array_ops.mirror_pad_grad(grad, op.inputs[1], mode=mode), None]
    
    
    @Composite(
        'NewMirrorPadGrad',
        inputs=['input_: T', 'paddings: Tpaddings'],
        attrs=['mode: {"REFLECT", "SYMMETRIC"}'],
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Oct 01 05:00:29 UTC 2021
    - 5.6K bytes
    - Viewed (0)
  10. tensorflow/compiler/jit/ops/xla_ops_grad.py

    # ==============================================================================
    
    from tensorflow.python.framework import ops
    
    
    @ops.RegisterGradient("XlaClusterOutput")
    def _XlaClusterOutputGrad(_, grad):
      del grad  # unused
      raise RuntimeError("Gradient computation of graph in xla.compile() is "
                         "prohibited because it can cause performance degradation."
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Sep 28 21:37:05 UTC 2021
    - 1.1K bytes
    - Viewed (0)
Back to top