- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 65 for Brad (0.05 sec)
-
src/database/sql/sql_test.go
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 18:42:28 UTC 2024 - 111.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfr/examples/mnist/ops_defs.py
@tf.RegisterGradient('NewConv2D') def _conv_add_relu_grad(op: ops.Operation, grad): act = op.get_attr('act') y = op.outputs[0] if act == 'RELU': grad = gen_nn_ops.relu_grad(grad, y) elif act == 'RELU6': grad = gen_nn_ops.relu6_grad(grad, y) elif act == 'TANH': y = math_ops.conj(y) grad = gen_math_ops.tanh_grad(y, grad) broadcast_shape = tf.shape(y) input_value_shape = tf.shape(op.inputs[2])
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Aug 31 20:23:51 UTC 2023 - 6.8K bytes - Viewed (0) -
tensorflow/cc/gradients/math_grad.cc
std::vector<Output>* grad_outputs) { auto grad = grad_inputs[0]; auto two_over_root_pi = Cast(scope, Const(scope, 2 / std::sqrt(M_PI)), grad.type()); Scope grad_scope = scope.WithControlDependencies(grad); auto x = ConjugateHelper(grad_scope, op.input(0)); // grad * 2/sqrt(pi) * exp(-x**2) auto dx = Mul(grad_scope, Mul(grad_scope, grad, two_over_root_pi),
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Aug 25 18:20:20 UTC 2023 - 50.7K bytes - Viewed (0) -
tensorflow/cc/gradients/nn_grad.cc
auto multiply_result = Multiply(scope, subtraction_result, logits_softmax); grad = Add(scope, grad, multiply_result); } auto minus_log_softmax = Multiply(scope, LogSoftmax(scope, logits), -1.0f); grad_outputs->push_back(grad); grad_outputs->push_back(BroadcastMul(scope, grad_loss, minus_log_softmax)); return scope.status(); }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 27 23:34:33 UTC 2022 - 24.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/decompose_resource_ops.mlir
// CHECK: [[GRAD_SQUARE:%.*]] = "tf.Mul"([[GRAD]], [[GRAD]]) : (tensor<f32>, tensor<f32>) -> tensor<f32> // CHECK: [[NEW_ACC:%.*]] = "tf.AddV2"([[OLD_ACC]], [[GRAD_SQUARE]]) : (tensor<*xf32>, tensor<f32>) -> tensor<*xf32> // CHECK: [[LR_MULTIPLY:%.*]] = "tf.Mul"([[LR]], [[GRAD]]) : (tensor<f32>, tensor<f32>) -> tensor<f32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 22 19:47:48 UTC 2024 - 51.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/decompose_resource_ops.td
(CreateTFReadVariableOp $src_op, $grad, $ms_resource), (TF_AddV2Op:$ms_new (TF_MulOp (TF_MulOp $grad, $grad), (TF_SubOp $one, $rho) ), (TF_MulOp (CreateTFReadVariableOp $src_op, $grad, $ms_resource), $rho ) ), (TF_AssignVariableOp $ms_resource, $ms_new, (CreateConstBoolAttrFalse)), // mg = grad * (one - rho) + mg * rho;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 22 19:47:48 UTC 2024 - 20.7K bytes - Viewed (0) -
tensorflow/c/experimental/gradients/math_grad.cc
/* Given upstream grad U and a Sub op A-B, the gradients are: * * dA = U * dB = -U * */ // Grad for A DCHECK(grad_outputs[0]); grad_inputs[0] = grad_outputs[0]; grad_inputs[0]->Ref(); // Grad for B // negate the upstream grad std::string name = "Neg_Sub_Grad_B"; TF_RETURN_IF_ERROR(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 28 13:53:47 UTC 2024 - 15.2K bytes - Viewed (0) -
tensorflow/c/experimental/gradients/nn_grad.cc
absl::Span<AbstractTensorHandle*> grad_inputs) override { // Grad for Softmax Input TF_RETURN_IF_ERROR(BroadcastMul( ctx, grad_outputs[0], forward_outputs_[1], grad_inputs.subspan(0, 1))); // upstream_grad * local softmax grad // Grad for labels is null grad_inputs[1] = nullptr; return absl::OkStatus(); }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Apr 09 06:38:45 UTC 2024 - 5.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfr/examples/pad/ops_defs.py
input_ = tf.raw_ops.Concat( concat_dim=i, values=[left_padding, input_, right_padding]) return input_ @tf.RegisterGradient('NewMirrorPad') def _mirror_pad_grad(op, grad): mode = op.get_attr('mode') return [gen_array_ops.mirror_pad_grad(grad, op.inputs[1], mode=mode), None] @Composite( 'NewMirrorPadGrad', inputs=['input_: T', 'paddings: Tpaddings'], attrs=['mode: {"REFLECT", "SYMMETRIC"}'],
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Oct 01 05:00:29 UTC 2021 - 5.6K bytes - Viewed (0) -
tensorflow/compiler/jit/ops/xla_ops_grad.py
# ============================================================================== from tensorflow.python.framework import ops @ops.RegisterGradient("XlaClusterOutput") def _XlaClusterOutputGrad(_, grad): del grad # unused raise RuntimeError("Gradient computation of graph in xla.compile() is " "prohibited because it can cause performance degradation."
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Sep 28 21:37:05 UTC 2021 - 1.1K bytes - Viewed (0)