- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 71 for dy (0.02 sec)
-
tensorflow/cc/gradients/math_grad.cc
grad_outputs->push_back(dx); auto dy = MatMul(scope, y0, y1, MatMul::TransposeA(adj_y0).TransposeB(adj_y1)); grad_outputs->push_back(dy); } else { auto dx = BatchMatMulV3(scope, x0, x1, x_data_type, BatchMatMulV3::AdjX(adj_x0).AdjY(adj_x1)); grad_outputs->push_back(dx); auto dy = BatchMatMulV3(scope, y0, y1, y_data_type,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Aug 25 18:20:20 UTC 2023 - 50.7K bytes - Viewed (0) -
src/image/draw/draw.go
*r = r.Intersect(src.Bounds().Add(orig.Sub(*sp))) if mask != nil { *r = r.Intersect(mask.Bounds().Add(orig.Sub(*mp))) } dx := r.Min.X - orig.X dy := r.Min.Y - orig.Y if dx == 0 && dy == 0 { return } sp.X += dx sp.Y += dy if mp != nil { mp.X += dx mp.Y += dy } } func processBackward(dst image.Image, r image.Rectangle, src image.Image, sp image.Point) bool { return dst == src &&
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 11 17:08:05 UTC 2024 - 33.9K bytes - Viewed (0) -
staging/src/k8s.io/apimachinery/pkg/util/duration/duration.go
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Sun Jan 29 09:44:02 UTC 2023 - 2.8K bytes - Viewed (0) -
tensorflow/cc/framework/gradients_test.cc
TF_ASSERT_OK(scope.status()); // Construct grad inputs. auto dx = Const(scope, 4, {4, 2}); auto dy = Const(scope, 5, {4, 2}); auto dz = Const(scope, 6, {4, 2}); if (expected) { // Construct backward graph. auto unpack_grad = Stack(scope, {dx, dy, dz}); auto pack_grad = Unstack(scope, unpack_grad.output, 3); } else { // Call AddSymbolicGradients.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Apr 15 15:13:38 UTC 2023 - 25K bytes - Viewed (0) -
tensorflow/cc/framework/gradients.cc
TF_RETURN_IF_ERROR(Initialize()); // Backward propagation. std::vector<Output> dy; while (!ready_.empty()) { // n has collected all gradients. Node* n = ready_.front(); ready_.pop_front(); // dy[i] is the sum of i-th output's backpropped gradients. const int num_y = n->num_outputs(); dy.clear(); dy.resize(num_y, {nullptr, 0}); std::vector<int> no_grad_dy_indices;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Apr 13 05:57:22 UTC 2024 - 22K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/lower_tf.td
// Sqrt op patterns. //===----------------------------------------------------------------------===// // SqrtGrad(y, dy) = dy * 0.5 / y def LowerSqrtGradOp : Pat< (TF_SqrtGradOp $y, $dy), (TF_DivOp (TF_MulOp $dy, (TF_ConstOp (GetScalarOfFloatType<"0.5"> $dy))), $y )>; //===----------------------------------------------------------------------===// // TanhGrad op patterns.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 04 13:30:42 UTC 2024 - 24.7K bytes - Viewed (0) -
tensorflow/cc/framework/gradient_checker.h
namespace tensorflow { /// Returns in 'max_error' the maximum element-wise error for dy/dx between the /// computed and numeric Jacobian matrices where 'xs' and 'ys' are tensors. /// X_T and Y_T are the c++ types for the x and y tensors, and JAC_T is a /// real-valued type to store the Jacobian derivatives dy/dx. /// This function adds operations to the graph associated with 'scope'. /// /// Examples:
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Oct 05 15:35:17 UTC 2022 - 2.8K bytes - Viewed (0) -
src/cmd/vendor/golang.org/x/mod/semver/semver.go
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Jul 12 20:38:21 UTC 2023 - 8.9K bytes - Viewed (0) -
internal/s3select/sql/timestampfuncs.go
} duration := ts2.Sub(ts1) y1, m1, d1 := ts1.Date() y2, m2, d2 := ts2.Date() switch timePart { case timePartYear: dy := int64(y2 - y1) if m2 > m1 || (m2 == m1 && d2 >= d1) { return FromInt(dy), nil } return FromInt(dy - 1), nil case timePartMonth: m1 += time.Month(12 * y1) m2 += time.Month(12 * y2) return FromInt(int64(m2 - m1)), nil case timePartDay:
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Sun Jan 02 17:15:06 UTC 2022 - 4.8K bytes - Viewed (0) -
tensorflow/c/experimental/ops/array_ops.cc
// backprop such that dx = g(dy). In Python, // // ```python // with tf.get_default_graph().gradient_override_map( // {'IdentityN': 'OverrideGradientWithG'}): // y, _ = identity_n([f(x), x]) // // @tf.RegisterGradient('OverrideGradientWithG') // def ApplyG(op, dy, _): // return [None, g(dy)] # Do not backprop to f(x). // ``` Status IdentityN(AbstractContext* ctx,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 10 19:11:36 UTC 2022 - 6.7K bytes - Viewed (0)