- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 27 for _backprop (0.28 sec)
-
tensorflow/cc/framework/while_gradients.cc
return result; } // The backprop loop counter and main backprop loop run in their own execution // frame (conceptually, the main forward loop and forward loop counter run // together in a frame, then the backprop loop counter and backprop loop run // together in a different frame). This returns the frame name to use for the // backprop while loops.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Apr 13 05:57:22 UTC 2024 - 8.1K bytes - Viewed (0) -
tensorflow/cc/framework/gradients.cc
DCHECK(while_ctx != nullptr); // Record 'summed_grads' as the backprop input associated with 'exit_node' std::map<Node*, Output>& backprops = while_backprops_[while_ctx]; DCHECK(backprops.find(exit_node) == backprops.end()); backprops[exit_node] = summed_grads; // Wait until we have all exit nodes' backprops collected before processing // the while loop.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Apr 13 05:57:22 UTC 2024 - 22K bytes - Viewed (0) -
pkg/ctrlz/assets/static/js/bootstrap-4.0.0.min.js
tScrollbar(),t(e._element).trigger(h.HIDDEN)})},p._removeBackdrop=function(){this._backdrop&&(t(this._backdrop).remove(),this._backdrop=null)},p._showBackdrop=function(e){var n=this,i=t(this._element).hasClass(d)?d:"";if(this._isShown&&this._config.backdrop){var s=P.supportsTransitionEnd()&&i;if(this._backdrop=document.createElement("div"),this._backdrop.className=u,i&&t(this._backdrop).addClass(i),t(this._backdrop).appendTo(document.body),t(this._element).on(h.CLICK_DISMISS,function(t){n._ignor...
Registered: Fri Jun 14 15:00:06 UTC 2024 - Last Modified: Tue May 23 17:08:31 UTC 2023 - 47.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfr/python/test_utils.py
# compute with op. with backprop.GradientTape() as gt: for var_ in vars_: gt.watch(var_) y = compute_op(**op_kwargs) # uses op and decomposites by the graph pass. grads = gt.gradient(y, vars_) # uses registered gradient function. # compute with composition with backprop.GradientTape() as gt: for var_ in vars_: gt.watch(var_)
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Jun 02 18:32:17 UTC 2023 - 1.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/tpu_space_to_depth_pass.cc
// Build new BackPropFilterOp. auto loc = backprop.getLoc(); auto new_backprop = builder.create<TF::Conv2DBackpropFilterOp>( loc, new_result_type, input, new_filter_sizes, backprop.getOutBackprop(), strides, backprop.getUseCudnnOnGpu(), backprop.getPadding(), backprop.getExplicitPaddings(), backprop.getDataFormat(), backprop.getDilations());
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 29.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/layout_optimization_layout_assignment_to_nchw.mlir
// CHECK: %[[RES_PERM:.*]] = "tf.Const"() // CHECK-SAME: <{value = dense<[0, 2, 3, 1]> : tensor<4xi64>}> // CHECK: %[[RES_TPOSE:[0-9]*]] = "tf.Transpose" // CHECK-SAME: (%x_backprop, %[[RES_PERM]]) // CHECK: return %[[RES_TPOSE]] %x_backprop, %scale_backprop, %offset_backprop, %reserve_1, %reserve_2 = "tf.FusedBatchNormGradV3"(%arg0, %arg1, %arg2, %arg2, %arg2, %arg2) { data_format = "NHWC",
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Oct 30 06:52:55 UTC 2023 - 9K bytes - Viewed (0) -
tensorflow/c/experimental/gradients/nn_grad_test.cc
absl::Span<AbstractTensorHandle*> outputs) { AbstractTensorHandle* loss; AbstractTensorHandle* backprop; TF_RETURN_IF_ERROR(ops::SparseSoftmaxCrossEntropyWithLogits( ctx, inputs[0], inputs[1], &loss, &backprop, "SparseSoftmaxCrossEntropyWithLogits")); // `gradient_checker` only works with model that returns only 1 tensor.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 28 13:53:47 UTC 2024 - 8.3K bytes - Viewed (0) -
tensorflow/c/experimental/ops/gen/cpp/golden/testing_ops.h.golden
// Status SparseSoftmaxCrossEntropyWithLogits(AbstractContext* ctx, AbstractTensorHandle* const features, AbstractTensorHandle* const labels, AbstractTensorHandle** loss, AbstractTensorHandle** backprop, const char* name = nullptr, const char* raw_device_name = nullptr); //
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Nov 16 19:04:03 UTC 2023 - 2.9K bytes - Viewed (0) -
tensorflow/c/experimental/ops/gen/cpp/golden/testing_ops.cc.golden
// Summary: // // Description: Status SparseSoftmaxCrossEntropyWithLogits(AbstractContext* ctx, AbstractTensorHandle* const features, AbstractTensorHandle* const labels, AbstractTensorHandle** loss, AbstractTensorHandle** backprop, const char* name, const char* raw_device_name) { AbstractOperationPtr op_ptr(ctx->CreateOperation()); TF_RETURN_IF_ERROR(op_ptr->Reset("SparseSoftmaxCrossEntropyWithLogits", raw_device_name));
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Nov 16 19:04:03 UTC 2023 - 6.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/lower_tf.td
// computes loss and backprop of the loss with respect to 'features'. // // Softmax cross entropy loss is defined as follows: // // loss = Sum(-labels * Log(Exp(features) / Sum(Exp(features))) // loss = Sum(-labels * LogSoftmax(features)) // // Computing gradient of the loss with respect to features gives us, // // backprop = (Exp(features) / Sum(Exp(features))) - labels
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 04 13:30:42 UTC 2024 - 24.7K bytes - Viewed (0)