- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 23 for AddN (0.04 sec)
-
tensorflow/cc/gradients/math_grad.cc
const std::vector<Output>& grad_inputs, std::vector<Output>* grad_outputs) { // AddN doesn't support broadcasting, so all the inputs must be the // same shape. // Note: // dy/dx_k = d(x_1 + x_2 + ... + x_n)/dx_k = 1 for all x_k // hence dx_k = dy for all x_k // So the gradient for AddN just transfers the incoming gradient to // all outgoing gradients.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Aug 25 18:20:20 UTC 2023 - 50.7K bytes - Viewed (0) -
tensorflow/c/eager/gradients.cc
if (gradient_tensors.size() == 1) { return gradient_tensors[0]; } AbstractOperationPtr op(ctx_->CreateOperation()); Status s = op->Reset("AddN", /*raw_device_name=*/nullptr); if (!s.ok()) { return nullptr; } s = op->AddInputList(gradient_tensors); if (!s.ok()) { return nullptr; } int num_outputs = 1;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 15 09:49:45 UTC 2024 - 19.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/lower_tf.mlir
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Jan 05 18:35:42 UTC 2024 - 92K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/tests/legalize-tf-with-tf2xla-hlo-importer.mlir
} // CHECK-LABEL: unsupported_dtype func.func @unsupported_dtype(%arg0: tensor<2x!tf_type.variant>) -> tensor<2x!tf_type.variant> { // CHECK: tf.AddN // expected-remark@+1 {{skipping legalization due to unsupported type 'tensor<2x!tf_type.variant>'}} %0 = "tf.AddN"(%arg0, %arg0) : (tensor<2x!tf_type.variant>, tensor<2x!tf_type.variant>) -> tensor<2x!tf_type.variant> func.return %0 : tensor<2x!tf_type.variant> }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Apr 06 15:32:52 UTC 2024 - 38.6K bytes - Viewed (0) -
tensorflow/cc/framework/gradients.cc
// Just one backprop edge. *grad = grads_to_keep[0]; } else { // Otherwise, adds backprop-ed gradients. // TODO(andydavis) Use a better accumulator here. *grad = ops::AddN(scope_, grads_to_keep); } return absl::OkStatus(); } bool SymbolicGradientBuilder::IsPrimitiveOpWithNoGrad(const string& opname) { ops::GradFunc grad_fn;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Apr 13 05:57:22 UTC 2024 - 22K bytes - Viewed (0) -
tensorflow/c/c_api_test.cc
ASSERT_EQ(TF_OK, TF_GetCode(s)) << TF_Message(s); // Test TF_Operation*() query functions. EXPECT_EQ(string("add"), string(TF_OperationName(add))); EXPECT_EQ(string("AddN"), string(TF_OperationOpType(add))); EXPECT_EQ(string(""), string(TF_OperationDevice(add))); EXPECT_EQ(1, TF_OperationNumOutputs(add)); EXPECT_EQ(TF_INT32, TF_OperationOutputType(TF_Output{add, 0}));
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Apr 15 03:35:10 UTC 2024 - 96.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/legalize-tf.mlir
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 05 01:54:33 UTC 2024 - 153.4K bytes - Viewed (0) -
tensorflow/c/c_api_function_test.cc
TF_Operation* add = Add(feed1, feed2, func_graph_, s_); DefineT(-1, {}, {{feed1, 0}, {feed2, 0}}, {{add, 3}}, {}, true); EXPECT_EQ(TF_OUT_OF_RANGE, TF_GetCode(s_)); EXPECT_EQ(string("Node 'add' (type: 'AddN', num of outputs: 1) does " "not have output 3\n\tEncountered while processing " "output 0 from function 'MyFunc'"), string(TF_Message(s_))); }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jul 20 22:08:54 UTC 2023 - 63.6K bytes - Viewed (0) -
tensorflow/compiler/jit/encapsulate_subgraphs_pass_test.cc
} Node* AddNLike(const std::vector<ops::NodeOut>& inputs, const GraphDefBuilder::Options& opts) { if (opts.HaveError()) return nullptr; NodeBuilder node_builder(opts.GetNameForOp("AddN"), "AddNLikeTest", opts.op_registry()); node_builder.Input(inputs); return opts.FinalizeBuilder(&node_builder); }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 08:47:20 UTC 2024 - 113.3K bytes - Viewed (0) -
tensorflow/compiler/jit/mark_for_compilation_pass.cc
"Less", "LessEqual", "SigmoidGrad", "SoftplusGrad", "SoftsignGrad", "TanhGrad", "Pow", "SquaredDifference", "ApproximateEqual", // Others "AddN", "Bitcast", "Cast", "ClipByValue", "Const", "Empty", "Identity", "IdentityN", "Relu", "Relu6", "ReluGrad", "Relu6Grad", "LeakyReluGrad", "Elu", "EluGrad", "Selu", "SeluGrad", "Select",
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 21 12:19:41 UTC 2024 - 85.3K bytes - Viewed (0)