- Sort Score
- Result 10 results
- Languages All
Results 21 - 30 of 32 for AddEdge (0.31 sec)
-
tensorflow/compiler/jit/encapsulate_xla_computations_pass.cc
graph->AddEdge(data_inputs[i].first, data_inputs[i].second, xla_launch, i); } for (Node* n : control_inputs) { graph->AddControlEdge(n, xla_launch); } for (int i = 0, end = data_outputs.size(); i < end; ++i) { for (const auto& successor : data_outputs[i]) { graph->AddEdge(xla_launch, i, successor.first, successor.second); }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Mar 12 06:33:33 UTC 2024 - 15.1K bytes - Viewed (0) -
tensorflow/compiler/jit/partially_decluster_pass.cc
cloned_node->set_assigned_device_name(n->assigned_device_name()); for (const Edge* in_edge : n->in_edges()) { graph->AddEdge(in_edge->src(), in_edge->src_output(), cloned_node, in_edge->dst_input()); } for (const Edge* out_edge_to_clone : out_edges_to_clone) { graph->AddEdge(cloned_node, out_edge_to_clone->src_output(), out_edge_to_clone->dst(), out_edge_to_clone->dst_input());
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Feb 09 11:36:41 UTC 2024 - 15.7K bytes - Viewed (0) -
tensorflow/compiler/jit/encapsulate_subgraphs_pass.cc
args_.push_back(arg); } Node* dst_node = edge->dst(); Node* dst_image = node_images.at(dst_node); int dst_slot = edge->dst_input(); args_by_dst_[InputTensor(dst_node, dst_slot)] = arg_index; graph_->AddEdge(args_[arg_index], 0, dst_image, dst_slot); return absl::OkStatus(); } Status Encapsulator::Subgraph::RecordControlResult( const Edge* edge, const absl::flat_hash_map<const Node*, Node*>& node_images) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 08:47:20 UTC 2024 - 51K bytes - Viewed (0) -
tensorflow/cc/ops/while_loop.cc
// Create the backedges from the NextIteration nodes to the Merge nodes. for (size_t i = 0; i < num_loop_vars; ++i) { const int merge_backedge_output_index = 1; scope.graph()->AddEdge(next_outputs[i].node(), next_outputs[i].index(), merge_outputs[i].node(), merge_backedge_output_index); } outputs->resize(num_loop_vars);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Feb 26 01:01:21 UTC 2024 - 9.5K bytes - Viewed (0) -
android/guava/src/com/google/common/graph/Graphs.java
for (E edge : network.outEdges(node)) { N successorNode = network.incidentNodes(edge).adjacentNode(node); if (subgraph.nodes().contains(successorNode)) { subgraph.addEdge(node, successorNode, edge); } } } return subgraph; } /** Creates a mutable copy of {@code graph} with the same nodes and edges. */
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Mon Apr 01 16:15:01 UTC 2024 - 21.2K bytes - Viewed (0) -
tensorflow/compiler/jit/build_xla_ops_pass_test.cc
auto var = ops::VarHandleOp(root.WithOpName("var"), DT_INT32, TensorShape({})); auto int32_on_device = ops::ReadVariableOp(root.WithOpName("int32_on_device"), var, DT_INT32); root.graph()->AddEdge(int32_on_device.node(), 0, call, 0); std::unique_ptr<Graph> graph; TF_ASSERT_OK(BuildXlaOps(root, fdef_lib, &graph)); Node* stateful_partitioned_call_op = nullptr; for (Node* n : graph->op_nodes()) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 08:47:20 UTC 2024 - 12.2K bytes - Viewed (0) -
guava/src/com/google/common/graph/Graphs.java
for (E edge : network.outEdges(node)) { N successorNode = network.incidentNodes(edge).adjacentNode(node); if (subgraph.nodes().contains(successorNode)) { subgraph.addEdge(node, successorNode, edge); } } } return subgraph; } /** Creates a mutable copy of {@code graph} with the same nodes and edges. */
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Mon Apr 01 16:15:01 UTC 2024 - 21.7K bytes - Viewed (0) -
tensorflow/compiler/jit/shape_inference.cc
graph->RemoveEdge(e); } else { Node* dst = e->dst(); int dst_input = e->dst_input(); graph->RemoveEdge(e); graph->AddEdge(const_node, 0, dst, dst_input); } } } } } // Merge node causes a loop so we remove NextIteration->Merge edge before
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 31 00:41:19 UTC 2024 - 13K bytes - Viewed (0) -
src/runtime/mklockrank.go
if err != nil { log.Fatal(err) } var out []byte if *flagDot { var b bytes.Buffer g.TransitiveReduction() // Add cyclic edges for visualization. for k := range cyclicRanks { g.AddEdge(k, k) } // Reverse the graph. It's much easier to read this as // a "<" partial order than a ">" partial order. This // ways, locks are acquired from the top going down
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 08 17:47:01 UTC 2024 - 9.1K bytes - Viewed (0) -
tensorflow/compiler/jit/increase_dynamism_for_auto_jit_pass.cc
DCHECK(e->src_output() == 0 || e->src_output() == Graph::kControlSlot); int src_output = e->src_output(); int dst_input = e->dst_input(); Node* dst = e->dst(); g->RemoveEdge(e); g->AddEdge(static_shaped_slice, src_output, dst, dst_input); } for (const Edge* e : slice->in_edges()) { if (e->IsControlEdge()) { g->AddControlEdge(e->src(), static_shaped_slice); } }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Mar 12 06:33:33 UTC 2024 - 13.8K bytes - Viewed (0)