- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 69 for op_registry (0.14 sec)
-
tensorflow/compiler/jit/mark_for_compilation_pass_test_helper.cc
/*static*/ Status MarkForCompilationPassTestHelper::MarkForCompilation( std::unique_ptr<Graph>* graph, MarkForCompilationPassTestHelper::Options options) { FunctionDefLibrary flib; FunctionLibraryDefinition flib_def((*graph)->op_registry(), flib); return MarkForCompilation(graph, &flib_def, options); }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 09 19:51:48 UTC 2023 - 3.1K bytes - Viewed (0) -
tensorflow/c/c_api_internal.h
}; struct TF_OperationDescription { TF_OperationDescription(TF_Graph* g, const char* op_type, const char* node_name) : node_builder(node_name, op_type, g->graph.op_registry()), graph(g) {} tensorflow::NodeBuilder node_builder; TF_Graph* graph; std::set<tensorflow::string> colocation_constraints; }; struct TF_Operation { tensorflow::Node node; private:
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat May 13 00:49:12 UTC 2023 - 7.6K bytes - Viewed (0) -
tensorflow/compiler/jit/partially_decluster_pass.cc
continue; } DeviceType device_type(""); TF_RETURN_IF_ERROR( DeviceNameToDeviceType(n->assigned_device_name(), &device_type)); TF_RETURN_IF_ERROR(MemoryTypesForNode(graph.op_registry(), device_type, n->def(), &input_mtypes, &output_mtypes)); for (const Edge* e : n->out_edges()) { Node* dst = e->dst();
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Feb 09 11:36:41 UTC 2024 - 15.7K bytes - Viewed (0) -
tensorflow/cc/framework/scope.cc
Scope Scope::NewRootScope() { Graph* graph = new Graph(OpRegistry::Global()); ShapeRefiner* refiner = new ShapeRefiner(graph->versions(), graph->op_registry()); return Scope(new Impl(graph, new Status, new Impl::NameMap, refiner, /* disable_shape_inference */ false)); } Scope Scope::DisabledShapeInferenceScope() { Graph* graph = new Graph(OpRegistry::Global()); ShapeRefiner* refiner =
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Apr 13 05:57:22 UTC 2024 - 20.9K bytes - Viewed (0) -
tensorflow/compiler/jit/encapsulate_subgraphs_pass_test.cc
add2.node()->AddAttr("_cluster", "cluster2"); auto out = ops::Mul(root.WithOpName("mul"), add1, add2); Graph graph_before_encapsulation(OpRegistry::Global()); TF_ASSERT_OK(root.ToGraph(&graph_before_encapsulation)); FunctionLibraryDefinition library(OpRegistry::Global(), FunctionDefLibrary()); std::unique_ptr<Graph> graph; TF_ASSERT_OK(EncapsulateSubgraphsInFunctions( "_cluster", graph_before_encapsulation,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 08:47:20 UTC 2024 - 113.3K bytes - Viewed (0) -
tensorflow/compiler/jit/build_xla_ops_pass_test.cc
Status BuildXlaOps(const Scope& s, const FunctionDefLibrary& fdef_lib, std::unique_ptr<Graph>* result) { auto graph = std::make_unique<Graph>(OpRegistry::Global()); TF_RETURN_IF_ERROR(s.ToGraph(graph.get())); FunctionLibraryDefinition flib_def(graph->op_registry(), fdef_lib); // Assign all nodes to the CPU device. static const char* kCpuDevice = "/job:localhost/replica:0/task:0/cpu:0"; for (Node* n : graph->nodes()) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 08:47:20 UTC 2024 - 12.2K bytes - Viewed (0) -
tensorflow/compiler/jit/shape_inference.cc
const tensorflow::FunctionLibraryDefinition* fnlib_def, GraphShapeInfo* shape_info) { ShapeRefiner shape_refiner(graph->versions(), graph->op_registry()); shape_refiner.set_require_shape_inference_fns(false); // TODO(dlibenzi): Verify if it is worth trying to infer shaped within // functions. Some functions can be called at multiple locations with
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 31 00:41:19 UTC 2024 - 13K bytes - Viewed (0) -
tensorflow/compiler/jit/mark_for_compilation_pass_test.cc
std::unique_ptr<Graph> graph(new Graph(OpRegistry::Global())); { GraphDefBuilder builder(GraphDefBuilder::kFailImmediately); Node* a = ops::SourceOp("UncompilableNullary", builder.opts().WithName("A")); // Builds a Symbolic gradient for Supported NodeBuilder b_builder("B", "SymbolicGradient", builder.opts().op_registry()); NameAttrList b_name_attr;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 14 10:11:10 UTC 2024 - 79.6K bytes - Viewed (0) -
tensorflow/compiler/jit/encapsulate_subgraphs_pass.cc
return results_.at(OutputTensor(edge->src(), edge->src_output())); } Node* Encapsulator::Subgraph::MakeNodeImage(const Graph* graph_in, Node* node) { if (!graph_) { graph_.reset(new Graph(graph_in->op_registry())); graph_->set_versions(graph_in->versions()); } // TODO(b/116981129): Enhance how the device for the encapsulated subgraph is // determined. In case of hard placement, ensure all the encapsulated nodes
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 08:47:20 UTC 2024 - 51K bytes - Viewed (0) -
tensorflow/compiler/jit/encapsulate_xla_computations_pass.cc
e->src()->name(), ":", e->src_output(), " -> ", e->dst()->name(), ":", e->dst_input()); } } auto output = std::make_unique<Graph>((*graph)->op_registry()); TF_RETURN_WITH_CONTEXT_IF_ERROR( EncapsulateSubgraphsInFunctions( kXlaClusterIdAttr, **graph, RewriteSubgraph, /*reuse_existing_functions=*/true, &output, flib_def),
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Mar 12 06:33:33 UTC 2024 - 15.1K bytes - Viewed (0)