- Sort Score
- Result 10 results
- Languages All
Results 1 - 7 of 7 for LOG_FIRST_N (0.26 sec)
-
tensorflow/compiler/mlir/tfr/integration/graph_decompose_pass.cc
LOG_FIRST_N(INFO, 1) << "Skipping Graph Decomposition Pass, decomposition" " library was not found"; return absl::OkStatus(); } tf_core_op_expansion_graph_counter->GetCell()->IncrementBy(1); LOG_FIRST_N(INFO, 1) << "Run Graph Decomposition Passes"; TF_RETURN_IF_ERROR(DecomposeGraph(module)); LOG_FIRST_N(INFO, 1) << "Finish Graph Decomposition Passes";
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sun Feb 25 16:22:36 UTC 2024 - 2.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfr/integration/node_expansion_pass.cc
}; for (const char* skip : kOpsToSkip) { if (absl::StartsWith(orig_op->op_name(), skip)) return absl::OkStatus(); } tf_core_op_expansion_node_counter->GetCell()->IncrementBy(1); LOG_FIRST_N(INFO, 1) << "Run Node Expansion Passes"; // Get the FunctionDef and insert that into the context const NodeDef& ndef = orig_op->MutableAttrs()->BuildNodeDef(); auto& ctx = orig_op->EagerContext();
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sun Feb 25 16:22:36 UTC 2024 - 3.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/mlir_graph_optimization_pass.cc
if (overall_state == MlirOptimizationPassState::Disabled) { if (VLOG_IS_ON(1)) { LOG_FIRST_N(INFO, 1) << "None of the MLIR Optimization Passes are enabled " << "(registered " << registry_->passes().size() << ")"; } return absl::OkStatus(); } if (VLOG_IS_ON(1)) { LOG_FIRST_N(INFO, 1) << "MLIR Graph Optimization Passes." << " Enabled: " << num_passes_enabled
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 22:19:26 UTC 2024 - 18.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/internal/legalize_tf_to_hlo.cc
std::vector<std::vector<xla::Shape>>* per_core_arg_shapes, std::vector<std::unique_ptr<mlir::Pass>>& custom_legalization_passes, xla::CompileOnlyClient* client, XlaCompilationResult* compilation_result) { LOG_FIRST_N(INFO, 1) << "Compiling MLIR computation to XLA HLO using the " "Combined MLIR Tf2Xla Bridge."; absl::StatusOr<std::string> mlir_compilation = internal::CompileFromMlirToXlaHlo(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sun Apr 14 20:29:34 UTC 2024 - 3.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/internal/legalize_tf_mlir.cc
const std::vector<TensorShape>& arg_shapes, std::vector<ShardingAndIndex>* arg_core_mapping, std::vector<std::vector<xla::Shape>>* per_core_arg_shapes) { LOG_FIRST_N(INFO, 1) << "Compiling MLIR computation to XLA HLO using MLIR tf2xla bridge in " "the op by op fallback mode. This is Phase 2 of the TF2XLA Bridge. "
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sun Apr 14 20:29:34 UTC 2024 - 6.1K bytes - Viewed (0) -
tensorflow/compiler/jit/pjrt_device_context.cc
device_layout); if (first_try_buffer.ok()) { return std::move(*first_try_buffer); } if (first_try_buffer.status().code() == absl::StatusCode::kUnimplemented) { LOG_FIRST_N(WARNING, 1) << first_try_buffer.status() << "; fallback to BufferFromHostBuffer without device layout."; TF_ASSIGN_OR_RETURN( std::unique_ptr<xla::PjRtBuffer> second_try_buffer,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Apr 13 08:49:31 UTC 2024 - 11.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/api/v1/compile_tf_graph.cc
std::vector<tpu::ShardingAndIndex>* arg_core_mapping, std::vector<std::vector<xla::Shape>>* per_core_arg_shapes, xla::CompileOnlyClient* client, XlaCompiler::CompilationResult* compilation_result) { LOG_FIRST_N(INFO, 1) << "Compiling MLIR computation to XLA HLO using the " "old (non-MLIR) tf2xla bridge"; CompilationTimer timer; *compilation_result = {};
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 22:19:26 UTC 2024 - 14K bytes - Viewed (0)