- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 27 for input_node (0.51 sec)
-
tensorflow/compiler/mlir/lite/sparsity/sparsify_model.cc
namespace mlir { namespace lite { absl::Status SparsifyModel(const tflite::ModelT& input_model, flatbuffers::FlatBufferBuilder* builder) { MLIRContext context; StatusScopedDiagnosticHandler statusHandler(&context, /*propagate=*/true); // Import input_model to a MLIR module flatbuffers::FlatBufferBuilder input_builder;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Jun 10 20:16:40 UTC 2024 - 4.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/python/representative_dataset.py
' for representative samples.' ) sample = _RepresentativeDataSample() for input_name, input_value in repr_sample.items(): sample.tensor_proto_inputs[input_name].CopyFrom( tensor_util.make_tensor_proto(input_value) ) writer.write(sample.SerializeToString()) logging.info(
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Mar 22 22:55:22 UTC 2024 - 14.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/quantization/lite/quantize_weights.h
const tflite::Model* input_model, int64_t weights_min_num_elements, bool use_hybrid_evaluation = true); TfLiteStatus QuantizeWeights(flatbuffers::FlatBufferBuilder* builder, const tflite::Model* input_model, BufferType quant_type = BufferType::QUANTIZED_INT8,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 23:15:24 UTC 2024 - 4.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/common/quantization_lib/quantization_config.cc
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Mar 05 07:39:40 UTC 2024 - 5.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/sparsity/sparsify_model.h
#include "tensorflow/compiler/mlir/lite/schema/schema_generated.h" namespace mlir { namespace lite { // Sparsify the `input_model` and write the result to a flatbuffer `builder`. absl::Status SparsifyModel(const tflite::ModelT& input_model, flatbuffers::FlatBufferBuilder* builder); } // namespace lite } // namespace mlir
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Jun 10 20:16:40 UTC 2024 - 1.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/cc/convert_asset_args.cc
if (!asset_op) continue; const StringRef input_name = input_names[arg_idx]; ConvertMainArgAttrs(main_func_op, arg_idx, /*index_path=*/input_name); // This assumes that the final tensor name in `GraphDef` is equal to // `input_name`. asset_file_defs.emplace_back(CreateAssetFileDef( asset_op.getFilenameAttr(), /*tensor_name=*/input_name)); } return asset_file_defs; }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 5.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/python/save_model.py
for idx, input_name in enumerate(node.input): if input_name in output_renaming_map: node.input[idx] = output_renaming_map[input_name] # Update the control inputs referring to the pre-renaming node. updating_inputs = [] for input_name in reversed(node.input): if input_name.startswith('^') and input_name[1:] in output_renaming_map:
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 01:09:50 UTC 2024 - 12.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/insert_main_function.cc
for (StringRef input_name : fn_input_names) { if (input_name_set.contains(input_name)) { // Found a duplicated name, all input names will be prefixed by // their corresponding function names. need_prefix_for_input_name = true; } input_name_set.insert(input_name); fn_input_name_vec.push_back(function_name); }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 16.5K bytes - Viewed (0) -
tensorflow/cc/framework/gradients.cc
} std::unordered_set<int> stop_backprop_nodes = GetStopBackpropNodes(reachable_nodes, output_nodes); // Populate `input_nodes_` from Outputs in `inputs_`. input_nodes_.reserve(inputs_.size()); for (size_t i = 0; i < inputs_.size(); ++i) { input_nodes_.insert({inputs_[i], i}); } // TODO(andydavis) Consider a more efficient data structure for `pending_` to
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Apr 13 05:57:22 UTC 2024 - 22K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/stablehlo/odml_to_stablehlo.cc
// * Options for full/partial conversion, Op exceptions list. // * Option to serialize output to TFL flatbuffer format. using llvm::cl::opt; // NOLINTNEXTLINE opt<std::string> input_model(llvm::cl::Positional, llvm::cl::desc("<input model path>"), llvm::cl::Required); // NOLINTNEXTLINE
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 03 18:16:49 UTC 2024 - 14.1K bytes - Viewed (0)