Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 6 of 6 for DumpTensor (0.23 sec)

  1. tensorflow/compiler/mlir/quantization/tensorflow/tests/add_dump_tensor_op.mlir

    // IntPerLayer-DAG: "tf.DumpTensor"(%[[output1_quantized]]) <{enabled = true, file_name = "quantized_tensor_data.pb", func_name = "conv", log_dir_path = "/tmp/dumps/composite_conv2d_with_bias_and_relu6_fn_1", node_name = "Conv2D_1"}> : (tensor<*xf32>) -> ()
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Mar 22 22:55:22 UTC 2024
    - 37.9K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/quantization/tensorflow/tests/add_dump_tensor_op_stablehlo.mlir

    // WholeModel-DAG: "tf.DumpTensor"(%[[matmul0_q]]) <{enabled = true, file_name = "unquantized_tensor_data.pb", func_name = "composite_dot_general_with_bias_and_relu6_dynamic_fn_2", log_dir_path = "/tmp/dumps/composite_dot_general_with_bias_and_relu6_dynamic_fn_2", node_name = "_empty_node"}>...
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Mar 22 22:55:22 UTC 2024
    - 18K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/quantization/stablehlo/quantization_config.proto

      // NEXT ID: 4
      enum DebuggerType {
        DEBUGGER_TYPE_UNSPECIFIED = 0;
        // DEBUGGER_TYPE_WHOLE_MODEL creates two tf.Savedmodel - unquantized and
        // quantized model with DumpTensor added to outputs of quantizable layers.
        // The DumpTensor dumps entire value of its input to a specified file. When
        // DEBUGGER_TYPE_WHOLE_MODEL is used unquantized_dump_model_path has to be
        // specified.
        DEBUGGER_TYPE_WHOLE_MODEL = 1;
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 03:36:50 UTC 2024
    - 14.3K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/tensorflow/passes/add_dump_tensor_op.cc

        // the commandline for example).
        return "quant-add-dump-tensor-op";
      }
    
      StringRef getDescription() const final {
        // This is a brief description of the pass.
        return "Add DumpTensor ops after quantizable ops";
      }
    
      void getDependentDialects(DialectRegistry &registry) const override {
        registry.insert<TF::TensorFlowDialect>();
        registry.insert<quant::QuantizationDialect>();
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Mar 22 22:55:22 UTC 2024
    - 13K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/quantization/tensorflow/python/quantize_model.cc

      // debugger type is `DEBUGGER_TYPE_WHOLE_MODEL`. This is required
      // because in whole-model debugging mode the `DumpTensor` ops for the
      // unquantized tensors are only inserted in the unquantized model
      // whereas `DumpTensor` ops for the quantized tensors are only inserted
      // in the quantized model. Both models are required to be able to dump
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 03:36:50 UTC 2024
    - 23.8K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/quantization/tensorflow/python/BUILD

            "//tensorflow/compiler/mlir/quantization/tensorflow/cc:run_passes",
            "//tensorflow/compiler/mlir/quantization/tensorflow/debugging:dump_tensor_op",  # Required for DumpTensor op registration.
            "//tensorflow/compiler/mlir/tensorflow:mlir_import_options",
            "//tensorflow/compiler/mlir/tensorflow:translate_lib",
            "//tensorflow/compiler/mlir/tensorflow/transforms:tf_dialect_passes",
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 03:36:50 UTC 2024
    - 18.8K bytes
    - Viewed (0)
Back to top