Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 5 of 5 for debugger_type (0.2 sec)

  1. tensorflow/compiler/mlir/quantization/stablehlo/quantization_config.proto

        // both the quantized and unquantized layer.
        DEBUGGER_TYPE_FLOAT_PER_LAYER = 3;
      }
    
      DebuggerType debugger_type = 1;
    
      // Path to save unquantized model with dump tensor ops attached.
      // Used when debugger_type is WHOLE_MODEL.
      string unquantized_dump_model_path = 2;
    
      // Path to save debugger related logs. Defaults to '/tmp/dumps'.
      string log_dir_path = 3;
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 03:36:50 UTC 2024
    - 14.3K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/quantization/tensorflow/quantize_passes.cc

      // TODO: b/295140328 - Add debugger support for weight only
      if (quantization_options.has_debugger_config()) {
        pm.addPass(mlir::quant::CreateAddDumpTensorOpPass(
            quantization_options.debugger_config().debugger_type(),
            quantization_options.debugger_config().log_dir_path()));
      }
      pm.addNestedPass<mlir::func::FuncOp>(
          mlir::quant::CreateInsertCustomAggregationOpsPass(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 10 04:07:09 UTC 2024
    - 9.9K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/quantization/tensorflow/passes/passes.h

    // Create a pass that inserts dump tensor to quantizable layer's output.
    std::unique_ptr<OperationPass<ModuleOp>> CreateAddDumpTensorOpPass(
        ::stablehlo::quantization::DebuggerConfig::DebuggerType debugger_type,
        std::string log_dir_path);
    
    // Creates a pass that add QuantizationUnitLoc to quantizable layers.
    std::unique_ptr<OperationPass<func::FuncOp>> CreateAddQuantizationUnitLocPass();
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 10 04:07:09 UTC 2024
    - 12.3K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/tensorflow/python/quantize_model.py

            quantization_options.debugger_config.debugger_type
            == stablehlo_quant_config_pb2.DebuggerConfig.DebuggerType.DEBUGGER_TYPE_UNSPECIFIED
        ):
          raise ValueError(
              'Debugger is enabled but debugger type was not specified.'
          )
    
        if (
            quantization_options.debugger_config.debugger_type
            == stablehlo_quant_config_pb2.DebuggerConfig.DebuggerType.DEBUGGER_TYPE_WHOLE_MODEL
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 03:36:50 UTC 2024
    - 34.2K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/quantization/tensorflow/python/quantize_model.cc

      // both quantized and unquantized tensors and compare them offline.
      if (quantization_options.has_debugger_config() &&
          quantization_options.debugger_config().debugger_type() ==
              DebuggerConfig::DEBUGGER_TYPE_WHOLE_MODEL) {
        TF_ASSIGN_OR_RETURN(
            ExportedModel debugging_exported_model,
            ExportDebuggingModel(*module_ref, context.get(), quantization_options,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 03:36:50 UTC 2024
    - 23.8K bytes
    - Viewed (0)
Back to top