Search Options

Results per page
Sort
Preferred Languages
Advance

Results 31 - 40 of 48 for Quantile (0.17 sec)

  1. tensorflow/compiler/mlir/quantization/stablehlo/tests/passes/post_quantize.mlir

    // RUN: stablehlo-quant-opt %s -split-input-file -stablehlo-post-quantize | FileCheck %s
    
    // CHECK-LABEL: @remove_volatile_qdq
    func.func @remove_volatile_qdq() -> tensor<3x2xf32> {
      // CHECK: %[[CST:.*]] = stablehlo.constant
      // CHECK-NOT: "quantfork.qcast"
      // CHECK-NOT: "quantfork.dcast"
      // CHECK: return %[[CST]]
      %cst = stablehlo.constant dense<[[-0.960978984, -0.390246302], [-0.790828585, -0.601039409], [-1.0280807, -1.02731466]]> : tensor<3x2xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 18 20:32:46 UTC 2024
    - 4.4K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/quantization/lite/quantize_weights.h

    namespace mlir {
    namespace lite {
    
    // Supported resulting types from quantization process.
    enum class BufferType { QUANTIZED_INT8, QUANTIZED_FLOAT16 };
    
    // Stores information about how to quantize a user-specified custom operation.
    // CustomOpInfo contains info of its corresponding CustomOp registered in the
    // CustomOpMap. 'quantizable_input_indices' is used to determine which indices
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 23:15:24 UTC 2024
    - 4.2K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/quantization/tensorflow/quantization_options.proto

    // Defines various options to specify and control the behavior of the quantizer.
    // It consists of
    // 1) Model-wise quantization configuration as a default configuration. If it is
    // None, the default configuration is "do not quantize the model".
    // 2) A set of supported operations.
    // 3) Unit wise quantization precision.
    // 4) Target hardware name.
    // NEXT ID: 18
    message QuantizationOptions {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Mar 19 06:31:19 UTC 2024
    - 9.2K bytes
    - Viewed (0)
  4. cmd/admin-server-info.go

    					}
    				}
    			}
    		}
    	}
    
    	var memstats runtime.MemStats
    	runtime.ReadMemStats(&memstats)
    
    	gcStats := debug.GCStats{
    		// If stats.PauseQuantiles is non-empty, ReadGCStats fills
    		// it with quantiles summarizing the distribution of pause time.
    		// For example, if len(stats.PauseQuantiles) is 5, it will be
    		// filled with the minimum, 25%, 50%, 75%, and maximum pause times.
    		PauseQuantiles: make([]time.Duration, 5),
    	}
    Registered: Sun Jun 16 00:44:34 UTC 2024
    - Last Modified: Fri May 24 23:05:23 UTC 2024
    - 4.9K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/quantization/stablehlo/tests/passes/quantize/quantize_weight_only.mlir

    // RUN: stablehlo-quant-opt %s -split-input-file -stablehlo-quantize | FileCheck %s
    
    // Test that hybrid quantized dot_general is produced when q/dq pair only exists
    // for weight.
    
    module attributes {tf_saved_model.semantics} {
      func.func private @quantize_dot_general_fn(%arg0: tensor<1x2xf32>) -> tensor<1x3xf32> attributes {tf._original_func_name = "main_0"} {
        %cst = stablehlo.constant dense<3.000000e-01> : tensor<2x3xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 14 17:10:32 UTC 2024
    - 4.8K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/lite/common/tfl_pass_config.h

      // have side effects e.g. reduced flatbuffer size. Only certain type
      // conversions are supported.
      bool reduce_type_precision = false;
      // Whether to consider this model a quantized model with quantize/dequantize
      // ops and to convert kernels to quantized kernels wherever appropriate.
      quant::QDQConversionMode qdq_conversion_mode =
          quant::QDQConversionMode::kQDQNone;
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 08 19:05:30 UTC 2024
    - 6.5K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/quantization/stablehlo/cc/calibration/component.h

      const std::unordered_set<std::string> tags_;
    
      const absl::flat_hash_map<std::string, tensorflow::SignatureDef>
          signature_def_map_;
    
      // Signature keys to identify the functions to load & quantize.
      const std::vector<std::string> signature_keys_;
    };
    
    // Runs passes to prepare the calibration model.
    absl::Status RunCalibrationPasses(mlir::ModuleOp module_op, MLIRContext& ctx,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 14 06:31:57 UTC 2024
    - 5.4K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/quantization/common/uniform_quantized_types.h

    // whose storage type is 32-bit integer and expressed type is f32.
    bool IsI32F32UniformQuantizedPerAxisType(Type type);
    
    // Determines whether the storage type of a quantized type is supported by
    // `tfl.quantize` or `tfl.dequantize` ops. ui8, i8 and i16 are supported.
    bool IsSupportedByTfliteQuantizeOrDequantizeOps(IntegerType storage_type);
    
    // Returns true if a type is quantized tensor type.
    bool IsQuantizedTensorType(Type type);
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 5.7K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/lite/quantization/lite/quantize_model.cc

      pm.addPass(TFL::CreatePostQuantizeRemoveQDQPass());
      if (failed(pm.run(module.get()))) {
        const std::string err(statusHandler.ConsumeStatus().message());
        LOG(ERROR) << "Failed to quantize: " << err;
        return kTfLiteError;
      }
    
      // Export the results.
      tflite::FlatbufferExportOptions options;
      options.toco_flags.set_force_select_tf_ops(false);
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 23:15:24 UTC 2024
    - 6.3K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/quantization/common/attrs_and_constraints.cc

      if (dot_general_op == nullptr) return std::nullopt;
      const int64_t filter_rank =
          mlir::dyn_cast<ShapedType>(dot_general_op.getOperand(1).getType())
              .getRank();
    
      // To quantize rhs per-channel, we currently only consider the case where
      // `stablehlo.dot_general` is legalizable to `tfl.fully_connected`.
      const bool is_per_axis_quantizable =
          IsDotGeneralFullyConnected(dot_general_op).value();
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 6.8K bytes
    - Viewed (0)
Back to top