- Sort Score
- Result 10 results
- Languages All
Results 51 - 60 of 499 for weighted (0.13 sec)
-
src/internal/profile/graph.go
ret += edge.Weight } return ret } type edgeList []*Edge func (el edgeList) Len() int { return len(el) } func (el edgeList) Less(i, j int) bool { if el[i].Weight != el[j].Weight { return abs64(el[i].Weight) > abs64(el[j].Weight) } from1 := el[i].Src.Info.PrintableName() from2 := el[j].Src.Info.PrintableName() if from1 != from2 { return from1 < from2 }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Feb 05 20:59:15 UTC 2024 - 13.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/common/quantization_lib/quantization_config.h
// Whether to allow weight-only quantization. This scheme quantizes // weights but will dequantize them back at runtime which is useful for // memory bound case without kernel support available in lower precisions. // Used in MLIR dynamic range quantizer. bool weight_only_quantization = false; // The minimum number of elements in a weights array required to apply
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Mar 13 10:16:19 UTC 2024 - 10.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/passes/insert_weight_param.cc
namespace { using ::stablehlo::quantization::Method; using ::stablehlo::quantization::QuantizedType; using ::stablehlo::quantization::WeightOnlyPtq; // Inserts quantization parameters of weights for weight-only quantization and // dynamic range quantization of `stablehlo.convolution` and // `stablehlo.dot_general`. class InsertWeightParamPass : public impl::InsertWeightParamPassBase<InsertWeightParamPass> { public:
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 09 05:56:10 UTC 2024 - 10.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/end2end/fake_quant_without_identity.pbtxt
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 13.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/end2end/fake_quant_without_identity_4bit.pbtxt
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 13.8K bytes - Viewed (0) -
guava/src/com/google/common/cache/LocalCache.java
ValueReference<K, V> previous = entry.getValueReference(); int weight = map.weigher.weigh(key, value); checkState(weight >= 0, "Weights must be non-negative"); ValueReference<K, V> valueReference = map.valueStrength.referenceValue(this, entry, value, weight); entry.setValueReference(valueReference); recordWrite(entry, weight, now); previous.notifyNewValue(value); }
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Sat May 18 03:24:34 UTC 2024 - 149.2K bytes - Viewed (0) -
pilot/pkg/config/kube/gateway/conversion.go
if forwardTo == nil { return nil, nil, nil } weights := []int{} action := []k8s.BackendRef{} for _, w := range forwardTo { wt := int(ptr.OrDefault(w.Weight, 1)) if wt == 0 { continue } action = append(action, w) weights = append(weights, wt) } if len(weights) == 1 { weights = []int{0} } var invalidBackendErr *ConfigError
Registered: Fri Jun 14 15:00:06 UTC 2024 - Last Modified: Fri Jun 14 04:34:37 UTC 2024 - 84.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/quantization_options.proto
// determined. The activation and weight are quantized to INT8 while bias is // quantized to INT32. METHOD_STATIC_RANGE_INT8 = 2; // Dynamic range quantization. Quantized tensor values' ranges are // determined in the graph executions. The weights are quantized during // conversion. METHOD_DYNAMIC_RANGE_INT8 = 3; // Weight-only quantization. Only weights are quantized during conversion.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Mar 19 06:31:19 UTC 2024 - 9.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/utils/lstm_utils.h
// that also contains other supporting ops needed to construct the operands for // the fused op. The caller provides the containing FuncOp as input with // arguments specifying the input, weight, projection and bias. // The weight, projection, bias and layer norm scale all need to be // RankedTensorType. // This class sets the layer norm coefficients to NoneType. class ConvertLSTMCellSimpleToFusedLSTM { public:
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Jun 03 00:14:05 UTC 2023 - 7.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/end2end/fake_quant_per_channel.pbtxt
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 18.1K bytes - Viewed (0)