- Sort Score
- Result 10 results
- Languages All
Results 41 - 50 of 136 for weighted (0.33 sec)
-
tensorflow/compiler/mlir/quantization/tensorflow/cc/constant_fold_test.cc
%scale = "tf.Const"() {value = dense<2.0> : tensor<f32>} : () -> tensor<f32> %weight = "tf.Const"() {value = dense<1> : tensor<1024x24x24x3xi8>} : () -> tensor<1024x24x24x3xi8> %input_i32 = "tf.Cast"(%weight) : (tensor<1024x24x24x3xi8>) -> tensor<1024x24x24x3xi32> %output = "tf.Sub"(%input_i32, %zp) : (tensor<1024x24x24x3xi32>, tensor<i32>) -> tensor<1024x24x24x3xi32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 04 07:19:09 UTC 2024 - 10.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/quantization_config.proto
// value {dimension_specs {dimension: 3}}} // }} // } // ``` // // This preset: // * Applies per-channel quantization for weights (input index 1) of // convolution quantizable unit family. The quantization dimension is 3, the // channel dimension, which assumes the weight tensor is in NHWC format. // * Applies static-range PTQ for all other ops. // // Next ID: 4 message StaticRangePtqPreset {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 03:36:50 UTC 2024 - 14.3K bytes - Viewed (0) -
pkg/scheduler/extender_test.go
Prioritizers: []tf.PriorityConfig{{Function: tf.Node1PrioritizerExtender, Weight: 10}}, Weight: 1, }, { ExtenderName: "FakeExtender2", Predicates: []tf.FitPredicate{tf.TruePredicateExtender}, Prioritizers: []tf.PriorityConfig{{Function: tf.Node2PrioritizerExtender, Weight: 10}}, Weight: 5, }, }, nodes: []string{"node1", "node2"},
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Mon Feb 26 19:07:19 UTC 2024 - 16.7K bytes - Viewed (0) -
pkg/scheduler/testing/framework/fake_extender.go
for _, prioritizer := range f.Prioritizers { weight := prioritizer.Weight if weight == 0 { continue } priorityFunc := prioritizer.Function prioritizedList, err := priorityFunc(pod, nodes) if err != nil { return &extenderv1.HostPriorityList{}, 0, err } for _, hostEntry := range *prioritizedList { combinedScores[hostEntry.Name] += hostEntry.Score * weight } }
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Mon Feb 26 19:07:19 UTC 2024 - 13.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/experimental/tac/transforms/device_transform_patterns.cc
// TFL fully_connected basically does: // Weight * Input + bias. // Input layout is : [..., depth] // Weight layout is : [output, depth] // Bias is [output]. // // While conv2d is: // Filter: [NHWC] // Input is also: [NHWC] // Bias is [N] // // So to perform the transform, we need to insert a few reshape ops: // // Input weight bias // \ / / // FC // |
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 25.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/tests/legalize-tf-quant.mlir
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 09 01:25:29 UTC 2024 - 37.3K bytes - Viewed (0) -
pkg/scheduler/apis/config/validation/validation.go
extenderManagedResources := sets.New[string]() for i, extender := range extenders { path := fldPath.Index(i) if len(extender.PrioritizeVerb) > 0 && extender.Weight <= 0 { errs = append(errs, field.Invalid(path.Child("weight"), extender.Weight, "must have a positive weight applied to it")) } if extender.BindVerb != "" { binders++ } for j, resource := range extender.ManagedResources {
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Thu Apr 25 06:27:01 UTC 2024 - 11.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/prepare_lifting.cc
return ConstantFoldOpIfPossible(reshape_op).front(); } // Checks if a value can be symmetrically quantized. bool CanBeSymmetricallyQuantized(Value weight) { auto dq_op = weight.getDefiningOp<quantfork::DequantizeCastOp>(); if (!dq_op) return true; auto qtype = mlir::cast<TensorType>(dq_op.getArg().getType()).getElementType();
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 17:58:54 UTC 2024 - 13.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/quantize.cc
// quantizable ops, Q-DQ ops need to be preserved. bool shouldKeepUnusedQdqPattern(); void runOnOperation() override; private: QuantizationSpecs quant_specs_; Option<bool> weight_quantization_{ *this, "weight-quantization", llvm::cl::init(false), llvm::cl::desc("Whether to enable weight quantization.")}; Option<OpSet> target_opset_{
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Mar 22 05:52:39 UTC 2024 - 23.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/prepare_quantize.cc
// prepare_quantize_ptq_per_channel.mlir. Option<bool> enable_per_channel_quantization_{ *this, "enable-per-channel-quantization", llvm::cl::init(false), llvm::cl::desc("Whether enable per-channel quantized weights.")}; }; bool PrepareQuantizePass::SetInputNodesQuantizationParams(func::FuncOp func) { StringRef func_name = func.getName(); auto has_quantize_op = [&](const Value arg) { return (arg.hasOneUse() &&
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 17.2K bytes - Viewed (0)