- Sort Score
- Result 10 results
- Languages All
Results 61 - 70 of 660 for weights (0.14 sec)
-
tensorflow/compiler/mlir/lite/utils/const_tensor_utils.cc
} storage_type = mlir::cast<mlir::IntegerType>(raw_elem_type); } // TFlite uses narrow-range [u]int8 for constant buffers of quantized weights. // Since we don't know which ones are weights, we represent this optimization // as a change in the storage bounds for the type for all constants of this // type. const int bitwidth = storage_type.getIntOrFloatBitWidth();
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 07 23:04:40 UTC 2024 - 16.6K bytes - Viewed (0) -
pkg/scheduler/apis/config/types_pluginargs.go
// Type selects which strategy to run. Type ScoringStrategyType // Resources to consider when scoring. // The default resource set includes "cpu" and "memory" with an equal weight. // Allowed weights go from 1 to 100. // Weight defaults to 1 if not specified or explicitly set to 0. Resources []ResourceSpec // Arguments specific to RequestedToCapacityRatio strategy.
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Fri Jan 13 23:15:53 UTC 2023 - 8.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/g3doc/space_to_depth.md
c * (block_size ** 2)]) ``` `SpaceToDepthOp` can be called on the host to perform the transform. 1. Weight Transformation Weight Transformation is similar to Input Transform. Weight transform is needed to apply space to depth optimization for a model that needs to load a pre-train checkpoint. This transform can be done on the host or TPU device
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Oct 24 02:51:43 UTC 2020 - 8.3K bytes - Viewed (0) -
src/cmd/vendor/golang.org/x/text/language/parse.go
continue } entry, weight := split(entry, ';') // Scan the language. t, err := Parse(entry) if err != nil { id, ok := acceptFallback[entry] if !ok { return nil, nil, err } t = makeTag(language.Tag{LangID: id}) } // Scan the optional weight. w := 1.0 if weight != "" { weight = consume(weight, 'q') weight = consume(weight, '=')
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Jan 24 13:01:26 UTC 2024 - 7.5K bytes - Viewed (0) -
pilot/pkg/networking/core/networkfilter_test.go
Registered: Fri Jun 14 15:00:06 UTC 2024 - Last Modified: Wed Apr 17 22:20:44 UTC 2024 - 25.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/tensorflow/passes/prepare_quantize_drq.cc
OpSet op_set_; Option<bool> enable_per_channel_quantization_{ *this, "enable-per-channel-quantization", llvm::cl::init(false), llvm::cl::desc("Whether enable per-channel quantized weights.")}; }; // If the weight is applicable to dynamic range quantization, insert Quantize // and Dequantize ops with per-tensor scale. class PrepareDRQQuantizableOp : public OpRewritePattern<arith::ConstantOp> { public:
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 11.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/schema/schema_v3b.fbs
SPARSE = 1, DENSE = 2, } table LSHProjectionOptions { type: LSHProjectionType; } table SVDFOptions { rank:int; fused_activation_function:ActivationFunctionType; // For weights-only quantization, use asymmetric quantization for non // constant inputs at evaluation time. asymmetric_quantize_inputs:bool; } // An implementation of TensorFlow RNNCell. table RNNOptions {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 28 14:28:27 UTC 2024 - 30K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/optimize.mlir
// NoFusing-LABEL: FuseMulWithFullyConnectedNoBias // NoFusing-DAG: %[[MWEIGHTS:.*]] = arith.constant dense<2.000000e+00> : tensor<512xf32> // NoFusing-DAG: %[[WEIGHTS:.*]] = arith.constant dense<3.000000e+00> : tensor<1024x512xf32> // NoFusing-DAG: %[[BIAS:.*]] = "tfl.no_value"() <{value}> : () -> none
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 16 20:31:41 UTC 2024 - 284.1K bytes - Viewed (0) -
pkg/test/loadbalancersim/lb_test.go
wg := sync.WaitGroup{} clientLatencies := make([]timeseries.Data, len(s.mesh.Clients())) for i, client := range s.mesh.Clients() { i := i client := client wg.Add(1) go func() { // Assign weights to the endpoints. var conns []*loadbalancer.WeightedConnection for _, n := range s.mesh.Nodes() { conns = append(conns, s.newWeightedConnection(client, n)) } // Create a load balancer
Registered: Fri Jun 14 15:00:06 UTC 2024 - Last Modified: Thu May 19 23:29:30 UTC 2022 - 11K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tf_to_tfl_flatbuffer.cc
&q_builder, input_model, quantized_type, use_updated_hybrid_scheme, ::tflite::optimize::QuantizerType::OLD_QUANTIZER) != kTfLiteOk) { return absl::InvalidArgumentError( "Quantize weights transformation failed."); } const uint8_t* q_buffer = q_builder.GetBufferPointer(); *result = std::string(reinterpret_cast<const char*>(q_buffer), q_builder.GetSize()); return absl::OkStatus(); }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 03 18:01:23 UTC 2024 - 23.8K bytes - Viewed (0)