- Sort Score
- Result 10 results
- Languages All
Results 51 - 60 of 136 for ROUND (0.89 sec)
-
docs/metrics/v3.md
| `minio_cluster_iam_plugin_authn_service_succ_avg_rtt_ms_minute` | `counter` | When plugin authentication is configured, returns average round-trip-time of successful requests in the last full minute | | | `minio_cluster_iam_plugin_authn_service_succ_max_rtt_ms_minute` | `counter` | When plugin authentication is configured, returns maximum round-trip-time of successful requests in the last full minute | |
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Thu Jun 06 09:36:25 UTC 2024 - 40.9K bytes - Viewed (0) -
src/archive/tar/writer.go
tw.hdr.Typeflag = TypeReg } } // Round ModTime and ignore AccessTime and ChangeTime unless // the format is explicitly chosen. // This ensures nominal usage of WriteHeader (without specifying the format) // does not always result in the PAX format being chosen, which // causes a 1KiB increase to every header. if tw.hdr.Format == FormatUnknown { tw.hdr.ModTime = tw.hdr.ModTime.Round(time.Second) tw.hdr.AccessTime = time.Time{}
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 01:00:11 UTC 2024 - 19.6K bytes - Viewed (0) -
src/encoding/json/decode.go
// can be set to nil. func indirect(v reflect.Value, decodingNull bool) (Unmarshaler, encoding.TextUnmarshaler, reflect.Value) { // Issue #24153 indicates that it is generally not a guaranteed property // that you may round-trip a reflect.Value by calling Value.Addr().Elem() // and expect the value to still be settable for values derived from // unexported embedded struct fields. //
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 00:18:55 UTC 2024 - 35.3K bytes - Viewed (0) -
pkg/controller/cronjob/utils.go
// minute - 59, hour - 23, dom - 31, month - 2, and dow is optional, clearly 31 is invalid // In this case the timeBetweenTwoSchedules will be 0, and we error out the invalid schedule timeBetweenTwoSchedules := int64(t2.Sub(t1).Round(time.Second).Seconds()) if timeBetweenTwoSchedules < 1 { return earliestTime, nil, missedSchedules, fmt.Errorf("time difference between two schedules is less than 1 second") }
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Thu May 09 03:34:25 UTC 2024 - 10.7K bytes - Viewed (0) -
staging/src/k8s.io/apimachinery/pkg/api/resource/quantity.go
// of an amount. Arguably, this should be inf.RoundHalfUp (normal rounding), but that would have // the side effect of rounding values < .5n to zero. if v, ok := amount.Unscaled(); v != int64(0) || !ok { amount.Round(amount, Nano.infScale(), inf.RoundUp) } // The max is just a simple cap. // TODO: this prevents accumulating quantities greater than int64, for instance quota across a cluster
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed May 29 21:48:10 UTC 2024 - 23.8K bytes - Viewed (0) -
pkg/kubelet/images/image_gc_manager_test.go
require.NoError(t, err) require.Equal(t, len(images), 1) // Simulate pod having just used this image, but having been GC'd images[0].lastUsed = fakeClock.Now() // First GC round should not GC remaining image, as it was used too recently. assert := assert.New(t) oldStartTime := fakeClock.Now() images, err = manager.freeOldImages(ctx, images, oldStartTime, oldStartTime) require.NoError(t, err)
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed May 22 15:38:20 UTC 2024 - 29.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfr/passes/decompose.cc
OpBuilder builder) { double scale = mlir::cast<FloatAttr>(scale_attr).getValueAsDouble(); int64_t zp = mlir::cast<IntegerAttr>(zp_attr).getInt(); int quantized = static_cast<int>(std::round(value / scale) + zp); quantized = std::min(quantized, static_cast<int>(std::numeric_limits<int8_t>::max())); quantized = std::max(quantized, static_cast<int>(std::numeric_limits<int8_t>::min()));
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 14.6K bytes - Viewed (0) -
src/archive/tar/common.go
func alignSparseEntries(src []sparseEntry, size int64) []sparseEntry { dst := src[:0] for _, s := range src { pos, end := s.Offset, s.endOffset() pos += blockPadding(+pos) // Round-up to nearest blockSize if end != size { end -= blockPadding(-end) // Round-down to nearest blockSize } if pos < end { dst = append(dst, sparseEntry{Offset: pos, Length: end - pos}) } } return dst }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Mar 15 16:01:50 UTC 2024 - 24.7K bytes - Viewed (0) -
src/runtime/iface.go
// so that we are guaranteed an empty slot (for termination). newN := n * 2 // make it at most 50% full newN = 1 << sys.Len64(uint64(newN-1)) // round up to a power of 2 // Allocate the new table. newSize := unsafe.Sizeof(abi.TypeAssertCache{}) + uintptr(newN-1)*unsafe.Sizeof(abi.TypeAssertCacheEntry{}) newC := (*abi.TypeAssertCache)(mallocgc(newSize, nil, true))
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 22.5K bytes - Viewed (0) -
pkg/controller/statefulset/stateful_set_utils.go
// getStatefulSetMaxUnavailable calculates the real maxUnavailable number according to the replica count // and maxUnavailable from rollingUpdateStrategy. The number defaults to 1 if the maxUnavailable field is // not set, and it will be round down to at least 1 if the maxUnavailable value is a percentage. // Note that API validation has already guaranteed the maxUnavailable field to be >1 if it is an integer
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Jun 05 19:06:41 UTC 2024 - 26.7K bytes - Viewed (0)