- Sort Score
- Result 10 results
- Languages All
Results 31 - 40 of 74 for FLOAT64 (0.05 sec)
-
logger/sql.go
} case int, int8, int16, int32, int64, uint, uint8, uint16, uint32, uint64: vars[idx] = utils.ToString(v) case float32: vars[idx] = strconv.FormatFloat(float64(v), 'f', -1, 32) case float64: vars[idx] = strconv.FormatFloat(v, 'f', -1, 64) case string: vars[idx] = escaper + strings.ReplaceAll(v, escaper, escaper+escaper) + escaper default: rv := reflect.ValueOf(v)
Registered: Sun Sep 07 09:35:13 UTC 2025 - Last Modified: Thu Mar 21 08:00:02 UTC 2024 - 5K bytes - Viewed (0) -
cmd/callhome.go
// callhome was disabled or context was canceled return } // callhome running on a different node. // sleep for some time and try again. duration := max(time.Duration(r.Float64()*float64(globalCallhomeConfig.FrequencyDur())), // Make sure to sleep at least a second to avoid high CPU ticks. time.Second) time.Sleep(duration) } }() }
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 5.3K bytes - Viewed (0) -
tests/lru_test.go
if i%2 == 0 { l.Add(trace[i], trace[i]) } else { if _, ok := l.Get(trace[i]); ok { hit++ } else { miss++ } } } b.Logf("hit: %d miss: %d ratio: %f", hit, miss, float64(hit)/float64(hit+miss)) } func BenchmarkLRU_Freq_NoExpire(b *testing.B) { l := lru.NewLRU[int64, int64](8192, nil, 0) trace := make([]int64, b.N*2) for i := 0; i < b.N*2; i++ { if i%2 == 0 {
Registered: Sun Sep 07 09:35:13 UTC 2025 - Last Modified: Sun May 25 07:40:40 UTC 2025 - 10.4K bytes - Viewed (0) -
cmd/metrics-v2_test.go
histBuckets := []float64{0.50, 5.00} labels := []string{"PutObject", "CopyObject"} ttfbHist := prometheus.NewHistogramVec( prometheus.HistogramOpts{ Name: "s3_ttfb_seconds", Help: "Time taken by requests served by current MinIO server instance", Buckets: histBuckets, }, []string{"api"}, ) observations := []struct { val float64 label string }{ {
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Sun Mar 30 00:56:02 UTC 2025 - 6.8K bytes - Viewed (0) -
schema/field_test.go
ID int64 INT int INT8 int8 INT16 int16 INT32 int32 INT64 int64 UINT uint UINT8 uint8 UINT16 uint16 UINT32 uint32 UINT64 uint64 FLOAT32 float32 FLOAT64 float64 BOOL bool STRING string TIME time.Time BYTES []byte TypeAlias struct { ID INT `gorm:"column:fint"` INT8 `gorm:"column:fint8"` INT16 `gorm:"column:fint16"`
Registered: Sun Sep 07 09:35:13 UTC 2025 - Last Modified: Sat Feb 19 09:02:53 UTC 2022 - 12.7K bytes - Viewed (0) -
cmd/data-usage-utils.go
metrics = append(metrics, MetricV2{ Description: getClusterTransitionedBytesMD(), Value: float64(st.TotalSize), VariableLabels: map[string]string{"tier": tier}, }) metrics = append(metrics, MetricV2{ Description: getClusterTransitionedObjectsMD(), Value: float64(st.NumObjects), VariableLabels: map[string]string{"tier": tier}, }) metrics = append(metrics, MetricV2{
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Sun Mar 10 09:15:15 UTC 2024 - 6.4K bytes - Viewed (0) -
cmd/speedtest.go
} sendResult() break } // We break if we did not see 2.5% growth rate in total GET // requests, we have reached our peak at this point. doBreak := float64(totalGet-throughputHighestGet)/float64(totalGet) < 0.025 throughputHighestGet = totalGet throughputHighestResults = results throughputHighestPut = totalPut if doBreak { sendResult() break }
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Tue May 27 15:19:03 UTC 2025 - 9.2K bytes - Viewed (0) -
cmd/tier_test.go
expFailure := 5 for range expSuccess { globalTierMetrics.logSuccess(tier) } for range expFailure { globalTierMetrics.logFailure(tier) } metrics := globalTierMetrics.Report() var succ, fail float64 for _, metric := range metrics { switch metric.Description.Name { case tierRequestsSuccess: succ += metric.Value case tierRequestsFailure: fail += metric.Value } } if int(succ) != expSuccess {
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 1.5K bytes - Viewed (0) -
cmd/erasure.go
// This goroutine owns the cache. var saverWg sync.WaitGroup saverWg.Add(1) go func() { // Add jitter to the update time so multiple sets don't sync up. updateTime := 30*time.Second + time.Duration(float64(10*time.Second)*rand.Float64()) t := time.NewTicker(updateTime) defer t.Stop() defer saverWg.Done() var lastSave time.Time for { select { case <-t.C: if cache.Info.LastUpdate.Equal(lastSave) {
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 16.1K bytes - Viewed (0) -
cmd/tier.go
metrics = append(metrics, MetricV2{ Description: tierRequestsSuccessMD, Value: float64(stat.success), VariableLabels: map[string]string{"tier": tier}, }) metrics = append(metrics, MetricV2{ Description: tierRequestsFailureMD, Value: float64(stat.failure), VariableLabels: map[string]string{"tier": tier}, }) } return metrics }
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Fri Aug 29 02:39:48 UTC 2025 - 15.6K bytes - Viewed (0)