- Sort Score
- Result 10 results
- Languages All
Results 351 - 360 of 761 for fake (0.03 sec)
-
cmd/metacache.go
metacacheBlockSize = 5000 // metacacheSharePrefix controls whether prefixes on dirty paths are always shared. // This will make `test/a` and `test/b` share listings if they are concurrent. // Enabling this will make cache sharing more likely and cause less IO, // but may cause additional latency to some calls. metacacheSharePrefix = false ) //go:generate msgp -file $GOFILE -unexported
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Fri Jul 12 16:23:16 UTC 2024 - 6K bytes - Viewed (0) -
tensorflow/api_template.__init__.py
from tensorflow.python.util.lazy_loader import KerasLazyLoader as _KerasLazyLoader # Make sure code inside the TensorFlow codebase can use tf2.enabled() at import. _os.environ["TF2_BEHAVIOR"] = "1" from tensorflow.python import tf2 as _tf2 _tf2.enable() # API IMPORTS PLACEHOLDER # WRAPPER_PLACEHOLDER # Make sure directory containing top level submodules is in
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Wed Oct 02 22:16:02 UTC 2024 - 6.8K bytes - Viewed (0) -
cmd/data-usage-cache_gen.go
case "ts": var zb0002 uint32 zb0002, err = dc.ReadMapHeader() if err != nil { err = msgp.WrapError(err, "Tiers") return } if z.Tiers == nil { z.Tiers = make(map[string]tierStats, zb0002) } else if len(z.Tiers) > 0 { for key := range z.Tiers { delete(z.Tiers, key) } } for zb0002 > 0 { zb0002-- var za0001 string var za0002 tierStats
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Tue Oct 22 15:30:50 UTC 2024 - 75K bytes - Viewed (0) -
src/bufio/bufio.go
func NewReaderSize(rd io.Reader, size int) *Reader { // Is it already a Reader? b, ok := rd.(*Reader) if ok && len(b.buf) >= size { return b } r := new(Reader) r.reset(make([]byte, max(size, minReadBufferSize)), rd) return r } // NewReader returns a new [Reader] whose buffer has the default size. func NewReader(rd io.Reader) *Reader { return NewReaderSize(rd, defaultBufSize) }
Registered: Tue Nov 05 11:13:11 UTC 2024 - Last Modified: Thu Oct 12 14:39:08 UTC 2023 - 21.8K bytes - Viewed (0) -
cmd/os-instrumented.go
var m madmin.OSMetrics m.CollectedAt = time.Now() m.LifeTimeOps = make(map[string]uint64, osMetricLast) for i := osMetric(0); i < osMetricLast; i++ { if n := atomic.LoadUint64(&o.operations[i]); n > 0 { m.LifeTimeOps[i.String()] = n } } if len(m.LifeTimeOps) == 0 { m.LifeTimeOps = nil } m.LastMinute.Operations = make(map[string]madmin.TimedAction, osMetricLast)
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Thu Feb 15 01:09:38 UTC 2024 - 6.3K bytes - Viewed (0) -
internal/kms/context.go
// heuristic value that seems like a good starting size. b := bytes.NewBuffer(make([]byte, 0, 128)) if len(c) == 1 { for k, v := range c { b.WriteString(`{"`) escapeStringJSON(b, k) b.WriteString(`":"`) escapeStringJSON(b, v) b.WriteString(`"}`) } return b.Bytes(), nil } sortedKeys := make([]string, 0, len(c)) for k := range c { sortedKeys = append(sortedKeys, k) }
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Sun Jan 02 17:15:06 UTC 2022 - 6K bytes - Viewed (0) -
istioctl/pkg/writer/envoy/configdump/cluster.go
func (c *ConfigWriter) PrintClusterDump(filter ClusterFilter, outputFormat string) error { _, clusters, err := c.setupClusterConfigWriter() if err != nil { return err } filteredClusters := make(proto.MessageSlice, 0, len(clusters)) for _, cluster := range clusters { if filter.Verify(cluster) { filteredClusters = append(filteredClusters, cluster) } }
Registered: Wed Nov 06 22:53:10 UTC 2024 - Last Modified: Thu May 11 05:38:17 UTC 2023 - 5.9K bytes - Viewed (0) -
cmd/object-handlers_test.go
{"nothing", []int64{0}, make(map[string]string)}, {"small-0", []int64{11}, make(map[string]string)}, {"small-1", []int64{509}, make(map[string]string)}, {"small-2", []int64{5 * oneMiB}, make(map[string]string)}, // // // cases 4-7: multipart part objects {"mp-0", []int64{5 * oneMiB, 10}, make(map[string]string)}, {"mp-1", []int64{5*oneMiB + 1, 10}, make(map[string]string)},
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Thu Sep 12 20:10:44 UTC 2024 - 163.2K bytes - Viewed (0) -
src/bytes/buffer.go
}() // TODO(http://golang.org/issue/51462): We should rely on the append-make // pattern so that the compiler can call runtime.growslice. For example: // return append(b, make([]byte, n)...) // This avoids unnecessary zero-ing of the first len(b) bytes of the // allocated slice, but this pattern causes b to escape onto the heap. // // Instead use the append-make pattern with a nil slice to ensure that
Registered: Tue Nov 05 11:13:11 UTC 2024 - Last Modified: Tue Oct 29 16:47:05 UTC 2024 - 15.7K bytes - Viewed (0) -
cmd/bootstrap-messages.go
bs.mu.Lock() defer bs.mu.Unlock() if len(bs.info) > bootstrapTraceLimit { return } bs.info = append(bs.info, info) } func (bs *bootstrapTracer) Events() []madmin.TraceInfo { traceInfo := make([]madmin.TraceInfo, 0, bootstrapTraceLimit) bs.mu.RLock() for _, i := range bs.info { traceInfo = append(traceInfo, i) } bs.mu.RUnlock() return traceInfo }
Registered: Sun Nov 03 19:28:11 UTC 2024 - Last Modified: Wed Aug 23 10:07:06 UTC 2023 - 1.7K bytes - Viewed (0)