Search Options

Results per page
Sort
Preferred Languages
Advance

Results 351 - 360 of 761 for fake (0.03 sec)

  1. cmd/metacache.go

    	metacacheBlockSize = 5000
    
    	// metacacheSharePrefix controls whether prefixes on dirty paths are always shared.
    	// This will make `test/a` and `test/b` share listings if they are concurrent.
    	// Enabling this will make cache sharing more likely and cause less IO,
    	// but may cause additional latency to some calls.
    	metacacheSharePrefix = false
    )
    
    //go:generate msgp -file $GOFILE -unexported
    
    Registered: Sun Nov 03 19:28:11 UTC 2024
    - Last Modified: Fri Jul 12 16:23:16 UTC 2024
    - 6K bytes
    - Viewed (0)
  2. tensorflow/api_template.__init__.py

    from tensorflow.python.util.lazy_loader import KerasLazyLoader as _KerasLazyLoader
    
    # Make sure code inside the TensorFlow codebase can use tf2.enabled() at import.
    _os.environ["TF2_BEHAVIOR"] = "1"
    from tensorflow.python import tf2 as _tf2
    _tf2.enable()
    
    # API IMPORTS PLACEHOLDER
    
    # WRAPPER_PLACEHOLDER
    
    # Make sure directory containing top level submodules is in
    Registered: Tue Nov 05 12:39:12 UTC 2024
    - Last Modified: Wed Oct 02 22:16:02 UTC 2024
    - 6.8K bytes
    - Viewed (0)
  3. cmd/data-usage-cache_gen.go

    		case "ts":
    			var zb0002 uint32
    			zb0002, err = dc.ReadMapHeader()
    			if err != nil {
    				err = msgp.WrapError(err, "Tiers")
    				return
    			}
    			if z.Tiers == nil {
    				z.Tiers = make(map[string]tierStats, zb0002)
    			} else if len(z.Tiers) > 0 {
    				for key := range z.Tiers {
    					delete(z.Tiers, key)
    				}
    			}
    			for zb0002 > 0 {
    				zb0002--
    				var za0001 string
    				var za0002 tierStats
    Registered: Sun Nov 03 19:28:11 UTC 2024
    - Last Modified: Tue Oct 22 15:30:50 UTC 2024
    - 75K bytes
    - Viewed (0)
  4. src/bufio/bufio.go

    func NewReaderSize(rd io.Reader, size int) *Reader {
    	// Is it already a Reader?
    	b, ok := rd.(*Reader)
    	if ok && len(b.buf) >= size {
    		return b
    	}
    	r := new(Reader)
    	r.reset(make([]byte, max(size, minReadBufferSize)), rd)
    	return r
    }
    
    // NewReader returns a new [Reader] whose buffer has the default size.
    func NewReader(rd io.Reader) *Reader {
    	return NewReaderSize(rd, defaultBufSize)
    }
    
    Registered: Tue Nov 05 11:13:11 UTC 2024
    - Last Modified: Thu Oct 12 14:39:08 UTC 2023
    - 21.8K bytes
    - Viewed (0)
  5. cmd/os-instrumented.go

    	var m madmin.OSMetrics
    	m.CollectedAt = time.Now()
    	m.LifeTimeOps = make(map[string]uint64, osMetricLast)
    	for i := osMetric(0); i < osMetricLast; i++ {
    		if n := atomic.LoadUint64(&o.operations[i]); n > 0 {
    			m.LifeTimeOps[i.String()] = n
    		}
    	}
    	if len(m.LifeTimeOps) == 0 {
    		m.LifeTimeOps = nil
    	}
    
    	m.LastMinute.Operations = make(map[string]madmin.TimedAction, osMetricLast)
    Registered: Sun Nov 03 19:28:11 UTC 2024
    - Last Modified: Thu Feb 15 01:09:38 UTC 2024
    - 6.3K bytes
    - Viewed (0)
  6. internal/kms/context.go

    	// heuristic value that seems like a good starting size.
    	b := bytes.NewBuffer(make([]byte, 0, 128))
    	if len(c) == 1 {
    		for k, v := range c {
    			b.WriteString(`{"`)
    			escapeStringJSON(b, k)
    			b.WriteString(`":"`)
    			escapeStringJSON(b, v)
    			b.WriteString(`"}`)
    		}
    		return b.Bytes(), nil
    	}
    
    	sortedKeys := make([]string, 0, len(c))
    	for k := range c {
    		sortedKeys = append(sortedKeys, k)
    	}
    Registered: Sun Nov 03 19:28:11 UTC 2024
    - Last Modified: Sun Jan 02 17:15:06 UTC 2022
    - 6K bytes
    - Viewed (0)
  7. istioctl/pkg/writer/envoy/configdump/cluster.go

    func (c *ConfigWriter) PrintClusterDump(filter ClusterFilter, outputFormat string) error {
    	_, clusters, err := c.setupClusterConfigWriter()
    	if err != nil {
    		return err
    	}
    	filteredClusters := make(proto.MessageSlice, 0, len(clusters))
    	for _, cluster := range clusters {
    		if filter.Verify(cluster) {
    			filteredClusters = append(filteredClusters, cluster)
    		}
    	}
    Registered: Wed Nov 06 22:53:10 UTC 2024
    - Last Modified: Thu May 11 05:38:17 UTC 2023
    - 5.9K bytes
    - Viewed (0)
  8. cmd/object-handlers_test.go

    		{"nothing", []int64{0}, make(map[string]string)},
    		{"small-0", []int64{11}, make(map[string]string)},
    		{"small-1", []int64{509}, make(map[string]string)},
    		{"small-2", []int64{5 * oneMiB}, make(map[string]string)},
    		// // // cases 4-7: multipart part objects
    		{"mp-0", []int64{5 * oneMiB, 10}, make(map[string]string)},
    		{"mp-1", []int64{5*oneMiB + 1, 10}, make(map[string]string)},
    Registered: Sun Nov 03 19:28:11 UTC 2024
    - Last Modified: Thu Sep 12 20:10:44 UTC 2024
    - 163.2K bytes
    - Viewed (0)
  9. src/bytes/buffer.go

    	}()
    	// TODO(http://golang.org/issue/51462): We should rely on the append-make
    	// pattern so that the compiler can call runtime.growslice. For example:
    	//	return append(b, make([]byte, n)...)
    	// This avoids unnecessary zero-ing of the first len(b) bytes of the
    	// allocated slice, but this pattern causes b to escape onto the heap.
    	//
    	// Instead use the append-make pattern with a nil slice to ensure that
    Registered: Tue Nov 05 11:13:11 UTC 2024
    - Last Modified: Tue Oct 29 16:47:05 UTC 2024
    - 15.7K bytes
    - Viewed (0)
  10. cmd/bootstrap-messages.go

    	bs.mu.Lock()
    	defer bs.mu.Unlock()
    
    	if len(bs.info) > bootstrapTraceLimit {
    		return
    	}
    	bs.info = append(bs.info, info)
    }
    
    func (bs *bootstrapTracer) Events() []madmin.TraceInfo {
    	traceInfo := make([]madmin.TraceInfo, 0, bootstrapTraceLimit)
    
    	bs.mu.RLock()
    	for _, i := range bs.info {
    		traceInfo = append(traceInfo, i)
    	}
    	bs.mu.RUnlock()
    
    	return traceInfo
    }
    
    Registered: Sun Nov 03 19:28:11 UTC 2024
    - Last Modified: Wed Aug 23 10:07:06 UTC 2023
    - 1.7K bytes
    - Viewed (0)
Back to top