Search Options

Results per page
Sort
Preferred Languages
Advance

Results 71 - 80 of 108 for race (1.26 sec)

  1. src/syscall/fs_wasip1.go

    		// determine ahead of time if the path we are about to open is a
    		// directory, so instead we fallback to a second call to path_open with
    		// a more limited set of rights.
    		//
    		// This approach is subject to a race if the file system is modified
    		// concurrently, so we also inject OFLAG_DIRECTORY to ensure that we do
    		// not accidentally open a file which is not a directory.
    		errno = path_open(
    			dirFd,
    			LOOKUP_SYMLINK_FOLLOW,
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 15 19:57:43 UTC 2024
    - 24.1K bytes
    - Viewed (0)
  2. pilot/pkg/xds/delta.go

    		peerAddr = peerInfo.Addr.String()
    	}
    
    	if err := s.WaitForRequestLimit(stream.Context()); err != nil {
    		deltaLog.Warnf("ADS: %q exceeded rate limit: %v", peerAddr, err)
    		return status.Errorf(codes.ResourceExhausted, "request rate limit exceeded: %v", err)
    	}
    
    	ids, err := s.authenticate(ctx)
    	if err != nil {
    		return status.Error(codes.Unauthenticated, err.Error())
    	}
    	if ids != nil {
    Registered: Fri Jun 14 15:00:06 UTC 2024
    - Last Modified: Fri Jun 14 04:34:37 UTC 2024
    - 25.6K bytes
    - Viewed (0)
  3. src/cmd/compile/internal/types2/named.go

    func (n *Named) expandUnderlying() Type {
    	check := n.check
    	if check != nil && check.conf.Trace {
    		check.trace(n.obj.pos, "-- Named.expandUnderlying %s", n)
    		check.indent++
    		defer func() {
    			check.indent--
    			check.trace(n.obj.pos, "=> %s (tparams = %s, under = %s)", n, n.tparams.list(), n.underlying)
    		}()
    	}
    
    	assert(n.inst.orig.underlying != nil)
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 15 20:03:31 UTC 2024
    - 23.9K bytes
    - Viewed (0)
  4. tensorflow/compiler/jit/xla_launch_util.cc

          // separate stream from the main compute stream. We must ensure the
          // compute stream is synchronized with the host->device transfer
          // stream now otherwise we will create a race condition.
          auto* gpu_device_context =
              static_cast<GPUDeviceContext*>(ctx->op_device_context());
          TF_RETURN_IF_ERROR(gpu_device_context->stream()->WaitFor(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 16 00:36:08 UTC 2024
    - 40.4K bytes
    - Viewed (0)
  5. src/encoding/gob/type.go

    	if ui, ok := userTypeCache.Load(rt); ok {
    		return ui.(*userTypeInfo), nil
    	}
    
    	// Construct a new userTypeInfo and atomically add it to the userTypeCache.
    	// If we lose the race, we'll waste a little CPU and create a little garbage
    	// but return the existing value anyway.
    
    	ut := new(userTypeInfo)
    	ut.base = rt
    	ut.user = rt
    	// A type that is just a cycle of pointers (such as type T *T) cannot
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 16 02:00:26 UTC 2024
    - 27.2K bytes
    - Viewed (0)
  6. cmd/kubeadm/app/phases/upgrade/staticpods.go

    	fmt.Printf("[upgrade/staticpods] This can take up to %v\n", kubeadmapi.GetActiveTimeouts().UpgradeManifests.Duration)
    
    	// Wait for the mirror Pod hash to change; otherwise we'll run into race conditions here when the kubelet hasn't had time to
    	// notice the removal of the Static Pod, leading to a false positive below where we check that the API endpoint is healthy
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Wed May 29 10:07:41 UTC 2024
    - 30.7K bytes
    - Viewed (0)
  7. src/go/types/named.go

    func (n *Named) expandUnderlying() Type {
    	check := n.check
    	if check != nil && check.conf._Trace {
    		check.trace(n.obj.pos, "-- Named.expandUnderlying %s", n)
    		check.indent++
    		defer func() {
    			check.indent--
    			check.trace(n.obj.pos, "=> %s (tparams = %s, under = %s)", n, n.tparams.list(), n.underlying)
    		}()
    	}
    
    	assert(n.inst.orig.underlying != nil)
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 15 20:03:31 UTC 2024
    - 24K bytes
    - Viewed (0)
  8. src/go/types/resolver.go

    						// do not import them!
    						if token.IsExported(name) {
    							// declare dot-imported object
    							// (Do not use check.declare because it modifies the object
    							// via Object.setScopePos, which leads to a race condition;
    							// the object may be imported into more than one file scope
    							// concurrently. See go.dev/issue/32154.)
    							if alt := fileScope.Lookup(name); alt != nil {
    								err := check.newError(DuplicateDecl)
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri May 17 16:22:59 UTC 2024
    - 26.1K bytes
    - Viewed (0)
  9. src/go/types/issues_test.go

    // underlying type of b.T (which is the underlying type of a.T, which is int)
    // we must not set the underlying type of a.T again since that would lead to
    // a race condition if package b is imported elsewhere, in a package that is
    // concurrently type-checked.
    func TestIssue34921(t *testing.T) {
    	defer func() {
    		if r := recover(); r != nil {
    			t.Error(r)
    		}
    	}()
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 30 19:19:55 UTC 2024
    - 28.1K bytes
    - Viewed (0)
  10. pkg/kubelet/kubelet_node_status.go

    	unschedulableTaint := v1.Taint{
    		Key:    v1.TaintNodeUnschedulable,
    		Effect: v1.TaintEffectNoSchedule,
    	}
    
    	// Taint node with TaintNodeUnschedulable when initializing
    	// node to avoid race condition; refer to #63897 for more detail.
    	if node.Spec.Unschedulable &&
    		!taintutil.TaintExists(nodeTaints, &unschedulableTaint) {
    		nodeTaints = append(nodeTaints, unschedulableTaint)
    	}
    
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Tue Jun 04 06:25:43 UTC 2024
    - 31.1K bytes
    - Viewed (0)
Back to top