Search Options

Results per page
Sort
Preferred Languages
Advance

Results 61 - 70 of 147 for preemption (0.19 sec)

  1. manifests/charts/istiod-remote/values.yaml

        # system-node-critical, it is better to configure this in order to make sure your Istio pods
        # will not be killed because of low priority class.
        # Refer to https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/#priorityclass
        # for more detail.
        priorityClassName: ""
        proxy:
          image: proxyv2
          # This controls the 'policy' in the sidecar injector.
          autoInject: enabled
    Registered: Fri Jun 14 15:00:06 UTC 2024
    - Last Modified: Mon Jun 03 01:55:05 UTC 2024
    - 21K bytes
    - Viewed (0)
  2. src/cmd/compile/internal/ssa/op.go

    	zeroWidth         bool      // op never translates into any machine code. example: copy, which may sometimes translate to machine code, is not zero-width.
    	unsafePoint       bool      // this op is an unsafe point, i.e. not safe for async preemption
    	symEffect         SymEffect // effect this op has on symbol in aux
    	scale             uint8     // amd64/386 indexed load scale
    }
    
    type inputInfo struct {
    	idx  int     // index in Args array
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Apr 22 15:29:10 UTC 2024
    - 18.7K bytes
    - Viewed (0)
  3. manifests/charts/istio-control/istio-discovery/values.yaml

        # system-node-critical, it is better to configure this in order to make sure your Istio pods
        # will not be killed because of low priority class.
        # Refer to https://kubernetes.io/docs/concepts/configuration/pod-priority-preemption/#priorityclass
        # for more detail.
        priorityClassName: ""
    
        proxy:
          image: proxyv2
    
          # This controls the 'policy' in the sidecar injector.
          autoInject: enabled
    
    Registered: Fri Jun 14 15:00:06 UTC 2024
    - Last Modified: Mon Jun 03 01:55:05 UTC 2024
    - 21.1K bytes
    - Viewed (0)
  4. src/runtime/gc_test.go

    	}
    }
    
    func TestGCTestMoveStackOnNextCall(t *testing.T) {
    	t.Parallel()
    	var onStack int
    	// GCTestMoveStackOnNextCall can fail in rare cases if there's
    	// a preemption. This won't happen many times in quick
    	// succession, so just retry a few times.
    	for retry := 0; retry < 5; retry++ {
    		runtime.GCTestMoveStackOnNextCall()
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed Jun 05 22:33:52 UTC 2024
    - 17.6K bytes
    - Viewed (0)
  5. src/runtime/stubs.go

    //
    // frameSize is the total size of the argument frame at stackArgs and must
    // therefore be >= stackArgsSize. It must include additional space for spilling
    // register arguments for stack growth and preemption.
    //
    // TODO(mknyszek): Once we don't need the additional spill space, remove frameSize,
    // since frameSize will be redundant with stackArgsSize.
    //
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 17:58:53 UTC 2024
    - 20.2K bytes
    - Viewed (0)
  6. src/runtime/arena.go

    		// Redirect allocations that don't fit into a chunk well directly
    		// from the heap.
    		if cap >= 0 {
    			return newarray(typ, cap)
    		}
    		return newobject(typ)
    	}
    
    	// Prevent preemption as we set up the space for a new object.
    	//
    	// Act like we're allocating.
    	mp := acquirem()
    	if mp.mallocing != 0 {
    		throw("malloc deadlock")
    	}
    	if mp.gsignal == getg() {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 08 17:44:56 UTC 2024
    - 37.9K bytes
    - Viewed (0)
  7. src/cmd/internal/obj/loong64/obj.go

    	}
    	p.To.Type = obj.TYPE_REG
    	p.To.Reg = REG_R20
    
    	// Mark the stack bound check and morestack call async nonpreemptible.
    	// If we get preempted here, when resumed the preemption request is
    	// cleared, but we'll still call morestack, which will double the stack
    	// unnecessarily. See issue #35470.
    	p = c.ctxt.StartUnsafePoint(p, c.newprog)
    
    	var q *obj.Prog
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Tue Nov 21 19:22:18 UTC 2023
    - 19.7K bytes
    - Viewed (0)
  8. src/cmd/internal/obj/arm/obj5.go

    	}
    	p.To.Type = obj.TYPE_REG
    	p.To.Reg = REG_R1
    
    	// Mark the stack bound check and morestack call async nonpreemptible.
    	// If we get preempted here, when resumed the preemption request is
    	// cleared, but we'll still call morestack, which will double the stack
    	// unnecessarily. See issue #35470.
    	p = c.ctxt.StartUnsafePoint(p, c.newprog)
    
    	if framesize <= abi.StackSmall {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Nov 20 17:19:36 UTC 2023
    - 21.4K bytes
    - Viewed (0)
  9. src/cmd/compile/internal/ssa/writebarrier.go

    				mem = bEnd.NewValue3A(pos, OpStore, types.TypeMem, w.Aux, ptr, val, mem)
    			}
    		}
    
    		// The last store becomes the WBend marker. This marker is used by the liveness
    		// pass to determine what parts of the code are preemption-unsafe.
    		// All subsequent memory operations use this memory, so we have to sacrifice the
    		// previous last memory op to become this new value.
    		bEnd.Values = append(bEnd.Values, last)
    		last.Block = bEnd
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri Sep 08 19:09:14 UTC 2023
    - 23.5K bytes
    - Viewed (0)
  10. src/runtime/mheap.go

    	// then release mheap_.lock.
    	allArenas []arenaIdx
    
    	// sweepArenas is a snapshot of allArenas taken at the
    	// beginning of the sweep cycle. This can be read safely by
    	// simply blocking GC (by disabling preemption).
    	sweepArenas []arenaIdx
    
    	// markArenas is a snapshot of allArenas taken at the beginning
    	// of the mark cycle. Because allArenas is append-only, neither
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 22 22:31:00 UTC 2024
    - 78K bytes
    - Viewed (0)
Back to top