Search Options

Results per page
Sort
Preferred Languages
Advance

Results 31 - 40 of 465 for Atack (0.04 sec)

  1. src/runtime/os_netbsd.go

    	// signal stack of the creating thread. We always create a
    	// new signal stack here, to avoid having two Go threads using
    	// the same signal stack. This breaks the case of a thread
    	// created in C that calls sigaltstack and then calls a Go
    	// function, because we will lose track of the C code's
    	// sigaltstack, but it's the best we can do.
    	signalstack(&gp.m.gsignal.stack)
    	gp.m.newSigstack = true
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Mar 25 19:53:03 UTC 2024
    - 10.1K bytes
    - Viewed (0)
  2. src/cmd/compile/internal/ssagen/pgen.go

    			fmt.Fprintf(os.Stderr, "%s: %d bytes of stack space saved via stack slot merging (%d nonpointer %d pointer)\n", ir.FuncName(fn), savedNP+savedP, savedNP, savedP)
    			if base.Debug.MergeLocalsTrace > 1 {
    				fmt.Fprintf(os.Stderr, "=-= merge locals state for %v:\n%v",
    					fn, mls)
    			}
    		}
    		leaders = make(map[*ir.Name]int64)
    	}
    
    	// Use sort.SliceStable instead of sort.Slice so stack layout (and thus
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 15 15:44:14 UTC 2024
    - 13.1K bytes
    - Viewed (0)
  3. src/runtime/extern.go

    	execution tracer to double-check stack ownership before taking a stack trace.
    
    	asyncpreemptoff: asyncpreemptoff=1 disables signal-based
    	asynchronous goroutine preemption. This makes some loops
    	non-preemptible for long periods, which may delay GC and
    	goroutine scheduling. This is useful for debugging GC issues
    	because it also disables the conservative stack scanning used
    	for asynchronously preempted goroutines.
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 30 17:52:17 UTC 2024
    - 18.9K bytes
    - Viewed (0)
  4. pkg/test/framework/components/echo/config.go

    	// the same pod.
    	IncludeExtAuthz bool
    
    	// IPFamily for the service. This is optional field. Mainly is used for dual stack testing
    	IPFamilies string
    
    	// IPFamilyPolicy. This is optional field. Mainly is used for dual stack testing.
    	IPFamilyPolicy string
    
    	DualStack bool
    
    Registered: Fri Jun 14 15:00:06 UTC 2024
    - Last Modified: Thu May 09 12:26:52 UTC 2024
    - 18.2K bytes
    - Viewed (0)
  5. cmd/object-lambda-handlers.go

    // Copyright (c) 2015-2023 MinIO, Inc.
    //
    // This file is part of MinIO Object Storage stack
    //
    // This program is free software: you can redistribute it and/or modify
    // it under the terms of the GNU Affero General Public License as published by
    // the Free Software Foundation, either version 3 of the License, or
    // (at your option) any later version.
    //
    // This program is distributed in the hope that it will be useful
    Registered: Sun Jun 16 00:44:34 UTC 2024
    - Last Modified: Fri May 24 23:05:23 UTC 2024
    - 10.2K bytes
    - Viewed (0)
  6. src/cmd/vendor/golang.org/x/tools/go/analysis/passes/asmdecl/asmdecl.go

    	asmArchMipsLE   = asmArch{name: "mipsle", bigEndian: false, stack: "R29", lr: true}
    	asmArchMips64   = asmArch{name: "mips64", bigEndian: true, stack: "R29", lr: true}
    	asmArchMips64LE = asmArch{name: "mips64le", bigEndian: false, stack: "R29", lr: true}
    	asmArchPpc64    = asmArch{name: "ppc64", bigEndian: true, stack: "R1", lr: true, retRegs: []string{"R3", "F1"}}
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri May 03 02:38:00 UTC 2024
    - 22.8K bytes
    - Viewed (0)
  7. src/runtime/gc_test.go

    	// further move the stack.
    	new2 := uintptr(unsafe.Pointer(new))
    
    	t.Logf("old stack pointer %x, new stack pointer %x", old, new2)
    	if new2 == old {
    		// Check that we didn't screw up the test's escape analysis.
    		if cls := runtime.GCTestPointerClass(unsafe.Pointer(new)); cls != "stack" {
    			t.Fatalf("test bug: new (%#x) should be a stack pointer, not %s", new2, cls)
    		}
    		// This was a real failure.
    		return false
    	}
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed Jun 05 22:33:52 UTC 2024
    - 17.6K bytes
    - Viewed (0)
  8. src/cmd/compile/internal/walk/order.go

    			}
    		}
    	}
    	return replaced
    }
    
    type ordermarker int
    
    // markTemp returns the top of the temporary variable stack.
    func (o *orderState) markTemp() ordermarker {
    	return ordermarker(len(o.temp))
    }
    
    // popTemp pops temporaries off the stack until reaching the mark,
    // which must have been returned by markTemp.
    func (o *orderState) popTemp(mark ordermarker) {
    	for _, n := range o.temp[mark:] {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri Mar 08 02:00:33 UTC 2024
    - 42.7K bytes
    - Viewed (0)
  9. src/runtime/chan.go

    	// There are unlocked sudogs that point into gp's stack. Stack
    	// copying must lock the channels of those sudogs.
    	// Set activeStackChans here instead of before we try parking
    	// because we could self-deadlock in stack growth on the
    	// channel lock.
    	gp.activeStackChans = true
    	// Mark that it's safe for stack shrinking to occur now,
    	// because any thread acquiring this G's stack for shrinking
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 23 01:16:50 UTC 2024
    - 25.9K bytes
    - Viewed (0)
  10. src/cmd/compile/internal/ssa/op.go

    // package (assembler).
    func ObjRegForAbiReg(r abi.RegIndex, c *Config) int16 {
    	m := archRegForAbiReg(r, c)
    	return c.registers[m].objNum
    }
    
    // ArgWidth returns the amount of stack needed for all the inputs
    // and outputs of a function or method, including ABI-defined parameter
    // slots and ABI-defined spill slots for register-resident parameters.
    //
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Apr 22 15:29:10 UTC 2024
    - 18.7K bytes
    - Viewed (0)
Back to top