Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 5 of 5 for Scavenge (0.18 sec)

  1. src/runtime/mklockrank.go

    const ranks = `
    # Sysmon
    NONE
    < sysmon
    < scavenge, forcegc;
    
    # Defer
    NONE < defer;
    
    # GC
    NONE <
      sweepWaiters,
      assistQueue,
      sweep;
    
    # Test only
    NONE < testR, testW;
    
    NONE < timerSend;
    
    # Scheduler, timers, netpoll
    NONE < allocmW, execW, cpuprof, pollCache, pollDesc, wakeableSleep;
    scavenge, sweep, testR, wakeableSleep, timerSend < hchan;
    assistQueue,
      cpuprof,
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 08 17:47:01 UTC 2024
    - 9.1K bytes
    - Viewed (0)
  2. src/runtime/mpagealloc_32bit.go

    	if test {
    		// Set up the scavenge index via sysAlloc so the test can free it later.
    		scavIndexSize := uintptr(len(scavengeIndexArray)) * unsafe.Sizeof(atomicScavChunkData{})
    		s.chunks = ((*[(1 << heapAddrBits) / pallocChunkBytes]atomicScavChunkData)(sysAlloc(scavIndexSize, sysStat)))[:]
    		mappedReady = scavIndexSize
    	} else {
    		// Set up the scavenge index.
    		s.chunks = scavengeIndexArray[:]
    	}
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu Apr 20 20:08:25 UTC 2023
    - 4.6K bytes
    - Viewed (0)
  3. src/runtime/mpagealloc_64bit.go

    		sysMap(unsafe.Pointer(need.base.addr()), need.size(), p.sysStat)
    		sysUsed(unsafe.Pointer(need.base.addr()), need.size(), need.size())
    		p.summaryMappedReady += need.size()
    	}
    
    	// Update the scavenge index.
    	p.summaryMappedReady += p.scav.index.sysGrow(base, limit, p.sysStat)
    }
    
    // sysGrow increases the index's backing store in response to a heap growth.
    //
    // Returns the amount of memory added to sysStat.
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed Jan 03 11:00:10 UTC 2024
    - 9.3K bytes
    - Viewed (0)
  4. src/runtime/mpagecache.go

    			cache: ^chunk.pages64(chunkPageIndex(addr)),
    			scav:  chunk.scavenged.block64(chunkPageIndex(addr)),
    		}
    	}
    
    	// Set the page bits as allocated and clear the scavenged bits, but
    	// be careful to only set and clear the relevant bits.
    	cpi := chunkPageIndex(c.base)
    	chunk.allocPages64(cpi, c.cache)
    	chunk.scavenged.clearBlock64(cpi, c.cache&c.scav /* free and scavenged */)
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed Apr 19 14:30:00 UTC 2023
    - 5.6K bytes
    - Viewed (0)
  5. src/runtime/debug/garbage_test.go

    	// susceptible to collection.
    	big = nil
    
    	// FreeOSMemory runs a GC cycle before releasing memory,
    	// so it's fine to skip a GC here.
    	//
    	// It's possible the background scavenger runs concurrently
    	// with this function and does most of the work for it.
    	// If that happens, it's OK. What we want is a test that fails
    	// often if FreeOSMemory does not work correctly, and a test
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed Apr 19 14:30:00 UTC 2023
    - 7.2K bytes
    - Viewed (0)
Back to top