Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 28 for Alignment (0.2 sec)

  1. src/cmd/link/internal/ld/xcoff.go

    	Osntoc      int16    // Section Number For Toc
    	Osnloader   int16    // Section Number For Loader
    	Osnbss      int16    // Section Number For Bss
    	Oalgntext   int16    // Max Text Alignment
    	Oalgndata   int16    // Max Data Alignment
    	Omodtype    [2]byte  // Module Type Field
    	Ocpuflag    uint8    // Bit Flags - Cputypes Of Objects
    	Ocputype    uint8    // Reserved for CPU type
    	Otextpsize  uint8    // Requested text page size
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri Dec 01 19:58:23 UTC 2023
    - 51.8K bytes
    - Viewed (0)
  2. platforms/software/dependency-management/src/integTest/groovy/org/gradle/integtests/resolve/alignment/AlignmentIntegrationTest.groovy

     * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
     * See the License for the specific language governing permissions and
     * limitations under the License.
     */
    
    package org.gradle.integtests.resolve.alignment
    
    import org.gradle.integtests.fixtures.GradleMetadataResolveRunner
    import org.gradle.integtests.fixtures.RequiredFeature
    import spock.lang.Issue
    
    class AlignmentIntegrationTest extends AbstractAlignmentSpec {
    
    Registered: Wed Jun 12 18:38:38 UTC 2024
    - Last Modified: Tue Oct 10 21:10:11 UTC 2023
    - 59.6K bytes
    - Viewed (0)
  3. src/cmd/compile/internal/ssa/_gen/ARM.rules

    (Zero [s] {t} ptr mem)
    	&& s%4 == 0 && s > 4 && s <= 512
    	&& t.Alignment()%4 == 0 && !config.noDuffDevice =>
    	(DUFFZERO [4 * (128 - s/4)] ptr (MOVWconst [0]) mem)
    
    // Large zeroing uses a loop
    (Zero [s] {t} ptr mem)
    	&& (s > 512 || config.noDuffDevice) || t.Alignment()%4 != 0 =>
    	(LoweredZero [t.Alignment()]
    		ptr
    		(ADDconst <ptr.Type> ptr [int32(s-moveSize(t.Alignment(), config))])
    		(MOVWconst [0])
    		mem)
    
    // moves
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Nov 20 17:19:36 UTC 2023
    - 90.1K bytes
    - Viewed (0)
  4. src/cmd/compile/internal/ppc64/ssa.go

    		// eliminates some offset and alignment checking in the rules file.
    
    		fromAddr := obj.Addr{Type: obj.TYPE_MEM, Reg: v.Args[0].Reg()}
    		ssagen.AddAux(&fromAddr, v)
    
    		genAddr := false
    
    		switch fromAddr.Name {
    		case obj.NAME_EXTERN, obj.NAME_STATIC:
    			// Special case for a rule combines the bytes of gostring.
    			// The v alignment might seem OK, but we don't want to load it
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 22 19:59:38 UTC 2024
    - 55.4K bytes
    - Viewed (0)
  5. src/cmd/compile/internal/ssa/_gen/S390X.rules

    	(MOVDload  [off1+off2] {mergeSym(sym1,sym2)} base mem)
    (MOVWZload  [off1] {sym1} (MOVDaddr <t> [off2] {sym2} base) mem) && is32Bit(int64(off1)+int64(off2)) && canMergeSym(sym1, sym2) && (base.Op != OpSB || (t.IsPtr() && t.Elem().Alignment()%4 == 0 && (off1+off2)%4 == 0)) =>
    	(MOVWZload  [off1+off2] {mergeSym(sym1,sym2)} base mem)
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu Oct 12 18:09:26 UTC 2023
    - 74.3K bytes
    - Viewed (0)
  6. src/crypto/internal/nistec/p256_asm_ppc64le.s

    // 2. ppc64 does not have a multiply high and low
    // like s390x, so those were implemented using
    // macros to compute the equivalent values.
    
    // 3. The LVX, STVX instructions on ppc64 require
    // 16 byte alignment of the data.  To avoid that
    // requirement, data is loaded using LXVD2X and
    // STXVD2X with VPERM to reorder bytes correctly.
    
    // I have identified some areas where I believe
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Mar 04 17:29:44 UTC 2024
    - 56.5K bytes
    - Viewed (0)
  7. src/cmd/link/internal/loader/loader.go

    	}
    }
    
    // SymAlign returns the alignment for a symbol.
    func (l *Loader) SymAlign(i Sym) int32 {
    	if int(i) >= len(l.align) {
    		// align is extended lazily -- it the sym in question is
    		// outside the range of the existing slice, then we assume its
    		// alignment has not yet been set.
    		return 0
    	}
    	// TODO: would it make sense to return an arch-specific
    	// alignment depending on section type? E.g. STEXT => 32,
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri May 24 20:26:10 UTC 2024
    - 81.5K bytes
    - Viewed (0)
  8. src/runtime/mgc.go

    //go:linkname writeBarrier
    var writeBarrier struct {
    	enabled bool    // compiler emits a check of this before calling write barrier
    	pad     [3]byte // compiler uses 32-bit load for "enabled" field
    	alignme uint64  // guarantee alignment so that compiler can use a 32 or 64-bit load
    }
    
    // gcBlackenEnabled is 1 if mutator assists and background mark
    // workers are allowed to blacken objects. This must only be set when
    // gcphase == _GCmark.
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 16:25:21 UTC 2024
    - 62K bytes
    - Viewed (0)
  9. src/runtime/malloc.go

    // aligned to align bytes. It may reserve either n or n+align bytes,
    // so it returns the size that was reserved.
    func sysReserveAligned(v unsafe.Pointer, size, align uintptr) (unsafe.Pointer, uintptr) {
    	// Since the alignment is rather large in uses of this
    	// function, we're not likely to get it by chance, so we ask
    	// for a larger region and remove the parts we don't need.
    	retries := 0
    retry:
    	p := uintptr(sysReserve(v, size+align))
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 17:58:53 UTC 2024
    - 59.6K bytes
    - Viewed (0)
  10. src/runtime/map.go

    	// Because of minimum alignment rules, bucketCnt is known to be at least 8.
    	// Represent as loadFactorNum/loadFactorDen, to allow integer math.
    	loadFactorDen = 2
    	loadFactorNum = loadFactorDen * abi.MapBucketCount * 13 / 16
    
    	// data offset should be the size of the bmap struct, but needs to be
    	// aligned correctly. For amd64p32 this means 64-bit alignment
    	// even though pointers are 32 bit.
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 17:58:53 UTC 2024
    - 57.6K bytes
    - Viewed (0)
Back to top