Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 21 for offset2 (0.39 sec)

  1. src/cmd/internal/obj/arm64/asm7.go

    /* checkoffset checks whether the immediate offset is valid for VLD[1-4].P and VST[1-4].P */
    func (c *ctxt7) checkoffset(p *obj.Prog, as obj.As) {
    	var offset, list, n, expect int64
    	switch as {
    	case AVLD1, AVLD2, AVLD3, AVLD4, AVLD1R, AVLD2R, AVLD3R, AVLD4R:
    		offset = p.From.Offset
    		list = p.To.Offset
    	case AVST1, AVST2, AVST3, AVST4:
    		offset = p.To.Offset
    		list = p.From.Offset
    	default:
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 15 15:44:14 UTC 2024
    - 201.1K bytes
    - Viewed (0)
  2. src/cmd/internal/obj/ppc64/asm9.go

    			rel.Type = objabi.R_ADDRPOWER_PCREL34
    		} else if c.opform(inst) == DS_FORM && v&0x3 != 0 {
    			log.Fatalf("invalid offset for DS form load/store %v", p)
    		}
    
    	case 75: // 32 bit offset symbol loads (got/toc/addr)
    		var rel *obj.Reloc
    		v := p.From.Offset
    
    		// Offsets in DS form loads must be a multiple of 4
    		inst := c.opload(p.As)
    		switch p.From.Name {
    		case obj.NAME_GOTREF, obj.NAME_TOCREF:
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 15 13:55:28 UTC 2024
    - 156.1K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/flatbuffer_export.cc

      // Append constant and custom op buffers at the end of the flatbuffer and
      // calculate the offsets
      void AppendBufferData(absl::Cord& result);
    
      // Update constant & custom op buffer offsets
      // Return false if fail to update offset
      bool UpdateBufferOffsets(tflite::Model* mutable_model);
    
      // check if Flatbuffer builder can no longer hold the given amount of the data
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 21:41:49 UTC 2024
    - 164.5K bytes
    - Viewed (0)
  4. src/reflect/value.go

    				switch st.kind {
    				case abiStepIntReg:
    					offset := add(s, st.offset, "precomputed value offset")
    					intFromReg(&regArgs, st.ireg, st.size, offset)
    				case abiStepPointer:
    					s := add(s, st.offset, "precomputed value offset")
    					*((*unsafe.Pointer)(s)) = regArgs.Ptrs[st.ireg]
    				case abiStepFloatReg:
    					offset := add(s, st.offset, "precomputed value offset")
    					floatFromReg(&regArgs, st.freg, st.size, offset)
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 22 21:17:41 UTC 2024
    - 119.9K bytes
    - Viewed (0)
  5. src/cmd/link/internal/ld/data.go

    				}
    			}
    
    			o = ldr.SymValue(rs) + r.Add()
    			if rt == objabi.R_PEIMAGEOFF {
    				// The R_PEIMAGEOFF offset is a RVA, so subtract
    				// the base address for the executable.
    				o -= PEBASE
    			}
    
    			// On amd64, 4-byte offsets will be sign-extended, so it is impossible to
    			// access more than 2GB of static data; fail at link time is better than
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed Jun 12 15:10:50 UTC 2024
    - 100.5K bytes
    - Viewed (1)
  6. src/cmd/internal/obj/x86/asm6.go

    // starting at specified offset (e.g. z counter value).
    // Trailing 0 is not written.
    //
    // Intended to be used for literal Z cases.
    // Literal Z cases usually have "Zlit" in their name (Zlit, Zlitr_m, Zlitm_r).
    func (ab *AsmBuf) PutOpBytesLit(offset int, op *opBytes) {
    	for int(op[offset]) != 0 {
    		ab.Put1(byte(op[offset]))
    		offset++
    	}
    }
    
    // Insert inserts b at offset i.
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 15 15:44:14 UTC 2024
    - 146.9K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/lite/tests/legalize-tf.mlir

      func.return %0 : tensor<1x2x2x5xf32>
      // CHECK-LABEL: strided_slice
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 05 01:54:33 UTC 2024
    - 153.4K bytes
    - Viewed (0)
  8. src/cmd/vendor/github.com/ianlancetaylor/demangle/ast.go

    	pmc = &PtrMemCast{
    		Type:   typ,
    		Expr:   expr,
    		Offset: pmc.Offset,
    	}
    	if r := fn(pmc); r != nil {
    		return r
    	}
    	return pmc
    }
    
    func (pmc *PtrMemCast) GoString() string {
    	return pmc.goString(0, "")
    }
    
    func (pmc *PtrMemCast) goString(indent int, field string) string {
    	return fmt.Sprintf("%*s%sPtrMemCast:\n%s\n%s\n%*sOffset: %d",
    		indent, "", field,
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri May 31 19:48:28 UTC 2024
    - 105.8K bytes
    - Viewed (0)
  9. src/database/sql/sql_test.go

    		badReset bool
    	}{
    		{false, false},
    		{true, false},
    		{false, true},
    	}
    
    	t0 := time.Unix(1000000, 0)
    	offset := time.Duration(0)
    	offsetMu := sync.RWMutex{}
    
    	nowFunc = func() time.Time {
    		offsetMu.RLock()
    		defer offsetMu.RUnlock()
    		return t0.Add(offset)
    	}
    	defer func() { nowFunc = time.Now }()
    
    	ctx, cancel := context.WithCancel(context.Background())
    	defer cancel()
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 23 18:42:28 UTC 2024
    - 111.6K bytes
    - Viewed (0)
  10. src/reflect/all_test.go

    		})
    	}
    }
    
    func check2ndField(x any, offs uintptr, t *testing.T) {
    	s := ValueOf(x)
    	f := s.Type().Field(1)
    	if f.Offset != offs {
    		t.Error("mismatched offsets in structure alignment:", f.Offset, offs)
    	}
    }
    
    // Check that structure alignment & offsets viewed through reflect agree with those
    // from the compiler itself.
    func TestAlignment(t *testing.T) {
    	type T1inner struct {
    		a int
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 23 01:00:11 UTC 2024
    - 218.8K bytes
    - Viewed (0)
Back to top