/prebuilts/go/darwin-x86/src/sync/atomic/ |
asm_mipsx.s | 40 JMP runtime?internal?atomic·Xadd(SB) 43 JMP runtime?internal?atomic·Xadd(SB) 46 JMP runtime?internal?atomic·Xadd(SB)
|
/prebuilts/go/linux-x86/src/sync/atomic/ |
asm_mipsx.s | 40 JMP runtime?internal?atomic·Xadd(SB) 43 JMP runtime?internal?atomic·Xadd(SB) 46 JMP runtime?internal?atomic·Xadd(SB)
|
/prebuilts/go/darwin-x86/src/cmd/link/internal/arm64/ |
asm.go | 115 ld.Thearch.Vput(uint64(r.Xadd)) 124 ld.Thearch.Vput(uint64(r.Xadd)) 130 ld.Thearch.Vput(uint64(r.Xadd)) 141 ld.Thearch.Vput(uint64(r.Xadd)) 183 if r.Xadd != 0 { 184 ld.Errorf(s, "ld64 doesn't allow BR26 reloc with non-zero addend: %s+%d", rs.Name, r.Xadd) 193 // if r.Xadd is non-zero, add two MACHO_ARM64_RELOC_ADDEND. 194 if r.Xadd != 0 { 196 ld.Thearch.Lput((ld.MACHO_ARM64_RELOC_ADDEND << 28) | (2 << 25) | uint32(r.Xadd&0xffffff)) 200 if r.Xadd != 0 [all...] |
/prebuilts/go/linux-x86/src/cmd/link/internal/arm64/ |
asm.go | 115 ld.Thearch.Vput(uint64(r.Xadd)) 124 ld.Thearch.Vput(uint64(r.Xadd)) 130 ld.Thearch.Vput(uint64(r.Xadd)) 141 ld.Thearch.Vput(uint64(r.Xadd)) 183 if r.Xadd != 0 { 184 ld.Errorf(s, "ld64 doesn't allow BR26 reloc with non-zero addend: %s+%d", rs.Name, r.Xadd) 193 // if r.Xadd is non-zero, add two MACHO_ARM64_RELOC_ADDEND. 194 if r.Xadd != 0 { 196 ld.Thearch.Lput((ld.MACHO_ARM64_RELOC_ADDEND << 28) | (2 << 25) | uint32(r.Xadd&0xffffff)) 200 if r.Xadd != 0 [all...] |
/prebuilts/go/darwin-x86/src/runtime/internal/atomic/ |
asm_mipsx.s | 43 TEXT ·Xadd(SB),NOSPLIT,$0-12 86 JMP ·Xadd(SB)
|
asm_arm.s | 65 B runtime?internal?atomic·Xadd(SB)
|
atomic_arm.go | 43 func Xadd(val *uint32, delta int32) uint32 { 73 return Xadd(addr, 0) 84 return unsafe.Pointer(uintptr(Xadd((*uint32)(addr), 0)))
|
asm_386.s | 36 JMP runtime?internal?atomic·Xadd(SB) 83 // uint32 Xadd(uint32 volatile *val, int32 delta) 87 TEXT runtime?internal?atomic·Xadd(SB), NOSPLIT, $0-12
|
asm_amd64p32.s | 74 // uint32 Xadd(uint32 volatile *val, int32 delta) 78 TEXT runtime?internal?atomic·Xadd(SB), NOSPLIT, $0-12 99 JMP runtime?internal?atomic·Xadd(SB)
|
atomic_386.go | 47 func Xadd(ptr *uint32, delta int32) uint32
|
atomic_amd64x.go | 30 func Xadd(ptr *uint32, delta int32) uint32
|
atomic_arm64.go | 12 func Xadd(ptr *uint32, delta int32) uint32
|
atomic_mips64x.go | 12 func Xadd(ptr *uint32, delta int32) uint32
|
atomic_ppc64x.go | 12 func Xadd(ptr *uint32, delta int32) uint32
|
atomic_s390x.go | 55 func Xadd(ptr *uint32, delta int32) uint32
|
/prebuilts/go/linux-x86/src/runtime/internal/atomic/ |
asm_mipsx.s | 43 TEXT ·Xadd(SB),NOSPLIT,$0-12 86 JMP ·Xadd(SB)
|
asm_arm.s | 65 B runtime?internal?atomic·Xadd(SB)
|
atomic_arm.go | 43 func Xadd(val *uint32, delta int32) uint32 { 73 return Xadd(addr, 0) 84 return unsafe.Pointer(uintptr(Xadd((*uint32)(addr), 0)))
|
asm_386.s | 36 JMP runtime?internal?atomic·Xadd(SB) 83 // uint32 Xadd(uint32 volatile *val, int32 delta) 87 TEXT runtime?internal?atomic·Xadd(SB), NOSPLIT, $0-12
|
asm_amd64p32.s | 74 // uint32 Xadd(uint32 volatile *val, int32 delta) 78 TEXT runtime?internal?atomic·Xadd(SB), NOSPLIT, $0-12 99 JMP runtime?internal?atomic·Xadd(SB)
|
atomic_386.go | 47 func Xadd(ptr *uint32, delta int32) uint32
|
atomic_amd64x.go | 30 func Xadd(ptr *uint32, delta int32) uint32
|
atomic_arm64.go | 12 func Xadd(ptr *uint32, delta int32) uint32
|
/prebuilts/go/darwin-x86/src/cmd/link/internal/mips/ |
asm.go | 111 r.Xadd = r.Add 113 r.Xadd += ld.Symaddr(rs) - ld.Symaddr(rs.Outer) 121 applyrel(r, s, val, r.Xadd) 127 r.Xadd = r.Add
|
/prebuilts/go/linux-x86/src/cmd/link/internal/mips/ |
asm.go | 111 r.Xadd = r.Add 113 r.Xadd += ld.Symaddr(rs) - ld.Symaddr(rs.Outer) 121 applyrel(r, s, val, r.Xadd) 127 r.Xadd = r.Add
|