REQUIRES: aarch64
## Test handling of addends taken from the relocated word or instruction
## in AArch64 relocation sections of type SHT_REL. These can be generated
## by assemblers other than LLVM, in particular the legacy 'armasm'.
##
## llvm-mc will only generate SHT_RELA when targeting AArch64. So to make
## an input file with SHT_REL, we assemble our test source file, then
## round-trip via YAML and do some seddery to change the type of the
## relocation section. Since all the relocations were made manually with
## .reloc directives containing no addend, this succeeds.
# RUN: rm -rf %t && split-file %s %t && cd %t
# RUN: llvm-mc -filetype=obj -triple=aarch64 relocs.s -o rela.o
# RUN: obj2yaml rela.o -o rela.yaml
# RUN: sed "s/\.rela/\.rel/;s/SHT_RELA/SHT_REL/" rela.yaml > rel.yaml
# RUN: yaml2obj rel.yaml -o rel.o
# RUN: llvm-mc -filetype=obj -triple=aarch64 symbols.s -o symbols.o
# RUN: ld.lld rel.o symbols.o -o a.out --section-start=.data=0x100000 --section-start=.text=0x200000
# RUN: llvm-objdump -s a.out | FileCheck %s --check-prefix=DATALE
# RUN: llvm-objdump -d a.out | FileCheck %s --check-prefix=CODE
# RUN: llvm-mc -filetype=obj -triple=aarch64_be relocs.s -o rela_be.o
# RUN: obj2yaml rela_be.o -o rela_be.yaml
# RUN: sed "s/\.rela/\.rel/;s/SHT_RELA/SHT_REL/" rela_be.yaml > rel_be.yaml
# RUN: yaml2obj rel_be.yaml -o rel_be.o
# RUN: llvm-mc -filetype=obj -triple=aarch64_be symbols.s -o symbols_be.o
# RUN: ld.lld -EB rel_be.o symbols_be.o -o be.out --section-start=.data=0x100000 --section-start=.text=0x200000
# RUN: llvm-objdump -s be.out | FileCheck %s --check-prefix=DATABE
# RUN: llvm-objdump -d be.out | FileCheck %s --check-prefix=CODE
#--- symbols.s
// Source file containing the values of target symbols for the relocations. If
// we don't keep these in their own file, then llvm-mc is clever enough to
// resolve some of the relocations during assembly, even though they're written
// as explicit .reloc directives. But we want the relocations to be present in
// the object file, so that yaml2obj can change their type and we can test
// lld's handling of the result. So we ensure that llvm-mc can't see both the
// .reloc and the target symbol value at the same time.
.globl abs16
.globl abs32
.globl abs64
.globl big64
.globl pcrel
.globl data
.globl branchtarget
.globl calltarget
.equ abs16, 0x9999
.equ data, 0x100000
.equ branchtarget, 0x200100
.equ calltarget, 0x02000100
.equ pcrel, 0x245678
.equ abs32, 0x88888888
.equ abs64, 0x7777777777777777
.equ big64, 0x77ffffffffffff77
#--- relocs.s
// Source file containing the test instructions and their relocations, with the
// FileCheck comments interleaved.
// DATALE: Contents of section .data:
// DATABE: Contents of section .data:
.data
// First test absolute data relocations. For each one I show the expected
// value in a comment, and then expect a line in llvm-objdump -s containing
// all the values together.
// 0x7777777777777777 + 0x1234567887654321 = 0x89abcdeffedcba98
.reloc ., R_AARCH64_ABS64, abs64
.xword 0x1234567887654321
// 0x88888888 + 0x12344321 = 0x9abccba9
.reloc ., R_AARCH64_ABS32, abs32
.word 0x12344321
// 0x9999 + 0x1234 = 0xabcd
.reloc ., R_AARCH64_ABS16, abs16
.hword 0x1234
// DATALE-NEXT: 100000 98badcfe efcdab89 a9cbbc9a cdab
// DATABE-NEXT: 100000 89abcdef fedcba98 9abccba9 abcd
.balign 16
// Test relative data relocs, each subtracting the address of the relocated
// word.
// 0x100000 + 0x1234567887654321 - 0x100010 = 0x1234567887654311
.reloc ., R_AARCH64_PREL64, data
.xword 0x1234567887654321
// 0x100000 + 0x12344321 - 0x100018 = 0x12344309
.reloc ., R_AARCH64_PREL32, data
.word 0x12344321
// 0x100000 + 0x1234 - 0x10001c = 0x1218
.reloc ., R_AARCH64_PREL16, data
.hword 0x1234
// DATALE-NEXT: 100010 11436587 78563412 09433412 1812
// DATABE-NEXT: 100010 12345678 87654311 12344309 1218
// CODE: 0000000000200000 <_start>:
.text
.globl _start
_start:
// Full set of 4 instructions loading the constant 'abs64' and adding 0x1234 to
// it.
// Expected constant is 0x7777777777777777 + 0x1234 = 0x77777777777789ab
.reloc ., R_AARCH64_MOVW_UABS_G0_NC, abs64
movz x0, #0x1234
// CODE-NEXT: 200000: d2913560 mov x0, #0x89ab
.reloc ., R_AARCH64_MOVW_UABS_G1_NC, abs64
movk x0, #0x1234, lsl #16
// CODE-NEXT: 200004: f2aeeee0 movk x0, #0x7777, lsl #16
.reloc ., R_AARCH64_MOVW_UABS_G2_NC, abs64
movk x0, #0x1234, lsl #32
// CODE-NEXT: 200008: f2ceeee0 movk x0, #0x7777, lsl #32
.reloc ., R_AARCH64_MOVW_UABS_G3, abs64
movk x0, #0x1234, lsl #48
// CODE-NEXT: 20000c: f2eeeee0 movk x0, #0x7777, lsl #48
// The same, but this constant has ffff in the middle 32 bits, forcing carries
// to be propagated.
// Expected constant: 0x77ffffffffffff77 + 0x1234 = 0x78000000000011ab
.reloc ., R_AARCH64_MOVW_UABS_G0_NC, big64
movz x0, #0x1234
// CODE-NEXT: 200010: d2823560 mov x0, #0x11ab
.reloc ., R_AARCH64_MOVW_UABS_G1_NC, big64
movk x0, #0x1234, lsl #16
// CODE-NEXT: 200014: f2a00000 movk x0, #0x0, lsl #16
.reloc ., R_AARCH64_MOVW_UABS_G2_NC, big64
movk x0, #0x1234, lsl #32
// CODE-NEXT: 200018: f2c00000 movk x0, #0x0, lsl #32
.reloc ., R_AARCH64_MOVW_UABS_G3, big64
movk x0, #0x1234, lsl #48
// CODE-NEXT: 20001c: f2ef0000 movk x0, #0x7800, lsl #48
// Demonstrate that offsets are treated as signed: this one is taken to be
// -0x1234. (If it were +0xedcc then you'd be able to tell the difference by
// the carry into the second halfword.)
// Expected value: 0x7777777777777777 - 0x1234 = 0x7777777777776543
.reloc ., R_AARCH64_MOVW_UABS_G0_NC, abs64
movz x0, #0xedcc
// CODE-NEXT: 200020: d28ca860 mov x0, #0x6543
.reloc ., R_AARCH64_MOVW_UABS_G1_NC, abs64
movk x0, #0xedcc, lsl #16
// CODE-NEXT: 200024: f2aeeee0 movk x0, #0x7777, lsl #16
.reloc ., R_AARCH64_MOVW_UABS_G2_NC, abs64
movk x0, #0xedcc, lsl #32
// CODE-NEXT: 200028: f2ceeee0 movk x0, #0x7777, lsl #32
.reloc ., R_AARCH64_MOVW_UABS_G3, abs64
movk x0, #0xedcc, lsl #48
// CODE-NEXT: 20002c: f2eeeee0 movk x0, #0x7777, lsl #48
// Check various bits of the ADR immediate, including in particular the low 2
// bits, which are not contiguous with the rest in the encoding.
//
// These values are all 0x245678 + 2^n, except the last one, where the set bit
// of the addend is the top bit, counting as negative, i.e. we expect the value
// 0x254678 - 0x100000 = 0x145678.
.reloc ., R_AARCH64_ADR_PREL_LO21, pcrel
adr x0, .+1
// CODE-NEXT: 200030: 3022b240 adr x0, 0x245679
.reloc ., R_AARCH64_ADR_PREL_LO21, pcrel
adr x0, .+2
// CODE-NEXT: 200034: 5022b220 adr x0, 0x24567a
.reloc ., R_AARCH64_ADR_PREL_LO21, pcrel
adr x0, .+4
// CODE-NEXT: 200038: 1022b220 adr x0, 0x24567c
.reloc ., R_AARCH64_ADR_PREL_LO21, pcrel
adr x0, .+8
// CODE-NEXT: 20003c: 1022b220 adr x0, 0x245680
.reloc ., R_AARCH64_ADR_PREL_LO21, pcrel
adr x0, .+1<<19
// CODE-NEXT: 200040: 1062b1c0 adr x0, 0x2c5678
.reloc ., R_AARCH64_ADR_PREL_LO21, pcrel
adr x0, .-1<<20
// CODE-NEXT: 200044: 10a2b1a0 adr x0, 0x145678
// Now load the same set of values with ADRP+ADD. But because the real ADRP
// instruction shifts its immediate, we must account for that.
.reloc ., R_AARCH64_ADR_PREL_PG_HI21, pcrel
adrp x0, 1<<12
// CODE-NEXT: 200048: b0000220 adrp x0, 0x245000
.reloc ., R_AARCH64_ADD_ABS_LO12_NC, pcrel
add x0, x0, #1
// CODE-NEXT: 20004c: 9119e400 add x0, x0, #0x679
.reloc ., R_AARCH64_ADR_PREL_PG_HI21, pcrel
adrp x0, 2<<12
// CODE-NEXT: 200050: b0000220 adrp x0, 0x245000
.reloc ., R_AARCH64_ADD_ABS_LO12_NC, pcrel
add x0, x0, #2
// CODE-NEXT: 200054: 9119e800 add x0, x0, #0x67a
.reloc ., R_AARCH64_ADR_PREL_PG_HI21, pcrel
adrp x0, 4<<12
// CODE-NEXT: 200058: b0000220 adrp x0, 0x245000
.reloc ., R_AARCH64_ADD_ABS_LO12_NC, pcrel
add x0, x0, #4
// CODE-NEXT: 20005c: 9119f000 add x0, x0, #0x67c
.reloc ., R_AARCH64_ADR_PREL_PG_HI21, pcrel
adrp x0, 8<<12
// CODE-NEXT: 200060: b0000220 adrp x0, 0x245000
.reloc ., R_AARCH64_ADD_ABS_LO12_NC, pcrel
add x0, x0, #8
// CODE-NEXT: 200064: 911a0000 add x0, x0, #0x680
// Starting here, the high bits won't fit in the ADD immediate, so that
// becomes 0, and only the ADRP immediate shows evidence of the addend.
.reloc ., R_AARCH64_ADR_PREL_PG_HI21, pcrel
adrp x0, 1<<(19+12)
// CODE-NEXT: 200068: b0000620 adrp x0, 0x2c5000
.reloc ., R_AARCH64_ADD_ABS_LO12_NC, pcrel
add x0, x0, #0
// CODE-NEXT: 20006c: 9119e000 add x0, x0, #0x678
.reloc ., R_AARCH64_ADR_PREL_PG_HI21, pcrel
adrp x0, -1<<(20+12)
// CODE-NEXT: 200070: b0fffa20 adrp x0, 0x145000
.reloc ., R_AARCH64_ADD_ABS_LO12_NC, pcrel
add x0, x0, #0
// CODE-NEXT: 200074: 9119e000 add x0, x0, #0x678
// Finally, an example with a full 21-bit addend.
// Expected value = 0x245678 + 0xfedcb - 0x100000 = 0x244443
.reloc ., R_AARCH64_ADR_PREL_PG_HI21, pcrel
adrp x0, (0xfedcb-0x100000)<<12
// CODE-NEXT: 200078: 90000220 adrp x0, 0x244000
.reloc ., R_AARCH64_ADD_ABS_LO12_NC, pcrel
add x0, x0, #0xdcb
// CODE-NEXT: 20007c: 91110c00 add x0, x0, #0x443
// PC-relative loads, in which the 19-bit offset is shifted. The offsets are
// the same as the ADRs above, except for the first two, which can't be
// expressed by pc-relative LDR with an offset shifted left 2.
//
// (The input syntax is confusing here. I'd normally expect to write this as
// `ldr x0, [pc, #offset]`, but LLVM writes just `#offset`.)
.reloc ., R_AARCH64_LD_PREL_LO19, pcrel
ldr w0, #4
// CODE-NEXT: 200080: 1822afe0 ldr w0, 0x24567c
.reloc ., R_AARCH64_LD_PREL_LO19, pcrel
ldr w0, #8
// CODE-NEXT: 200084: 1822afe0 ldr w0, 0x245680
.reloc ., R_AARCH64_LD_PREL_LO19, pcrel
ldr w0, #1<<19
// CODE-NEXT: 200088: 1862af80 ldr w0, 0x2c5678
.reloc ., R_AARCH64_LD_PREL_LO19, pcrel
ldr w0, #-1<<20
// CODE-NEXT: 20008c: 18a2af60 ldr w0, 0x145678
// For these, the branch target is 0x200100 plus powers of 2, except the offset
// 2^15, which is negative, because the addend is treated as signed.
.reloc ., R_AARCH64_TSTBR14, branchtarget
tbnz x1, #63, #4
// CODE-NEXT: 200090: b7f803a1 tbnz x1, #0x3f, 0x200104
.reloc ., R_AARCH64_TSTBR14, branchtarget
tbnz x1, #62, #8
// CODE-NEXT: 200094: b7f003a1 tbnz x1, #0x3e, 0x200108
.reloc ., R_AARCH64_TSTBR14, branchtarget
tbnz x1, #61, #1<<14
// CODE-NEXT: 200098: b7ea0341 tbnz x1, #0x3d, 0x204100
.reloc ., R_AARCH64_TSTBR14, branchtarget
tbnz x1, #60, #-1<<15
// CODE-NEXT: 20009c: b7e40321 tbnz x1, #0x3c, 0x1f8100
// CONDBR19 is used for both cbz/cbnz and B.cond, so test both at once. Base
// offset is the same again (from 0x200100), but this time, offsets can go up
// to 2^20.
.reloc ., R_AARCH64_CONDBR19, branchtarget
cbnz x2, #4
// CODE-NEXT: 2000a0: b5000322 cbnz x2, 0x200104
.reloc ., R_AARCH64_CONDBR19, branchtarget
b.eq #8
// CODE-NEXT: 2000a4: 54000320 b.eq 0x200108
.reloc ., R_AARCH64_CONDBR19, branchtarget
cbz x2, #1<<19
// CODE-NEXT: 2000a8: b44002c2 cbz x2, 0x280100
.reloc ., R_AARCH64_CONDBR19, branchtarget
b.vs #-1<<20
// CODE-NEXT: 2000ac: 548002a6 b.vs 0x100100
// And for BL and B, the offsets go up to 2^25.
.reloc ., R_AARCH64_CALL26, calltarget
bl #4
// CODE-NEXT: 2000b0: 94780015 bl 0x2000104
.reloc ., R_AARCH64_CALL26, calltarget
bl #8
// CODE-NEXT: 2000b4: 94780015 bl 0x2000108
.reloc ., R_AARCH64_CALL26, calltarget
bl #1<<24
// CODE-NEXT: 2000b8: 94b80012 bl 0x3000100
.reloc ., R_AARCH64_CALL26, calltarget
bl #-1<<25
// CODE-NEXT: 2000bc: 97f80011 bl 0x100
.reloc ., R_AARCH64_JUMP26, calltarget
b #4
// CODE-NEXT: 2000c0: 14780011 b 0x2000104
.reloc ., R_AARCH64_JUMP26, calltarget
b #8
// CODE-NEXT: 2000c4: 14780011 b 0x2000108
.reloc ., R_AARCH64_JUMP26, calltarget
b #1<<24
// CODE-NEXT: 2000c8: 14b8000e b 0x3000100
.reloc ., R_AARCH64_JUMP26, calltarget
b #-1<<25
// CODE-NEXT: 2000cc: 17f8000d b 0x100