; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc -O2 < %s -mtriple=aarch64-linux-gnu | FileCheck %s --check-prefix=CHECKN
; RUN: llc -O2 < %s -mtriple=aarch64-linux-gnu -mattr=strict-align | FileCheck %s --check-prefix=CHECKS
declare i32 @bcmp(ptr, ptr, i64) nounwind readonly
declare i32 @memcmp(ptr, ptr, i64) nounwind readonly
define i1 @test_b2(ptr %s1, ptr %s2) {
; CHECKN-LABEL: test_b2:
; CHECKN: // %bb.0: // %entry
; CHECKN-NEXT: ldr x8, [x0]
; CHECKN-NEXT: ldr x9, [x1]
; CHECKN-NEXT: ldur x10, [x0, #7]
; CHECKN-NEXT: ldur x11, [x1, #7]
; CHECKN-NEXT: cmp x8, x9
; CHECKN-NEXT: ccmp x10, x11, #0, eq
; CHECKN-NEXT: cset w0, eq
; CHECKN-NEXT: ret
;
; CHECKS-LABEL: test_b2:
; CHECKS: // %bb.0: // %entry
; CHECKS-NEXT: str x30, [sp, #-16]! // 8-byte Folded Spill
; CHECKS-NEXT: .cfi_def_cfa_offset 16
; CHECKS-NEXT: .cfi_offset w30, -16
; CHECKS-NEXT: mov w2, #15 // =0xf
; CHECKS-NEXT: bl bcmp
; CHECKS-NEXT: cmp w0, #0
; CHECKS-NEXT: cset w0, eq
; CHECKS-NEXT: ldr x30, [sp], #16 // 8-byte Folded Reload
; CHECKS-NEXT: ret
entry:
%bcmp = call i32 @bcmp(ptr %s1, ptr %s2, i64 15)
%ret = icmp eq i32 %bcmp, 0
ret i1 %ret
}
; TODO: Four loads should be within the limit, but the heuristic isn't implemented.
define i1 @test_b2_align8(ptr align 8 %s1, ptr align 8 %s2) {
; CHECKN-LABEL: test_b2_align8:
; CHECKN: // %bb.0: // %entry
; CHECKN-NEXT: ldr x8, [x0]
; CHECKN-NEXT: ldr x9, [x1]
; CHECKN-NEXT: ldur x10, [x0, #7]
; CHECKN-NEXT: ldur x11, [x1, #7]
; CHECKN-NEXT: cmp x8, x9
; CHECKN-NEXT: ccmp x10, x11, #0, eq
; CHECKN-NEXT: cset w0, eq
; CHECKN-NEXT: ret
;
; CHECKS-LABEL: test_b2_align8:
; CHECKS: // %bb.0: // %entry
; CHECKS-NEXT: str x30, [sp, #-16]! // 8-byte Folded Spill
; CHECKS-NEXT: .cfi_def_cfa_offset 16
; CHECKS-NEXT: .cfi_offset w30, -16
; CHECKS-NEXT: mov w2, #15 // =0xf
; CHECKS-NEXT: bl bcmp
; CHECKS-NEXT: cmp w0, #0
; CHECKS-NEXT: cset w0, eq
; CHECKS-NEXT: ldr x30, [sp], #16 // 8-byte Folded Reload
; CHECKS-NEXT: ret
entry:
%bcmp = call i32 @bcmp(ptr %s1, ptr %s2, i64 15)
%ret = icmp eq i32 %bcmp, 0
ret i1 %ret
}
define i1 @test_bs(ptr %s1, ptr %s2) optsize {
; CHECKN-LABEL: test_bs:
; CHECKN: // %bb.0: // %entry
; CHECKN-NEXT: ldp x8, x11, [x1]
; CHECKN-NEXT: ldr x12, [x0, #16]
; CHECKN-NEXT: ldp x9, x10, [x0]
; CHECKN-NEXT: ldr x13, [x1, #16]
; CHECKN-NEXT: cmp x9, x8
; CHECKN-NEXT: ldur x8, [x0, #23]
; CHECKN-NEXT: ldur x9, [x1, #23]
; CHECKN-NEXT: ccmp x10, x11, #0, eq
; CHECKN-NEXT: ccmp x12, x13, #0, eq
; CHECKN-NEXT: ccmp x8, x9, #0, eq
; CHECKN-NEXT: cset w0, eq
; CHECKN-NEXT: ret
;
; CHECKS-LABEL: test_bs:
; CHECKS: // %bb.0: // %entry
; CHECKS-NEXT: str x30, [sp, #-16]! // 8-byte Folded Spill
; CHECKS-NEXT: .cfi_def_cfa_offset 16
; CHECKS-NEXT: .cfi_offset w30, -16
; CHECKS-NEXT: mov w2, #31 // =0x1f
; CHECKS-NEXT: bl memcmp
; CHECKS-NEXT: cmp w0, #0
; CHECKS-NEXT: cset w0, eq
; CHECKS-NEXT: ldr x30, [sp], #16 // 8-byte Folded Reload
; CHECKS-NEXT: ret
entry:
%memcmp = call i32 @memcmp(ptr %s1, ptr %s2, i64 31)
%ret = icmp eq i32 %memcmp, 0
ret i1 %ret
}