; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 5
; RUN: llc -mtriple=aarch64-linux-gnu -o - %s | FileCheck %s
; This file tests eliding stack slots when lowering the FSINCOS ISD node.
define { float, float } @sincos_f32_value_return(float %x) {
; CHECK-LABEL: sincos_f32_value_return:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: str x30, [sp, #-16]! // 8-byte Folded Spill
; CHECK-NEXT: .cfi_def_cfa_offset 16
; CHECK-NEXT: .cfi_offset w30, -16
; CHECK-NEXT: add x0, sp, #12
; CHECK-NEXT: add x1, sp, #8
; CHECK-NEXT: bl sincosf
; CHECK-NEXT: ldp s1, s0, [sp, #8]
; CHECK-NEXT: ldr x30, [sp], #16 // 8-byte Folded Reload
; CHECK-NEXT: ret
entry:
%sin = tail call float @llvm.sin.f32(float %x)
%cos = tail call float @llvm.cos.f32(float %x)
%ret_0 = insertvalue { float, float } poison, float %sin, 0
%ret_1 = insertvalue { float, float } %ret_0, float %cos, 1
ret { float, float } %ret_1
}
define void @sincos_f32_ptr_return(float %x, ptr noalias %out_sin, ptr noalias %out_cos) {
; CHECK-LABEL: sincos_f32_ptr_return:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: str x30, [sp, #-16]! // 8-byte Folded Spill
; CHECK-NEXT: .cfi_def_cfa_offset 16
; CHECK-NEXT: .cfi_offset w30, -16
; CHECK-NEXT: bl sincosf
; CHECK-NEXT: ldr x30, [sp], #16 // 8-byte Folded Reload
; CHECK-NEXT: ret
entry:
%sin = tail call float @llvm.sin.f32(float %x)
%cos = tail call float @llvm.cos.f32(float %x)
store float %sin, ptr %out_sin, align 4
store float %cos, ptr %out_cos, align 4
ret void
}
define float @sincos_f32_mixed_return(float %x, ptr %out_sin) {
; CHECK-LABEL: sincos_f32_mixed_return:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: str x30, [sp, #-16]! // 8-byte Folded Spill
; CHECK-NEXT: .cfi_def_cfa_offset 16
; CHECK-NEXT: .cfi_offset w30, -16
; CHECK-NEXT: add x1, sp, #12
; CHECK-NEXT: bl sincosf
; CHECK-NEXT: ldr s0, [sp, #12]
; CHECK-NEXT: ldr x30, [sp], #16 // 8-byte Folded Reload
; CHECK-NEXT: ret
entry:
%sin = tail call float @llvm.sin.f32(float %x)
%cos = tail call float @llvm.cos.f32(float %x)
store float %sin, ptr %out_sin, align 4
ret float %cos
}
define { double, double } @sincos_f64_value_return(double %x) {
; CHECK-LABEL: sincos_f64_value_return:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: sub sp, sp, #32
; CHECK-NEXT: str x30, [sp, #16] // 8-byte Folded Spill
; CHECK-NEXT: .cfi_def_cfa_offset 32
; CHECK-NEXT: .cfi_offset w30, -16
; CHECK-NEXT: add x0, sp, #24
; CHECK-NEXT: add x1, sp, #8
; CHECK-NEXT: bl sincos
; CHECK-NEXT: ldr d0, [sp, #24]
; CHECK-NEXT: ldr d1, [sp, #8]
; CHECK-NEXT: ldr x30, [sp, #16] // 8-byte Folded Reload
; CHECK-NEXT: add sp, sp, #32
; CHECK-NEXT: ret
entry:
%sin = tail call double @llvm.sin.f64(double %x)
%cos = tail call double @llvm.cos.f64(double %x)
%ret_0 = insertvalue { double, double } poison, double %sin, 0
%ret_1 = insertvalue { double, double } %ret_0, double %cos, 1
ret { double, double } %ret_1
}
define void @sincos_f64_ptr_return(double %x, ptr noalias %out_sin, ptr noalias %out_cos) {
; CHECK-LABEL: sincos_f64_ptr_return:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: str x30, [sp, #-16]! // 8-byte Folded Spill
; CHECK-NEXT: .cfi_def_cfa_offset 16
; CHECK-NEXT: .cfi_offset w30, -16
; CHECK-NEXT: bl sincos
; CHECK-NEXT: ldr x30, [sp], #16 // 8-byte Folded Reload
; CHECK-NEXT: ret
entry:
%sin = tail call double @llvm.sin.f64(double %x)
%cos = tail call double @llvm.cos.f64(double %x)
store double %sin, ptr %out_sin, align 8
store double %cos, ptr %out_cos, align 8
ret void
}
define double @sincos_f64_mixed_return(double %x, ptr %out_sin) {
; CHECK-LABEL: sincos_f64_mixed_return:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: str x30, [sp, #-16]! // 8-byte Folded Spill
; CHECK-NEXT: .cfi_def_cfa_offset 16
; CHECK-NEXT: .cfi_offset w30, -16
; CHECK-NEXT: add x1, sp, #8
; CHECK-NEXT: bl sincos
; CHECK-NEXT: ldr d0, [sp, #8]
; CHECK-NEXT: ldr x30, [sp], #16 // 8-byte Folded Reload
; CHECK-NEXT: ret
entry:
%sin = tail call double @llvm.sin.f64(double %x)
%cos = tail call double @llvm.cos.f64(double %x)
store double %sin, ptr %out_sin, align 8
ret double %cos
}
; Here %out_sin and %out_cos may alias so we can't replace both stores with the
; call to sincosf (as the order of stores in sincosf is not defined).
define void @sincos_may_alias(float %x, ptr %out_sin, ptr %out_cos) {
; CHECK-LABEL: sincos_may_alias:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: sub sp, sp, #32
; CHECK-NEXT: stp x30, x19, [sp, #16] // 16-byte Folded Spill
; CHECK-NEXT: .cfi_def_cfa_offset 32
; CHECK-NEXT: .cfi_offset w19, -8
; CHECK-NEXT: .cfi_offset w30, -16
; CHECK-NEXT: mov x19, x1
; CHECK-NEXT: add x1, sp, #12
; CHECK-NEXT: bl sincosf
; CHECK-NEXT: ldr s0, [sp, #12]
; CHECK-NEXT: str s0, [x19]
; CHECK-NEXT: ldp x30, x19, [sp, #16] // 16-byte Folded Reload
; CHECK-NEXT: add sp, sp, #32
; CHECK-NEXT: ret
entry:
%sin = tail call float @llvm.sin.f32(float %x)
%cos = tail call float @llvm.cos.f32(float %x)
store float %sin, ptr %out_sin, align 4
store float %cos, ptr %out_cos, align 4
ret void
}
; Here %out is used for both sin and cos (with the final value stored being cos).
define float @sincos_multiple_uses(float %x, ptr %out) {
; CHECK-LABEL: sincos_multiple_uses:
; CHECK: // %bb.0:
; CHECK-NEXT: str x30, [sp, #-16]! // 8-byte Folded Spill
; CHECK-NEXT: .cfi_def_cfa_offset 16
; CHECK-NEXT: .cfi_offset w30, -16
; CHECK-NEXT: mov x1, x0
; CHECK-NEXT: add x0, sp, #12
; CHECK-NEXT: bl sincosf
; CHECK-NEXT: ldr s0, [sp, #12]
; CHECK-NEXT: ldr x30, [sp], #16 // 8-byte Folded Reload
; CHECK-NEXT: ret
%sin = call float @llvm.sin.f32(float %x)
store float %sin, ptr %out, align 4
%reload = load float, ptr %out, align 4
%cos = call float @llvm.cos.f32(float %x)
store float %cos, ptr %out, align 4
ret float %reload
}
; Negative test. We can't fold volatile stores into the library call.
define void @sincos_volatile_result_stores(float %x, ptr noalias %out_sin, ptr noalias %out_cos) {
; CHECK-LABEL: sincos_volatile_result_stores:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: str x30, [sp, #-32]! // 8-byte Folded Spill
; CHECK-NEXT: stp x20, x19, [sp, #16] // 16-byte Folded Spill
; CHECK-NEXT: .cfi_def_cfa_offset 32
; CHECK-NEXT: .cfi_offset w19, -8
; CHECK-NEXT: .cfi_offset w20, -16
; CHECK-NEXT: .cfi_offset w30, -32
; CHECK-NEXT: mov x19, x1
; CHECK-NEXT: mov x20, x0
; CHECK-NEXT: add x0, sp, #12
; CHECK-NEXT: add x1, sp, #8
; CHECK-NEXT: bl sincosf
; CHECK-NEXT: ldp s1, s0, [sp, #8]
; CHECK-NEXT: str s0, [x20]
; CHECK-NEXT: str s1, [x19]
; CHECK-NEXT: ldp x20, x19, [sp, #16] // 16-byte Folded Reload
; CHECK-NEXT: ldr x30, [sp], #32 // 8-byte Folded Reload
; CHECK-NEXT: ret
entry:
%sin = tail call float @llvm.sin.f32(float %x)
%cos = tail call float @llvm.cos.f32(float %x)
store volatile float %sin, ptr %out_sin, align 4
store volatile float %cos, ptr %out_cos, align 4
ret void
}
; Negative test. We can't fold atomic stores into the library call.
define void @sincos_atomic_result_stores(float %x, ptr noalias %out_sin, ptr noalias %out_cos) {
; CHECK-LABEL: sincos_atomic_result_stores:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: str x30, [sp, #-32]! // 8-byte Folded Spill
; CHECK-NEXT: stp x20, x19, [sp, #16] // 16-byte Folded Spill
; CHECK-NEXT: .cfi_def_cfa_offset 32
; CHECK-NEXT: .cfi_offset w19, -8
; CHECK-NEXT: .cfi_offset w20, -16
; CHECK-NEXT: .cfi_offset w30, -32
; CHECK-NEXT: mov x19, x1
; CHECK-NEXT: mov x20, x0
; CHECK-NEXT: add x0, sp, #12
; CHECK-NEXT: add x1, sp, #8
; CHECK-NEXT: bl sincosf
; CHECK-NEXT: ldr w8, [sp, #12]
; CHECK-NEXT: str w8, [x20]
; CHECK-NEXT: ldr w8, [sp, #8]
; CHECK-NEXT: str w8, [x19]
; CHECK-NEXT: ldp x20, x19, [sp, #16] // 16-byte Folded Reload
; CHECK-NEXT: ldr x30, [sp], #32 // 8-byte Folded Reload
; CHECK-NEXT: ret
entry:
%sin = tail call float @llvm.sin.f32(float %x)
%cos = tail call float @llvm.cos.f32(float %x)
store atomic float %sin, ptr %out_sin unordered, align 4
store atomic float %cos, ptr %out_cos unordered, align 4
ret void
}
; Negative test. We can't fold misaligned stores into the library call.
define void @sincos_misaligned_result_stores(double %x, ptr noalias %out_sin, ptr noalias %out_cos) {
; CHECK-LABEL: sincos_misaligned_result_stores:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: sub sp, sp, #48
; CHECK-NEXT: str x30, [sp, #16] // 8-byte Folded Spill
; CHECK-NEXT: stp x20, x19, [sp, #32] // 16-byte Folded Spill
; CHECK-NEXT: .cfi_def_cfa_offset 48
; CHECK-NEXT: .cfi_offset w19, -8
; CHECK-NEXT: .cfi_offset w20, -16
; CHECK-NEXT: .cfi_offset w30, -32
; CHECK-NEXT: mov x19, x1
; CHECK-NEXT: mov x20, x0
; CHECK-NEXT: add x0, sp, #24
; CHECK-NEXT: add x1, sp, #8
; CHECK-NEXT: bl sincos
; CHECK-NEXT: ldr d0, [sp, #24]
; CHECK-NEXT: ldr d1, [sp, #8]
; CHECK-NEXT: ldr x30, [sp, #16] // 8-byte Folded Reload
; CHECK-NEXT: str d0, [x20]
; CHECK-NEXT: str d1, [x19]
; CHECK-NEXT: ldp x20, x19, [sp, #32] // 16-byte Folded Reload
; CHECK-NEXT: add sp, sp, #48
; CHECK-NEXT: ret
entry:
%sin = tail call double @llvm.sin.f64(double %x)
%cos = tail call double @llvm.cos.f64(double %x)
store double %sin, ptr %out_sin, align 4
store double %cos, ptr %out_cos, align 4
ret void
}