; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 5
; RUN: llc -mtriple=aarch64 -verify-machineinstrs %s -o - | FileCheck %s
define zeroext i16 @and_sext_v8i8_i16(<8 x i8> %x) {
; CHECK-LABEL: and_sext_v8i8_i16:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: and x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: and w8, w8, w9
; CHECK-NEXT: and w8, w8, w8, lsr #8
; CHECK-NEXT: sxtb w8, w8
; CHECK-NEXT: and w0, w8, #0xffff
; CHECK-NEXT: ret
entry:
%y = sext <8 x i8> %x to <8 x i16>
%z = call i16 @llvm.vector.reduce.and.v8i16(<8 x i16> %y)
ret i16 %z
}
define zeroext i16 @and_zext_v8i8_i16(<8 x i8> %x) {
; CHECK-LABEL: and_zext_v8i8_i16:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: and x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: and w8, w8, w9
; CHECK-NEXT: and w8, w8, w8, lsr #8
; CHECK-NEXT: and w0, w8, #0xff
; CHECK-NEXT: ret
entry:
%y = zext <8 x i8> %x to <8 x i16>
%z = call i16 @llvm.vector.reduce.and.v8i16(<8 x i16> %y)
ret i16 %z
}
define zeroext i16 @and_sext_v16i8_i16(<16 x i8> %x) {
; CHECK-LABEL: and_sext_v16i8_i16:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: and v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: and x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: and w8, w8, w9
; CHECK-NEXT: and w8, w8, w8, lsr #8
; CHECK-NEXT: sxtb w8, w8
; CHECK-NEXT: and w0, w8, #0xffff
; CHECK-NEXT: ret
entry:
%y = sext <16 x i8> %x to <16 x i16>
%z = call i16 @llvm.vector.reduce.and.v16i16(<16 x i16> %y)
ret i16 %z
}
define zeroext i16 @and_zext_v16i8_i16(<16 x i8> %x) {
; CHECK-LABEL: and_zext_v16i8_i16:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: and v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: and x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: and w8, w8, w9
; CHECK-NEXT: and w8, w8, w8, lsr #8
; CHECK-NEXT: and w0, w8, #0xff
; CHECK-NEXT: ret
entry:
%y = zext <16 x i8> %x to <16 x i16>
%z = call i16 @llvm.vector.reduce.and.v16i16(<16 x i16> %y)
ret i16 %z
}
define i32 @and_sext_v8i8_i32(<8 x i8> %x) {
; CHECK-LABEL: and_sext_v8i8_i32:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: and x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: and w8, w8, w9
; CHECK-NEXT: and w8, w8, w8, lsr #8
; CHECK-NEXT: sxtb w0, w8
; CHECK-NEXT: ret
entry:
%y = sext <8 x i8> %x to <8 x i32>
%z = call i32 @llvm.vector.reduce.and.v8i32(<8 x i32> %y)
ret i32 %z
}
define i32 @and_zext_v8i8_i32(<8 x i8> %x) {
; CHECK-LABEL: and_zext_v8i8_i32:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: and x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: and w8, w8, w9
; CHECK-NEXT: and w8, w8, w8, lsr #8
; CHECK-NEXT: and w0, w8, #0xff
; CHECK-NEXT: ret
entry:
%y = zext <8 x i8> %x to <8 x i32>
%z = call i32 @llvm.vector.reduce.and.v8i32(<8 x i32> %y)
ret i32 %z
}
define i32 @and_sext_v16i8_i32(<16 x i8> %x) {
; CHECK-LABEL: and_sext_v16i8_i32:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: and v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: and x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: and w8, w8, w9
; CHECK-NEXT: and w8, w8, w8, lsr #8
; CHECK-NEXT: sxtb w0, w8
; CHECK-NEXT: ret
entry:
%y = sext <16 x i8> %x to <16 x i32>
%z = call i32 @llvm.vector.reduce.and.v16i32(<16 x i32> %y)
ret i32 %z
}
define i32 @and_zext_v16i8_i32(<16 x i8> %x) {
; CHECK-LABEL: and_zext_v16i8_i32:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: and v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: and x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: and w8, w8, w9
; CHECK-NEXT: and w8, w8, w8, lsr #8
; CHECK-NEXT: and w0, w8, #0xff
; CHECK-NEXT: ret
entry:
%y = zext <16 x i8> %x to <16 x i32>
%z = call i32 @llvm.vector.reduce.and.v16i32(<16 x i32> %y)
ret i32 %z
}
define i64 @and_sext_v8i8_i64(<8 x i8> %x) {
; CHECK-LABEL: and_sext_v8i8_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: and x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: and w8, w8, w9
; CHECK-NEXT: and w8, w8, w8, lsr #8
; CHECK-NEXT: sxtb x0, w8
; CHECK-NEXT: ret
entry:
%y = sext <8 x i8> %x to <8 x i64>
%z = call i64 @llvm.vector.reduce.and.v8i64(<8 x i64> %y)
ret i64 %z
}
define i64 @and_zext_v8i8_i64(<8 x i8> %x) {
; CHECK-LABEL: and_zext_v8i8_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: and x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: and w8, w8, w9
; CHECK-NEXT: and w8, w8, w8, lsr #8
; CHECK-NEXT: and x0, x8, #0xff
; CHECK-NEXT: ret
entry:
%y = zext <8 x i8> %x to <8 x i64>
%z = call i64 @llvm.vector.reduce.and.v8i64(<8 x i64> %y)
ret i64 %z
}
define i64 @and_sext_v16i8_i64(<16 x i8> %x) {
; CHECK-LABEL: and_sext_v16i8_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: and v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: and x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: and w8, w8, w9
; CHECK-NEXT: and w8, w8, w8, lsr #8
; CHECK-NEXT: sxtb x0, w8
; CHECK-NEXT: ret
entry:
%y = sext <16 x i8> %x to <16 x i64>
%z = call i64 @llvm.vector.reduce.and.v16i64(<16 x i64> %y)
ret i64 %z
}
define i64 @and_zext_v16i8_i64(<16 x i8> %x) {
; CHECK-LABEL: and_zext_v16i8_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: and v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: and x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: and w8, w8, w9
; CHECK-NEXT: and w8, w8, w8, lsr #8
; CHECK-NEXT: and x0, x8, #0xff
; CHECK-NEXT: ret
entry:
%y = zext <16 x i8> %x to <16 x i64>
%z = call i64 @llvm.vector.reduce.and.v16i64(<16 x i64> %y)
ret i64 %z
}
define i32 @and_sext_v4i16_i32(<4 x i16> %x) {
; CHECK-LABEL: and_sext_v4i16_i32:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: and w8, w8, w9
; CHECK-NEXT: and w8, w8, w8, lsr #16
; CHECK-NEXT: sxth w0, w8
; CHECK-NEXT: ret
entry:
%y = sext <4 x i16> %x to <4 x i32>
%z = call i32 @llvm.vector.reduce.and.v4i32(<4 x i32> %y)
ret i32 %z
}
define i32 @and_zext_v4i16_i32(<4 x i16> %x) {
; CHECK-LABEL: and_zext_v4i16_i32:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: and x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: and w0, w8, w9
; CHECK-NEXT: ret
entry:
%y = zext <4 x i16> %x to <4 x i32>
%z = call i32 @llvm.vector.reduce.and.v4i32(<4 x i32> %y)
ret i32 %z
}
define i32 @and_sext_v8i16_i32(<8 x i16> %x) {
; CHECK-LABEL: and_sext_v8i16_i32:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: and v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: and w8, w8, w9
; CHECK-NEXT: and w8, w8, w8, lsr #16
; CHECK-NEXT: sxth w0, w8
; CHECK-NEXT: ret
entry:
%y = sext <8 x i16> %x to <8 x i32>
%z = call i32 @llvm.vector.reduce.and.v8i32(<8 x i32> %y)
ret i32 %z
}
define i32 @and_zext_v8i16_i32(<8 x i16> %x) {
; CHECK-LABEL: and_zext_v8i16_i32:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: and v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: and x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: and w0, w8, w9
; CHECK-NEXT: ret
entry:
%y = zext <8 x i16> %x to <8 x i32>
%z = call i32 @llvm.vector.reduce.and.v8i32(<8 x i32> %y)
ret i32 %z
}
define i64 @and_sext_v4i16_i64(<4 x i16> %x) {
; CHECK-LABEL: and_sext_v4i16_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: and w8, w8, w9
; CHECK-NEXT: and w8, w8, w8, lsr #16
; CHECK-NEXT: sxth x0, w8
; CHECK-NEXT: ret
entry:
%y = sext <4 x i16> %x to <4 x i64>
%z = call i64 @llvm.vector.reduce.and.v4i64(<4 x i64> %y)
ret i64 %z
}
define i64 @and_zext_v4i16_i64(<4 x i16> %x) {
; CHECK-LABEL: and_zext_v4i16_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: and x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: and w0, w8, w9
; CHECK-NEXT: ret
entry:
%y = zext <4 x i16> %x to <4 x i64>
%z = call i64 @llvm.vector.reduce.and.v4i64(<4 x i64> %y)
ret i64 %z
}
define i64 @and_sext_v8i16_i64(<8 x i16> %x) {
; CHECK-LABEL: and_sext_v8i16_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: and v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: and w8, w8, w9
; CHECK-NEXT: and w8, w8, w8, lsr #16
; CHECK-NEXT: sxth x0, w8
; CHECK-NEXT: ret
entry:
%y = sext <8 x i16> %x to <8 x i64>
%z = call i64 @llvm.vector.reduce.and.v8i64(<8 x i64> %y)
ret i64 %z
}
define i64 @and_zext_v8i16_i64(<8 x i16> %x) {
; CHECK-LABEL: and_zext_v8i16_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: and v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: and x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: and w0, w8, w9
; CHECK-NEXT: ret
entry:
%y = zext <8 x i16> %x to <8 x i64>
%z = call i64 @llvm.vector.reduce.and.v8i64(<8 x i64> %y)
ret i64 %z
}
define i64 @and_sext_v2i32_i64(<2 x i32> %x) {
; CHECK-LABEL: and_sext_v2i32_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: and w8, w8, w9
; CHECK-NEXT: sxtw x0, w8
; CHECK-NEXT: ret
entry:
%y = sext <2 x i32> %x to <2 x i64>
%z = call i64 @llvm.vector.reduce.and.v2i64(<2 x i64> %y)
ret i64 %z
}
define i64 @and_zext_v2i32_i64(<2 x i32> %x) {
; CHECK-LABEL: and_zext_v2i32_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: and w0, w8, w9
; CHECK-NEXT: ret
entry:
%y = zext <2 x i32> %x to <2 x i64>
%z = call i64 @llvm.vector.reduce.and.v2i64(<2 x i64> %y)
ret i64 %z
}
define i64 @and_sext_v4i32_i64(<4 x i32> %x) {
; CHECK-LABEL: and_sext_v4i32_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: and v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: and w8, w8, w9
; CHECK-NEXT: sxtw x0, w8
; CHECK-NEXT: ret
entry:
%y = sext <4 x i32> %x to <4 x i64>
%z = call i64 @llvm.vector.reduce.and.v4i64(<4 x i64> %y)
ret i64 %z
}
define i64 @and_zext_v4i32_i64(<4 x i32> %x) {
; CHECK-LABEL: and_zext_v4i32_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: and v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: and w0, w8, w9
; CHECK-NEXT: ret
entry:
%y = zext <4 x i32> %x to <4 x i64>
%z = call i64 @llvm.vector.reduce.and.v4i64(<4 x i64> %y)
ret i64 %z
}
define zeroext i16 @or_sext_v8i8_i16(<8 x i8> %x) {
; CHECK-LABEL: or_sext_v8i8_i16:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: orr x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: orr w8, w8, w9
; CHECK-NEXT: orr w8, w8, w8, lsr #8
; CHECK-NEXT: sxtb w8, w8
; CHECK-NEXT: and w0, w8, #0xffff
; CHECK-NEXT: ret
entry:
%y = sext <8 x i8> %x to <8 x i16>
%z = call i16 @llvm.vector.reduce.or.v8i16(<8 x i16> %y)
ret i16 %z
}
define zeroext i16 @or_zext_v8i8_i16(<8 x i8> %x) {
; CHECK-LABEL: or_zext_v8i8_i16:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: orr x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: orr w8, w8, w9
; CHECK-NEXT: orr w8, w8, w8, lsr #8
; CHECK-NEXT: and w0, w8, #0xff
; CHECK-NEXT: ret
entry:
%y = zext <8 x i8> %x to <8 x i16>
%z = call i16 @llvm.vector.reduce.or.v8i16(<8 x i16> %y)
ret i16 %z
}
define zeroext i16 @or_sext_v16i8_i16(<16 x i8> %x) {
; CHECK-LABEL: or_sext_v16i8_i16:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: orr v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: orr x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: orr w8, w8, w9
; CHECK-NEXT: orr w8, w8, w8, lsr #8
; CHECK-NEXT: sxtb w8, w8
; CHECK-NEXT: and w0, w8, #0xffff
; CHECK-NEXT: ret
entry:
%y = sext <16 x i8> %x to <16 x i16>
%z = call i16 @llvm.vector.reduce.or.v16i16(<16 x i16> %y)
ret i16 %z
}
define zeroext i16 @or_zext_v16i8_i16(<16 x i8> %x) {
; CHECK-LABEL: or_zext_v16i8_i16:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: orr v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: orr x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: orr w8, w8, w9
; CHECK-NEXT: orr w8, w8, w8, lsr #8
; CHECK-NEXT: and w0, w8, #0xff
; CHECK-NEXT: ret
entry:
%y = zext <16 x i8> %x to <16 x i16>
%z = call i16 @llvm.vector.reduce.or.v16i16(<16 x i16> %y)
ret i16 %z
}
define i32 @or_sext_v8i8_i32(<8 x i8> %x) {
; CHECK-LABEL: or_sext_v8i8_i32:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: orr x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: orr w8, w8, w9
; CHECK-NEXT: orr w8, w8, w8, lsr #8
; CHECK-NEXT: sxtb w0, w8
; CHECK-NEXT: ret
entry:
%y = sext <8 x i8> %x to <8 x i32>
%z = call i32 @llvm.vector.reduce.or.v8i32(<8 x i32> %y)
ret i32 %z
}
define i32 @or_zext_v8i8_i32(<8 x i8> %x) {
; CHECK-LABEL: or_zext_v8i8_i32:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: orr x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: orr w8, w8, w9
; CHECK-NEXT: orr w8, w8, w8, lsr #8
; CHECK-NEXT: and w0, w8, #0xff
; CHECK-NEXT: ret
entry:
%y = zext <8 x i8> %x to <8 x i32>
%z = call i32 @llvm.vector.reduce.or.v8i32(<8 x i32> %y)
ret i32 %z
}
define i32 @or_sext_v16i8_i32(<16 x i8> %x) {
; CHECK-LABEL: or_sext_v16i8_i32:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: orr v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: orr x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: orr w8, w8, w9
; CHECK-NEXT: orr w8, w8, w8, lsr #8
; CHECK-NEXT: sxtb w0, w8
; CHECK-NEXT: ret
entry:
%y = sext <16 x i8> %x to <16 x i32>
%z = call i32 @llvm.vector.reduce.or.v16i32(<16 x i32> %y)
ret i32 %z
}
define i32 @or_zext_v16i8_i32(<16 x i8> %x) {
; CHECK-LABEL: or_zext_v16i8_i32:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: orr v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: orr x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: orr w8, w8, w9
; CHECK-NEXT: orr w8, w8, w8, lsr #8
; CHECK-NEXT: and w0, w8, #0xff
; CHECK-NEXT: ret
entry:
%y = zext <16 x i8> %x to <16 x i32>
%z = call i32 @llvm.vector.reduce.or.v16i32(<16 x i32> %y)
ret i32 %z
}
define i64 @or_sext_v8i8_i64(<8 x i8> %x) {
; CHECK-LABEL: or_sext_v8i8_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: orr x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: orr w8, w8, w9
; CHECK-NEXT: orr w8, w8, w8, lsr #8
; CHECK-NEXT: sxtb x0, w8
; CHECK-NEXT: ret
entry:
%y = sext <8 x i8> %x to <8 x i64>
%z = call i64 @llvm.vector.reduce.or.v8i64(<8 x i64> %y)
ret i64 %z
}
define i64 @or_zext_v8i8_i64(<8 x i8> %x) {
; CHECK-LABEL: or_zext_v8i8_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: orr x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: orr w8, w8, w9
; CHECK-NEXT: orr w8, w8, w8, lsr #8
; CHECK-NEXT: and x0, x8, #0xff
; CHECK-NEXT: ret
entry:
%y = zext <8 x i8> %x to <8 x i64>
%z = call i64 @llvm.vector.reduce.or.v8i64(<8 x i64> %y)
ret i64 %z
}
define i64 @or_sext_v16i8_i64(<16 x i8> %x) {
; CHECK-LABEL: or_sext_v16i8_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: orr v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: orr x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: orr w8, w8, w9
; CHECK-NEXT: orr w8, w8, w8, lsr #8
; CHECK-NEXT: sxtb x0, w8
; CHECK-NEXT: ret
entry:
%y = sext <16 x i8> %x to <16 x i64>
%z = call i64 @llvm.vector.reduce.or.v16i64(<16 x i64> %y)
ret i64 %z
}
define i64 @or_zext_v16i8_i64(<16 x i8> %x) {
; CHECK-LABEL: or_zext_v16i8_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: orr v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: orr x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: orr w8, w8, w9
; CHECK-NEXT: orr w8, w8, w8, lsr #8
; CHECK-NEXT: and x0, x8, #0xff
; CHECK-NEXT: ret
entry:
%y = zext <16 x i8> %x to <16 x i64>
%z = call i64 @llvm.vector.reduce.or.v16i64(<16 x i64> %y)
ret i64 %z
}
define i32 @or_sext_v4i16_i32(<4 x i16> %x) {
; CHECK-LABEL: or_sext_v4i16_i32:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: orr w8, w8, w9
; CHECK-NEXT: orr w8, w8, w8, lsr #16
; CHECK-NEXT: sxth w0, w8
; CHECK-NEXT: ret
entry:
%y = sext <4 x i16> %x to <4 x i32>
%z = call i32 @llvm.vector.reduce.or.v4i32(<4 x i32> %y)
ret i32 %z
}
define i32 @or_zext_v4i16_i32(<4 x i16> %x) {
; CHECK-LABEL: or_zext_v4i16_i32:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: orr w8, w8, w9
; CHECK-NEXT: orr w8, w8, w8, lsr #16
; CHECK-NEXT: and w0, w8, #0xffff
; CHECK-NEXT: ret
entry:
%y = zext <4 x i16> %x to <4 x i32>
%z = call i32 @llvm.vector.reduce.or.v4i32(<4 x i32> %y)
ret i32 %z
}
define i32 @or_sext_v8i16_i32(<8 x i16> %x) {
; CHECK-LABEL: or_sext_v8i16_i32:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: orr v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: orr w8, w8, w9
; CHECK-NEXT: orr w8, w8, w8, lsr #16
; CHECK-NEXT: sxth w0, w8
; CHECK-NEXT: ret
entry:
%y = sext <8 x i16> %x to <8 x i32>
%z = call i32 @llvm.vector.reduce.or.v8i32(<8 x i32> %y)
ret i32 %z
}
define i32 @or_zext_v8i16_i32(<8 x i16> %x) {
; CHECK-LABEL: or_zext_v8i16_i32:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: orr v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: orr w8, w8, w9
; CHECK-NEXT: orr w8, w8, w8, lsr #16
; CHECK-NEXT: and w0, w8, #0xffff
; CHECK-NEXT: ret
entry:
%y = zext <8 x i16> %x to <8 x i32>
%z = call i32 @llvm.vector.reduce.or.v8i32(<8 x i32> %y)
ret i32 %z
}
define i64 @or_sext_v4i16_i64(<4 x i16> %x) {
; CHECK-LABEL: or_sext_v4i16_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: orr w8, w8, w9
; CHECK-NEXT: orr w8, w8, w8, lsr #16
; CHECK-NEXT: sxth x0, w8
; CHECK-NEXT: ret
entry:
%y = sext <4 x i16> %x to <4 x i64>
%z = call i64 @llvm.vector.reduce.or.v4i64(<4 x i64> %y)
ret i64 %z
}
define i64 @or_zext_v4i16_i64(<4 x i16> %x) {
; CHECK-LABEL: or_zext_v4i16_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: orr w8, w8, w9
; CHECK-NEXT: orr w8, w8, w8, lsr #16
; CHECK-NEXT: and x0, x8, #0xffff
; CHECK-NEXT: ret
entry:
%y = zext <4 x i16> %x to <4 x i64>
%z = call i64 @llvm.vector.reduce.or.v4i64(<4 x i64> %y)
ret i64 %z
}
define i64 @or_sext_v8i16_i64(<8 x i16> %x) {
; CHECK-LABEL: or_sext_v8i16_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: orr v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: orr w8, w8, w9
; CHECK-NEXT: orr w8, w8, w8, lsr #16
; CHECK-NEXT: sxth x0, w8
; CHECK-NEXT: ret
entry:
%y = sext <8 x i16> %x to <8 x i64>
%z = call i64 @llvm.vector.reduce.or.v8i64(<8 x i64> %y)
ret i64 %z
}
define i64 @or_zext_v8i16_i64(<8 x i16> %x) {
; CHECK-LABEL: or_zext_v8i16_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: orr v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: orr w8, w8, w9
; CHECK-NEXT: orr w8, w8, w8, lsr #16
; CHECK-NEXT: and x0, x8, #0xffff
; CHECK-NEXT: ret
entry:
%y = zext <8 x i16> %x to <8 x i64>
%z = call i64 @llvm.vector.reduce.or.v8i64(<8 x i64> %y)
ret i64 %z
}
define i64 @or_sext_v2i32_i64(<2 x i32> %x) {
; CHECK-LABEL: or_sext_v2i32_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: orr w8, w8, w9
; CHECK-NEXT: sxtw x0, w8
; CHECK-NEXT: ret
entry:
%y = sext <2 x i32> %x to <2 x i64>
%z = call i64 @llvm.vector.reduce.or.v2i64(<2 x i64> %y)
ret i64 %z
}
define i64 @or_zext_v2i32_i64(<2 x i32> %x) {
; CHECK-LABEL: or_zext_v2i32_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: orr w0, w8, w9
; CHECK-NEXT: ret
entry:
%y = zext <2 x i32> %x to <2 x i64>
%z = call i64 @llvm.vector.reduce.or.v2i64(<2 x i64> %y)
ret i64 %z
}
define i64 @or_sext_v4i32_i64(<4 x i32> %x) {
; CHECK-LABEL: or_sext_v4i32_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: orr v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: orr w8, w8, w9
; CHECK-NEXT: sxtw x0, w8
; CHECK-NEXT: ret
entry:
%y = sext <4 x i32> %x to <4 x i64>
%z = call i64 @llvm.vector.reduce.or.v4i64(<4 x i64> %y)
ret i64 %z
}
define i64 @or_zext_v4i32_i64(<4 x i32> %x) {
; CHECK-LABEL: or_zext_v4i32_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: orr v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: orr w0, w8, w9
; CHECK-NEXT: ret
entry:
%y = zext <4 x i32> %x to <4 x i64>
%z = call i64 @llvm.vector.reduce.or.v4i64(<4 x i64> %y)
ret i64 %z
}
define zeroext i16 @xor_sext_v8i8_i16(<8 x i8> %x) {
; CHECK-LABEL: xor_sext_v8i8_i16:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: eor x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: eor w8, w8, w9
; CHECK-NEXT: eor w8, w8, w8, lsr #8
; CHECK-NEXT: sxtb w8, w8
; CHECK-NEXT: and w0, w8, #0xffff
; CHECK-NEXT: ret
entry:
%y = sext <8 x i8> %x to <8 x i16>
%z = call i16 @llvm.vector.reduce.xor.v8i16(<8 x i16> %y)
ret i16 %z
}
define zeroext i16 @xor_zext_v8i8_i16(<8 x i8> %x) {
; CHECK-LABEL: xor_zext_v8i8_i16:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: eor x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: eor w8, w8, w9
; CHECK-NEXT: eor w8, w8, w8, lsr #8
; CHECK-NEXT: and w0, w8, #0xff
; CHECK-NEXT: ret
entry:
%y = zext <8 x i8> %x to <8 x i16>
%z = call i16 @llvm.vector.reduce.xor.v8i16(<8 x i16> %y)
ret i16 %z
}
define zeroext i16 @xor_sext_v16i8_i16(<16 x i8> %x) {
; CHECK-LABEL: xor_sext_v16i8_i16:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: eor v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: eor x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: eor w8, w8, w9
; CHECK-NEXT: eor w8, w8, w8, lsr #8
; CHECK-NEXT: sxtb w8, w8
; CHECK-NEXT: and w0, w8, #0xffff
; CHECK-NEXT: ret
entry:
%y = sext <16 x i8> %x to <16 x i16>
%z = call i16 @llvm.vector.reduce.xor.v16i16(<16 x i16> %y)
ret i16 %z
}
define zeroext i16 @xor_zext_v16i8_i16(<16 x i8> %x) {
; CHECK-LABEL: xor_zext_v16i8_i16:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: eor v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: eor x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: eor w8, w8, w9
; CHECK-NEXT: eor w8, w8, w8, lsr #8
; CHECK-NEXT: and w0, w8, #0xff
; CHECK-NEXT: ret
entry:
%y = zext <16 x i8> %x to <16 x i16>
%z = call i16 @llvm.vector.reduce.xor.v16i16(<16 x i16> %y)
ret i16 %z
}
define i32 @xor_sext_v8i8_i32(<8 x i8> %x) {
; CHECK-LABEL: xor_sext_v8i8_i32:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: eor x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: eor w8, w8, w9
; CHECK-NEXT: eor w8, w8, w8, lsr #8
; CHECK-NEXT: sxtb w0, w8
; CHECK-NEXT: ret
entry:
%y = sext <8 x i8> %x to <8 x i32>
%z = call i32 @llvm.vector.reduce.xor.v8i32(<8 x i32> %y)
ret i32 %z
}
define i32 @xor_zext_v8i8_i32(<8 x i8> %x) {
; CHECK-LABEL: xor_zext_v8i8_i32:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: eor x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: eor w8, w8, w9
; CHECK-NEXT: eor w8, w8, w8, lsr #8
; CHECK-NEXT: and w0, w8, #0xff
; CHECK-NEXT: ret
entry:
%y = zext <8 x i8> %x to <8 x i32>
%z = call i32 @llvm.vector.reduce.xor.v8i32(<8 x i32> %y)
ret i32 %z
}
define i32 @xor_sext_v16i8_i32(<16 x i8> %x) {
; CHECK-LABEL: xor_sext_v16i8_i32:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: eor v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: eor x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: eor w8, w8, w9
; CHECK-NEXT: eor w8, w8, w8, lsr #8
; CHECK-NEXT: sxtb w0, w8
; CHECK-NEXT: ret
entry:
%y = sext <16 x i8> %x to <16 x i32>
%z = call i32 @llvm.vector.reduce.xor.v16i32(<16 x i32> %y)
ret i32 %z
}
define i32 @xor_zext_v16i8_i32(<16 x i8> %x) {
; CHECK-LABEL: xor_zext_v16i8_i32:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: eor v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: eor x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: eor w8, w8, w9
; CHECK-NEXT: eor w8, w8, w8, lsr #8
; CHECK-NEXT: and w0, w8, #0xff
; CHECK-NEXT: ret
entry:
%y = zext <16 x i8> %x to <16 x i32>
%z = call i32 @llvm.vector.reduce.xor.v16i32(<16 x i32> %y)
ret i32 %z
}
define i64 @xor_sext_v8i8_i64(<8 x i8> %x) {
; CHECK-LABEL: xor_sext_v8i8_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: eor x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: eor w8, w8, w9
; CHECK-NEXT: eor w8, w8, w8, lsr #8
; CHECK-NEXT: sxtb x0, w8
; CHECK-NEXT: ret
entry:
%y = sext <8 x i8> %x to <8 x i64>
%z = call i64 @llvm.vector.reduce.xor.v8i64(<8 x i64> %y)
ret i64 %z
}
define i64 @xor_zext_v8i8_i64(<8 x i8> %x) {
; CHECK-LABEL: xor_zext_v8i8_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: eor x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: eor w8, w8, w9
; CHECK-NEXT: eor w8, w8, w8, lsr #8
; CHECK-NEXT: and x0, x8, #0xff
; CHECK-NEXT: ret
entry:
%y = zext <8 x i8> %x to <8 x i64>
%z = call i64 @llvm.vector.reduce.xor.v8i64(<8 x i64> %y)
ret i64 %z
}
define i64 @xor_sext_v16i8_i64(<16 x i8> %x) {
; CHECK-LABEL: xor_sext_v16i8_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: eor v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: eor x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: eor w8, w8, w9
; CHECK-NEXT: eor w8, w8, w8, lsr #8
; CHECK-NEXT: sxtb x0, w8
; CHECK-NEXT: ret
entry:
%y = sext <16 x i8> %x to <16 x i64>
%z = call i64 @llvm.vector.reduce.xor.v16i64(<16 x i64> %y)
ret i64 %z
}
define i64 @xor_zext_v16i8_i64(<16 x i8> %x) {
; CHECK-LABEL: xor_zext_v16i8_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: eor v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: eor x8, x8, x8, lsr #32
; CHECK-NEXT: lsr x9, x8, #16
; CHECK-NEXT: eor w8, w8, w9
; CHECK-NEXT: eor w8, w8, w8, lsr #8
; CHECK-NEXT: and x0, x8, #0xff
; CHECK-NEXT: ret
entry:
%y = zext <16 x i8> %x to <16 x i64>
%z = call i64 @llvm.vector.reduce.xor.v16i64(<16 x i64> %y)
ret i64 %z
}
define i32 @xor_sext_v4i16_i32(<4 x i16> %x) {
; CHECK-LABEL: xor_sext_v4i16_i32:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: eor w8, w8, w9
; CHECK-NEXT: eor w8, w8, w8, lsr #16
; CHECK-NEXT: sxth w0, w8
; CHECK-NEXT: ret
entry:
%y = sext <4 x i16> %x to <4 x i32>
%z = call i32 @llvm.vector.reduce.xor.v4i32(<4 x i32> %y)
ret i32 %z
}
define i32 @xor_zext_v4i16_i32(<4 x i16> %x) {
; CHECK-LABEL: xor_zext_v4i16_i32:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: eor w8, w8, w9
; CHECK-NEXT: eor w8, w8, w8, lsr #16
; CHECK-NEXT: and w0, w8, #0xffff
; CHECK-NEXT: ret
entry:
%y = zext <4 x i16> %x to <4 x i32>
%z = call i32 @llvm.vector.reduce.xor.v4i32(<4 x i32> %y)
ret i32 %z
}
define i32 @xor_sext_v8i16_i32(<8 x i16> %x) {
; CHECK-LABEL: xor_sext_v8i16_i32:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: eor v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: eor w8, w8, w9
; CHECK-NEXT: eor w8, w8, w8, lsr #16
; CHECK-NEXT: sxth w0, w8
; CHECK-NEXT: ret
entry:
%y = sext <8 x i16> %x to <8 x i32>
%z = call i32 @llvm.vector.reduce.xor.v8i32(<8 x i32> %y)
ret i32 %z
}
define i32 @xor_zext_v8i16_i32(<8 x i16> %x) {
; CHECK-LABEL: xor_zext_v8i16_i32:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: eor v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: eor w8, w8, w9
; CHECK-NEXT: eor w8, w8, w8, lsr #16
; CHECK-NEXT: and w0, w8, #0xffff
; CHECK-NEXT: ret
entry:
%y = zext <8 x i16> %x to <8 x i32>
%z = call i32 @llvm.vector.reduce.xor.v8i32(<8 x i32> %y)
ret i32 %z
}
define i64 @xor_sext_v4i16_i64(<4 x i16> %x) {
; CHECK-LABEL: xor_sext_v4i16_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: eor w8, w8, w9
; CHECK-NEXT: eor w8, w8, w8, lsr #16
; CHECK-NEXT: sxth x0, w8
; CHECK-NEXT: ret
entry:
%y = sext <4 x i16> %x to <4 x i64>
%z = call i64 @llvm.vector.reduce.xor.v4i64(<4 x i64> %y)
ret i64 %z
}
define i64 @xor_zext_v4i16_i64(<4 x i16> %x) {
; CHECK-LABEL: xor_zext_v4i16_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: eor w8, w8, w9
; CHECK-NEXT: eor w8, w8, w8, lsr #16
; CHECK-NEXT: and x0, x8, #0xffff
; CHECK-NEXT: ret
entry:
%y = zext <4 x i16> %x to <4 x i64>
%z = call i64 @llvm.vector.reduce.xor.v4i64(<4 x i64> %y)
ret i64 %z
}
define i64 @xor_sext_v8i16_i64(<8 x i16> %x) {
; CHECK-LABEL: xor_sext_v8i16_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: eor v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: eor w8, w8, w9
; CHECK-NEXT: eor w8, w8, w8, lsr #16
; CHECK-NEXT: sxth x0, w8
; CHECK-NEXT: ret
entry:
%y = sext <8 x i16> %x to <8 x i64>
%z = call i64 @llvm.vector.reduce.xor.v8i64(<8 x i64> %y)
ret i64 %z
}
define i64 @xor_zext_v8i16_i64(<8 x i16> %x) {
; CHECK-LABEL: xor_zext_v8i16_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: eor v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: eor w8, w8, w9
; CHECK-NEXT: eor w8, w8, w8, lsr #16
; CHECK-NEXT: and x0, x8, #0xffff
; CHECK-NEXT: ret
entry:
%y = zext <8 x i16> %x to <8 x i64>
%z = call i64 @llvm.vector.reduce.xor.v8i64(<8 x i64> %y)
ret i64 %z
}
define i64 @xor_sext_v2i32_i64(<2 x i32> %x) {
; CHECK-LABEL: xor_sext_v2i32_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: eor w8, w8, w9
; CHECK-NEXT: sxtw x0, w8
; CHECK-NEXT: ret
entry:
%y = sext <2 x i32> %x to <2 x i64>
%z = call i64 @llvm.vector.reduce.xor.v2i64(<2 x i64> %y)
ret i64 %z
}
define i64 @xor_zext_v2i32_i64(<2 x i32> %x) {
; CHECK-LABEL: xor_zext_v2i32_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: eor w0, w8, w9
; CHECK-NEXT: ret
entry:
%y = zext <2 x i32> %x to <2 x i64>
%z = call i64 @llvm.vector.reduce.xor.v2i64(<2 x i64> %y)
ret i64 %z
}
define i64 @xor_sext_v4i32_i64(<4 x i32> %x) {
; CHECK-LABEL: xor_sext_v4i32_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: eor v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: eor w8, w8, w9
; CHECK-NEXT: sxtw x0, w8
; CHECK-NEXT: ret
entry:
%y = sext <4 x i32> %x to <4 x i64>
%z = call i64 @llvm.vector.reduce.xor.v4i64(<4 x i64> %y)
ret i64 %z
}
define i64 @xor_zext_v4i32_i64(<4 x i32> %x) {
; CHECK-LABEL: xor_zext_v4i32_i64:
; CHECK: // %bb.0: // %entry
; CHECK-NEXT: ext v1.16b, v0.16b, v0.16b, #8
; CHECK-NEXT: eor v0.8b, v0.8b, v1.8b
; CHECK-NEXT: fmov x8, d0
; CHECK-NEXT: lsr x9, x8, #32
; CHECK-NEXT: eor w0, w8, w9
; CHECK-NEXT: ret
entry:
%y = zext <4 x i32> %x to <4 x i64>
%z = call i64 @llvm.vector.reduce.xor.v4i64(<4 x i64> %y)
ret i64 %z
}
declare i16 @llvm.vector.reduce.and.v16i16(<16 x i16>)
declare i16 @llvm.vector.reduce.and.v8i16(<8 x i16>)
declare i16 @llvm.vector.reduce.or.v16i16(<16 x i16>)
declare i16 @llvm.vector.reduce.or.v8i16(<8 x i16>)
declare i16 @llvm.vector.reduce.xor.v16i16(<16 x i16>)
declare i16 @llvm.vector.reduce.xor.v8i16(<8 x i16>)
declare i32 @llvm.vector.reduce.and.v16i32(<16 x i32>)
declare i32 @llvm.vector.reduce.and.v4i32(<4 x i32>)
declare i32 @llvm.vector.reduce.and.v8i32(<8 x i32>)
declare i32 @llvm.vector.reduce.or.v16i32(<16 x i32>)
declare i32 @llvm.vector.reduce.or.v4i32(<4 x i32>)
declare i32 @llvm.vector.reduce.or.v8i32(<8 x i32>)
declare i32 @llvm.vector.reduce.xor.v16i32(<16 x i32>)
declare i32 @llvm.vector.reduce.xor.v4i32(<4 x i32>)
declare i32 @llvm.vector.reduce.xor.v8i32(<8 x i32>)
declare i64 @llvm.vector.reduce.and.v16i64(<16 x i64>)
declare i64 @llvm.vector.reduce.and.v2i64(<2 x i64>)
declare i64 @llvm.vector.reduce.and.v4i64(<4 x i64>)
declare i64 @llvm.vector.reduce.and.v8i64(<8 x i64>)
declare i64 @llvm.vector.reduce.or.v16i64(<16 x i64>)
declare i64 @llvm.vector.reduce.or.v2i64(<2 x i64>)
declare i64 @llvm.vector.reduce.or.v4i64(<4 x i64>)
declare i64 @llvm.vector.reduce.or.v8i64(<8 x i64>)
declare i64 @llvm.vector.reduce.xor.v16i64(<16 x i64>)
declare i64 @llvm.vector.reduce.xor.v2i64(<2 x i64>)
declare i64 @llvm.vector.reduce.xor.v4i64(<4 x i64>)
declare i64 @llvm.vector.reduce.xor.v8i64(<8 x i64>)