; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc -mtriple=x86_64-linux -mcpu=skylake-avx512 < %s | FileCheck %s
define void @scatter_scale_512(ptr %result, <4 x i64> %idx, <4 x i1> %mask) {
; CHECK-LABEL: scatter_scale_512:
; CHECK: # %bb.0:
; CHECK-NEXT: vpslld $31, %xmm1, %xmm1
; CHECK-NEXT: vpmovd2m %xmm1, %k1
; CHECK-NEXT: vpsllq $9, %ymm0, %ymm0
; CHECK-NEXT: vpxor %xmm1, %xmm1, %xmm1
; CHECK-NEXT: vscatterqpd %ymm1, (%rdi,%ymm0) {%k1}
; CHECK-NEXT: vzeroupper
; CHECK-NEXT: retq
%gep = getelementptr inbounds [512 x i8], ptr %result, <4 x i64> %idx
call void @llvm.masked.scatter.v4f64.v4p0(<4 x double> zeroinitializer, <4 x ptr> %gep, i32 0, <4 x i1> %mask)
ret void
}
define void @scatter_scale_16(ptr %result, <4 x i64> %idx, <4 x i1> %mask) {
; CHECK-LABEL: scatter_scale_16:
; CHECK: # %bb.0:
; CHECK-NEXT: vpslld $31, %xmm1, %xmm1
; CHECK-NEXT: vpmovd2m %xmm1, %k1
; CHECK-NEXT: vpsllq $4, %ymm0, %ymm0
; CHECK-NEXT: vpxor %xmm1, %xmm1, %xmm1
; CHECK-NEXT: vscatterqpd %ymm1, (%rdi,%ymm0) {%k1}
; CHECK-NEXT: vzeroupper
; CHECK-NEXT: retq
%gep = getelementptr inbounds [16 x i8], ptr %result, <4 x i64> %idx
call void @llvm.masked.scatter.v4f64.v4p0(<4 x double> zeroinitializer, <4 x ptr> %gep, i32 0, <4 x i1> %mask)
ret void
}
define void @scatter_scale_8(ptr %result, <4 x i64> %idx, <4 x i1> %mask) {
; CHECK-LABEL: scatter_scale_8:
; CHECK: # %bb.0:
; CHECK-NEXT: vpslld $31, %xmm1, %xmm1
; CHECK-NEXT: vpmovd2m %xmm1, %k1
; CHECK-NEXT: vpxor %xmm1, %xmm1, %xmm1
; CHECK-NEXT: vscatterqpd %ymm1, (%rdi,%ymm0,8) {%k1}
; CHECK-NEXT: vzeroupper
; CHECK-NEXT: retq
%gep = getelementptr inbounds [8 x i8], ptr %result, <4 x i64> %idx
call void @llvm.masked.scatter.v4f64.v4p0(<4 x double> zeroinitializer, <4 x ptr> %gep, i32 0, <4 x i1> %mask)
ret void
}
define void @scatter_scale_4(ptr %result, <4 x i64> %idx, <4 x i1> %mask) {
; CHECK-LABEL: scatter_scale_4:
; CHECK: # %bb.0:
; CHECK-NEXT: vpslld $31, %xmm1, %xmm1
; CHECK-NEXT: vpmovd2m %xmm1, %k1
; CHECK-NEXT: vpxor %xmm1, %xmm1, %xmm1
; CHECK-NEXT: vscatterqpd %ymm1, (%rdi,%ymm0,4) {%k1}
; CHECK-NEXT: vzeroupper
; CHECK-NEXT: retq
%gep = getelementptr inbounds [4 x i8], ptr %result, <4 x i64> %idx
call void @llvm.masked.scatter.v4f64.v4p0(<4 x double> zeroinitializer, <4 x ptr> %gep, i32 0, <4 x i1> %mask)
ret void
}
define void @scatter_scale_3(ptr %result, <4 x i64> %idx, <4 x i1> %mask) {
; CHECK-LABEL: scatter_scale_3:
; CHECK: # %bb.0:
; CHECK-NEXT: vpslld $31, %xmm1, %xmm1
; CHECK-NEXT: vpmovd2m %xmm1, %k1
; CHECK-NEXT: vpaddq %ymm0, %ymm0, %ymm1
; CHECK-NEXT: vpaddq %ymm0, %ymm1, %ymm0
; CHECK-NEXT: vpxor %xmm1, %xmm1, %xmm1
; CHECK-NEXT: vscatterqpd %ymm1, (%rdi,%ymm0) {%k1}
; CHECK-NEXT: vzeroupper
; CHECK-NEXT: retq
%gep = getelementptr inbounds [3 x i8], ptr %result, <4 x i64> %idx
call void @llvm.masked.scatter.v4f64.v4p0(<4 x double> zeroinitializer, <4 x ptr> %gep, i32 0, <4 x i1> %mask)
ret void
}
define void @scatter_scale_1(ptr %result, <4 x i64> %idx, <4 x i1> %mask) {
; CHECK-LABEL: scatter_scale_1:
; CHECK: # %bb.0:
; CHECK-NEXT: vpslld $31, %xmm1, %xmm1
; CHECK-NEXT: vpmovd2m %xmm1, %k1
; CHECK-NEXT: vpxor %xmm1, %xmm1, %xmm1
; CHECK-NEXT: vscatterqpd %ymm1, (%rdi,%ymm0) {%k1}
; CHECK-NEXT: vzeroupper
; CHECK-NEXT: retq
%gep = getelementptr inbounds [1 x i8], ptr %result, <4 x i64> %idx
call void @llvm.masked.scatter.v4f64.v4p0(<4 x double> zeroinitializer, <4 x ptr> %gep, i32 0, <4 x i1> %mask)
ret void
}
define <4 x double> @gather_scale_512(ptr %result, <4 x i64> %idx, <4 x i1> %mask) {
; CHECK-LABEL: gather_scale_512:
; CHECK: # %bb.0:
; CHECK-NEXT: vpslld $31, %xmm1, %xmm1
; CHECK-NEXT: vpmovd2m %xmm1, %k1
; CHECK-NEXT: vpsllq $9, %ymm0, %ymm1
; CHECK-NEXT: vpxor %xmm0, %xmm0, %xmm0
; CHECK-NEXT: vgatherqpd (%rdi,%ymm1), %ymm0 {%k1}
; CHECK-NEXT: retq
%gep = getelementptr inbounds [512 x i8], ptr %result, <4 x i64> %idx
%res = call <4 x double> @llvm.masked.gather.v4f64.v4p0(<4 x ptr> %gep, i32 0, <4 x i1> %mask, <4 x double> zeroinitializer)
ret <4 x double> %res
}
define <4 x double> @gather_scale_16(ptr %result, <4 x i64> %idx, <4 x i1> %mask) {
; CHECK-LABEL: gather_scale_16:
; CHECK: # %bb.0:
; CHECK-NEXT: vpslld $31, %xmm1, %xmm1
; CHECK-NEXT: vpmovd2m %xmm1, %k1
; CHECK-NEXT: vpsllq $4, %ymm0, %ymm1
; CHECK-NEXT: vpxor %xmm0, %xmm0, %xmm0
; CHECK-NEXT: vgatherqpd (%rdi,%ymm1), %ymm0 {%k1}
; CHECK-NEXT: retq
%gep = getelementptr inbounds [16 x i8], ptr %result, <4 x i64> %idx
%res = call <4 x double> @llvm.masked.gather.v4f64.v4p0(<4 x ptr> %gep, i32 0, <4 x i1> %mask, <4 x double> zeroinitializer)
ret <4 x double> %res
}
define <4 x double> @gather_scale_8(ptr %result, <4 x i64> %idx, <4 x i1> %mask) {
; CHECK-LABEL: gather_scale_8:
; CHECK: # %bb.0:
; CHECK-NEXT: vpslld $31, %xmm1, %xmm1
; CHECK-NEXT: vpmovd2m %xmm1, %k1
; CHECK-NEXT: vpxor %xmm1, %xmm1, %xmm1
; CHECK-NEXT: vgatherqpd (%rdi,%ymm0,8), %ymm1 {%k1}
; CHECK-NEXT: vmovapd %ymm1, %ymm0
; CHECK-NEXT: retq
%gep = getelementptr inbounds [8 x i8], ptr %result, <4 x i64> %idx
%res = call <4 x double> @llvm.masked.gather.v4f64.v4p0(<4 x ptr> %gep, i32 0, <4 x i1> %mask, <4 x double> zeroinitializer)
ret <4 x double> %res
}
define <4 x double> @gather_scale_4(ptr %result, <4 x i64> %idx, <4 x i1> %mask) {
; CHECK-LABEL: gather_scale_4:
; CHECK: # %bb.0:
; CHECK-NEXT: vpslld $31, %xmm1, %xmm1
; CHECK-NEXT: vpmovd2m %xmm1, %k1
; CHECK-NEXT: vpxor %xmm1, %xmm1, %xmm1
; CHECK-NEXT: vgatherqpd (%rdi,%ymm0,4), %ymm1 {%k1}
; CHECK-NEXT: vmovapd %ymm1, %ymm0
; CHECK-NEXT: retq
%gep = getelementptr inbounds [4 x i8], ptr %result, <4 x i64> %idx
%res = call <4 x double> @llvm.masked.gather.v4f64.v4p0(<4 x ptr> %gep, i32 0, <4 x i1> %mask, <4 x double> zeroinitializer)
ret <4 x double> %res
}
define <4 x double> @gather_scale_3(ptr %result, <4 x i64> %idx, <4 x i1> %mask) {
; CHECK-LABEL: gather_scale_3:
; CHECK: # %bb.0:
; CHECK-NEXT: vpslld $31, %xmm1, %xmm1
; CHECK-NEXT: vpmovd2m %xmm1, %k1
; CHECK-NEXT: vpaddq %ymm0, %ymm0, %ymm1
; CHECK-NEXT: vpaddq %ymm0, %ymm1, %ymm1
; CHECK-NEXT: vpxor %xmm0, %xmm0, %xmm0
; CHECK-NEXT: vgatherqpd (%rdi,%ymm1), %ymm0 {%k1}
; CHECK-NEXT: retq
%gep = getelementptr inbounds [3 x i8], ptr %result, <4 x i64> %idx
%res = call <4 x double> @llvm.masked.gather.v4f64.v4p0(<4 x ptr> %gep, i32 0, <4 x i1> %mask, <4 x double> zeroinitializer)
ret <4 x double> %res
}
define <4 x double> @gather_scale_1(ptr %result, <4 x i64> %idx, <4 x i1> %mask) {
; CHECK-LABEL: gather_scale_1:
; CHECK: # %bb.0:
; CHECK-NEXT: vpslld $31, %xmm1, %xmm1
; CHECK-NEXT: vpmovd2m %xmm1, %k1
; CHECK-NEXT: vpxor %xmm1, %xmm1, %xmm1
; CHECK-NEXT: vgatherqpd (%rdi,%ymm0), %ymm1 {%k1}
; CHECK-NEXT: vmovapd %ymm1, %ymm0
; CHECK-NEXT: retq
%gep = getelementptr inbounds [1 x i8], ptr %result, <4 x i64> %idx
%res = call <4 x double> @llvm.masked.gather.v4f64.v4p0(<4 x ptr> %gep, i32 0, <4 x i1> %mask, <4 x double> zeroinitializer)
ret <4 x double> %res
}
declare void @llvm.masked.scatter.v4f64.v4p0(<4 x double>, <4 x ptr>, i32 immarg, <4 x i1>)
declare <4 x double> @llvm.masked.gather.v4f64.v4p0(<4 x ptr>, i32 immarg, <4 x i1>, <4 x double>)