llvm/llvm/test/CodeGen/X86/vec_shift4.ll

; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc < %s -mtriple=i686-unknown -mattr=+sse4.1 | FileCheck %s --check-prefix=X86
; RUN: llc < %s -mtriple=x86_64-unknown -mattr=+sse4.1 | FileCheck %s --check-prefix=X64

define <2 x i64> @shl1(<4 x i32> %r, <4 x i32> %a) nounwind readnone ssp {
; X86-LABEL: shl1:
; X86:       # %bb.0: # %entry
; X86-NEXT:    pslld $23, %xmm1
; X86-NEXT:    paddd {{\.?LCPI[0-9]+_[0-9]+}}, %xmm1
; X86-NEXT:    cvttps2dq %xmm1, %xmm1
; X86-NEXT:    pmulld %xmm1, %xmm0
; X86-NEXT:    retl
;
; X64-LABEL: shl1:
; X64:       # %bb.0: # %entry
; X64-NEXT:    pslld $23, %xmm1
; X64-NEXT:    paddd {{\.?LCPI[0-9]+_[0-9]+}}(%rip), %xmm1
; X64-NEXT:    cvttps2dq %xmm1, %xmm1
; X64-NEXT:    pmulld %xmm1, %xmm0
; X64-NEXT:    retq
entry:
  %shl = shl <4 x i32> %r, %a                     ; <<4 x i32>> [#uses=1]
  %tmp2 = bitcast <4 x i32> %shl to <2 x i64>     ; <<2 x i64>> [#uses=1]
  ret <2 x i64> %tmp2
}

define <2 x i64> @shl2(<16 x i8> %r, <16 x i8> %a) nounwind readnone ssp {
; X86-LABEL: shl2:
; X86:       # %bb.0: # %entry
; X86-NEXT:    movdqa %xmm1, %xmm2
; X86-NEXT:    movdqa %xmm0, %xmm1
; X86-NEXT:    psllw $5, %xmm2
; X86-NEXT:    movdqa %xmm0, %xmm3
; X86-NEXT:    psllw $4, %xmm3
; X86-NEXT:    pand {{\.?LCPI[0-9]+_[0-9]+}}, %xmm3
; X86-NEXT:    movdqa %xmm2, %xmm0
; X86-NEXT:    pblendvb %xmm0, %xmm3, %xmm1
; X86-NEXT:    movdqa %xmm1, %xmm3
; X86-NEXT:    psllw $2, %xmm3
; X86-NEXT:    pand {{\.?LCPI[0-9]+_[0-9]+}}, %xmm3
; X86-NEXT:    paddb %xmm2, %xmm2
; X86-NEXT:    movdqa %xmm2, %xmm0
; X86-NEXT:    pblendvb %xmm0, %xmm3, %xmm1
; X86-NEXT:    movdqa %xmm1, %xmm3
; X86-NEXT:    paddb %xmm1, %xmm3
; X86-NEXT:    paddb %xmm2, %xmm2
; X86-NEXT:    movdqa %xmm2, %xmm0
; X86-NEXT:    pblendvb %xmm0, %xmm3, %xmm1
; X86-NEXT:    movdqa %xmm1, %xmm0
; X86-NEXT:    retl
;
; X64-LABEL: shl2:
; X64:       # %bb.0: # %entry
; X64-NEXT:    movdqa %xmm0, %xmm2
; X64-NEXT:    psllw $5, %xmm1
; X64-NEXT:    movdqa %xmm0, %xmm3
; X64-NEXT:    psllw $4, %xmm3
; X64-NEXT:    pand {{\.?LCPI[0-9]+_[0-9]+}}(%rip), %xmm3
; X64-NEXT:    movdqa %xmm1, %xmm0
; X64-NEXT:    pblendvb %xmm0, %xmm3, %xmm2
; X64-NEXT:    movdqa %xmm2, %xmm3
; X64-NEXT:    psllw $2, %xmm3
; X64-NEXT:    pand {{\.?LCPI[0-9]+_[0-9]+}}(%rip), %xmm3
; X64-NEXT:    paddb %xmm1, %xmm1
; X64-NEXT:    movdqa %xmm1, %xmm0
; X64-NEXT:    pblendvb %xmm0, %xmm3, %xmm2
; X64-NEXT:    movdqa %xmm2, %xmm3
; X64-NEXT:    paddb %xmm2, %xmm3
; X64-NEXT:    paddb %xmm1, %xmm1
; X64-NEXT:    movdqa %xmm1, %xmm0
; X64-NEXT:    pblendvb %xmm0, %xmm3, %xmm2
; X64-NEXT:    movdqa %xmm2, %xmm0
; X64-NEXT:    retq
entry:
  %shl = shl <16 x i8> %r, %a                     ; <<16 x i8>> [#uses=1]
  %tmp2 = bitcast <16 x i8> %shl to <2 x i64>     ; <<2 x i64>> [#uses=1]
  ret <2 x i64> %tmp2
}