; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc < %s -mtriple=i686-- | FileCheck %s --check-prefixes=X86
; RUN: llc < %s -mtriple=x86_64-- | FileCheck %s --check-prefixes=X64
;
; fixed avg(x,y) = sub(or(x,y),lshr(xor(x,y),1))
;
; ext avg(x,y) = trunc(lshr(add(zext(x),zext(y),1),1))
;
define i8 @test_fixed_i8(i8 %a0, i8 %a1) nounwind {
; X86-LABEL: test_fixed_i8:
; X86: # %bb.0:
; X86-NEXT: movzbl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movzbl {{[0-9]+}}(%esp), %ecx
; X86-NEXT: leal 1(%ecx,%eax), %eax
; X86-NEXT: shrl %eax
; X86-NEXT: # kill: def $al killed $al killed $eax
; X86-NEXT: retl
;
; X64-LABEL: test_fixed_i8:
; X64: # %bb.0:
; X64-NEXT: movzbl %sil, %eax
; X64-NEXT: movzbl %dil, %ecx
; X64-NEXT: leal 1(%rcx,%rax), %eax
; X64-NEXT: shrl %eax
; X64-NEXT: # kill: def $al killed $al killed $eax
; X64-NEXT: retq
%or = or i8 %a0, %a1
%xor = xor i8 %a0, %a1
%shift = lshr i8 %xor, 1
%res = sub i8 %or, %shift
ret i8 %res
}
define i8 @test_ext_i8(i8 %a0, i8 %a1) nounwind {
; X86-LABEL: test_ext_i8:
; X86: # %bb.0:
; X86-NEXT: movzbl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movzbl {{[0-9]+}}(%esp), %ecx
; X86-NEXT: leal 1(%ecx,%eax), %eax
; X86-NEXT: shrl %eax
; X86-NEXT: # kill: def $al killed $al killed $eax
; X86-NEXT: retl
;
; X64-LABEL: test_ext_i8:
; X64: # %bb.0:
; X64-NEXT: movzbl %sil, %eax
; X64-NEXT: movzbl %dil, %ecx
; X64-NEXT: leal 1(%rcx,%rax), %eax
; X64-NEXT: shrl %eax
; X64-NEXT: # kill: def $al killed $al killed $eax
; X64-NEXT: retq
%x0 = zext i8 %a0 to i16
%x1 = zext i8 %a1 to i16
%sum = add i16 %x0, %x1
%sum1 = add i16 %sum, 1
%shift = lshr i16 %sum1, 1
%res = trunc i16 %shift to i8
ret i8 %res
}
define i16 @test_fixed_i16(i16 %a0, i16 %a1) nounwind {
; X86-LABEL: test_fixed_i16:
; X86: # %bb.0:
; X86-NEXT: movzwl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movzwl {{[0-9]+}}(%esp), %ecx
; X86-NEXT: leal 1(%ecx,%eax), %eax
; X86-NEXT: shrl %eax
; X86-NEXT: # kill: def $ax killed $ax killed $eax
; X86-NEXT: retl
;
; X64-LABEL: test_fixed_i16:
; X64: # %bb.0:
; X64-NEXT: movzwl %si, %eax
; X64-NEXT: movzwl %di, %ecx
; X64-NEXT: leal 1(%rcx,%rax), %eax
; X64-NEXT: shrl %eax
; X64-NEXT: # kill: def $ax killed $ax killed $eax
; X64-NEXT: retq
%or = or i16 %a0, %a1
%xor = xor i16 %a0, %a1
%shift = lshr i16 %xor, 1
%res = sub i16 %or, %shift
ret i16 %res
}
define i16 @test_ext_i16(i16 %a0, i16 %a1) nounwind {
; X86-LABEL: test_ext_i16:
; X86: # %bb.0:
; X86-NEXT: movzwl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movzwl {{[0-9]+}}(%esp), %ecx
; X86-NEXT: leal 1(%ecx,%eax), %eax
; X86-NEXT: shrl %eax
; X86-NEXT: # kill: def $ax killed $ax killed $eax
; X86-NEXT: retl
;
; X64-LABEL: test_ext_i16:
; X64: # %bb.0:
; X64-NEXT: movzwl %si, %eax
; X64-NEXT: movzwl %di, %ecx
; X64-NEXT: leal 1(%rcx,%rax), %eax
; X64-NEXT: shrl %eax
; X64-NEXT: # kill: def $ax killed $ax killed $eax
; X64-NEXT: retq
%x0 = zext i16 %a0 to i32
%x1 = zext i16 %a1 to i32
%sum = add i32 %x0, %x1
%sum1 = add i32 %sum, 1
%shift = lshr i32 %sum1, 1
%res = trunc i32 %shift to i16
ret i16 %res
}
define i32 @test_fixed_i32(i32 %a0, i32 %a1) nounwind {
; X86-LABEL: test_fixed_i32:
; X86: # %bb.0:
; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx
; X86-NEXT: movl {{[0-9]+}}(%esp), %edx
; X86-NEXT: movl %edx, %eax
; X86-NEXT: orl %ecx, %eax
; X86-NEXT: xorl %ecx, %edx
; X86-NEXT: shrl %edx
; X86-NEXT: subl %edx, %eax
; X86-NEXT: retl
;
; X64-LABEL: test_fixed_i32:
; X64: # %bb.0:
; X64-NEXT: movl %esi, %eax
; X64-NEXT: movl %edi, %ecx
; X64-NEXT: leaq 1(%rcx,%rax), %rax
; X64-NEXT: shrq %rax
; X64-NEXT: # kill: def $eax killed $eax killed $rax
; X64-NEXT: retq
%or = or i32 %a0, %a1
%xor = xor i32 %a1, %a0
%shift = lshr i32 %xor, 1
%res = sub i32 %or, %shift
ret i32 %res
}
define i32 @test_ext_i32(i32 %a0, i32 %a1) nounwind {
; X86-LABEL: test_ext_i32:
; X86: # %bb.0:
; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx
; X86-NEXT: movl {{[0-9]+}}(%esp), %edx
; X86-NEXT: movl %edx, %eax
; X86-NEXT: orl %ecx, %eax
; X86-NEXT: xorl %ecx, %edx
; X86-NEXT: shrl %edx
; X86-NEXT: subl %edx, %eax
; X86-NEXT: retl
;
; X64-LABEL: test_ext_i32:
; X64: # %bb.0:
; X64-NEXT: movl %esi, %eax
; X64-NEXT: movl %edi, %ecx
; X64-NEXT: leaq 1(%rcx,%rax), %rax
; X64-NEXT: shrq %rax
; X64-NEXT: # kill: def $eax killed $eax killed $rax
; X64-NEXT: retq
%x0 = zext i32 %a0 to i64
%x1 = zext i32 %a1 to i64
%sum = add i64 %x0, %x1
%sum1 = add i64 %sum, 1
%shift = lshr i64 %sum1, 1
%res = trunc i64 %shift to i32
ret i32 %res
}
define i64 @test_fixed_i64(i64 %a0, i64 %a1) nounwind {
; X86-LABEL: test_fixed_i64:
; X86: # %bb.0:
; X86-NEXT: pushl %ebx
; X86-NEXT: pushl %edi
; X86-NEXT: pushl %esi
; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx
; X86-NEXT: movl {{[0-9]+}}(%esp), %esi
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl {{[0-9]+}}(%esp), %edx
; X86-NEXT: movl %eax, %edi
; X86-NEXT: xorl %ecx, %edi
; X86-NEXT: movl %edx, %ebx
; X86-NEXT: xorl %esi, %ebx
; X86-NEXT: shrdl $1, %ebx, %edi
; X86-NEXT: orl %esi, %edx
; X86-NEXT: shrl %ebx
; X86-NEXT: orl %ecx, %eax
; X86-NEXT: subl %edi, %eax
; X86-NEXT: sbbl %ebx, %edx
; X86-NEXT: popl %esi
; X86-NEXT: popl %edi
; X86-NEXT: popl %ebx
; X86-NEXT: retl
;
; X64-LABEL: test_fixed_i64:
; X64: # %bb.0:
; X64-NEXT: movq %rdi, %rax
; X64-NEXT: orq %rsi, %rax
; X64-NEXT: xorq %rsi, %rdi
; X64-NEXT: shrq %rdi
; X64-NEXT: subq %rdi, %rax
; X64-NEXT: retq
%or = or i64 %a0, %a1
%xor = xor i64 %a1, %a0
%shift = lshr i64 %xor, 1
%res = sub i64 %or, %shift
ret i64 %res
}
define i64 @test_ext_i64(i64 %a0, i64 %a1) nounwind {
; X86-LABEL: test_ext_i64:
; X86: # %bb.0:
; X86-NEXT: pushl %ebx
; X86-NEXT: pushl %edi
; X86-NEXT: pushl %esi
; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx
; X86-NEXT: movl {{[0-9]+}}(%esp), %esi
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
; X86-NEXT: movl {{[0-9]+}}(%esp), %edx
; X86-NEXT: movl %eax, %edi
; X86-NEXT: xorl %ecx, %edi
; X86-NEXT: movl %edx, %ebx
; X86-NEXT: xorl %esi, %ebx
; X86-NEXT: shrdl $1, %ebx, %edi
; X86-NEXT: orl %esi, %edx
; X86-NEXT: shrl %ebx
; X86-NEXT: orl %ecx, %eax
; X86-NEXT: subl %edi, %eax
; X86-NEXT: sbbl %ebx, %edx
; X86-NEXT: popl %esi
; X86-NEXT: popl %edi
; X86-NEXT: popl %ebx
; X86-NEXT: retl
;
; X64-LABEL: test_ext_i64:
; X64: # %bb.0:
; X64-NEXT: movq %rdi, %rax
; X64-NEXT: orq %rsi, %rax
; X64-NEXT: xorq %rsi, %rdi
; X64-NEXT: shrq %rdi
; X64-NEXT: subq %rdi, %rax
; X64-NEXT: retq
%x0 = zext i64 %a0 to i128
%x1 = zext i64 %a1 to i128
%sum = add i128 %x0, %x1
%sum1 = add i128 %sum, 1
%shift = lshr i128 %sum1, 1
%res = trunc i128 %shift to i64
ret i64 %res
}