; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc -O0 -mtriple=x86_64-unknown -mcpu=skx -o - %s | FileCheck %s --check-prefix=X64-O0
; RUN: llc -mtriple=x86_64-unknown -mcpu=skx -o - %s | FileCheck %s --check-prefix=X64
; RUN: llc -O0 -mtriple=i686-unknown -mcpu=skx -o - %s | FileCheck %s --check-prefix=X86-O0
; RUN: llc -mtriple=i686-unknown -mcpu=skx -o - %s | FileCheck %s --check-prefix=X86
@c = external dso_local constant i8, align 1
define void @foo() {
; X64-O0-LABEL: foo:
; X64-O0: # %bb.0: # %entry
; X64-O0-NEXT: movzbl c, %ecx
; X64-O0-NEXT: xorl %eax, %eax
; X64-O0-NEXT: subl %ecx, %eax
; X64-O0-NEXT: movslq %eax, %rcx
; X64-O0-NEXT: xorl %eax, %eax
; X64-O0-NEXT: # kill: def $rax killed $eax
; X64-O0-NEXT: subq %rcx, %rax
; X64-O0-NEXT: # kill: def $al killed $al killed $rax
; X64-O0-NEXT: cmpb $0, %al
; X64-O0-NEXT: setne %al
; X64-O0-NEXT: andb $1, %al
; X64-O0-NEXT: movb %al, -{{[0-9]+}}(%rsp)
; X64-O0-NEXT: cmpb $0, c
; X64-O0-NEXT: setne %al
; X64-O0-NEXT: xorb $-1, %al
; X64-O0-NEXT: xorb $-1, %al
; X64-O0-NEXT: andb $1, %al
; X64-O0-NEXT: movzbl %al, %eax
; X64-O0-NEXT: movzbl c, %ecx
; X64-O0-NEXT: cmpl %ecx, %eax
; X64-O0-NEXT: setle %al
; X64-O0-NEXT: andb $1, %al
; X64-O0-NEXT: movzbl %al, %eax
; X64-O0-NEXT: movl %eax, -{{[0-9]+}}(%rsp)
; X64-O0-NEXT: retq
;
; X64-LABEL: foo:
; X64: # %bb.0: # %entry
; X64-NEXT: movzbl c(%rip), %eax
; X64-NEXT: xorl %ecx, %ecx
; X64-NEXT: testl %eax, %eax
; X64-NEXT: setne -{{[0-9]+}}(%rsp)
; X64-NEXT: setne %cl
; X64-NEXT: xorl %edx, %edx
; X64-NEXT: cmpl %eax, %ecx
; X64-NEXT: setle %dl
; X64-NEXT: movl %edx, -{{[0-9]+}}(%rsp)
; X64-NEXT: retq
;
; X86-O0-LABEL: foo:
; X86-O0: # %bb.0: # %entry
; X86-O0-NEXT: subl $8, %esp
; X86-O0-NEXT: .cfi_def_cfa_offset 12
; X86-O0-NEXT: movb c, %al
; X86-O0-NEXT: cmpb $0, %al
; X86-O0-NEXT: setne %al
; X86-O0-NEXT: andb $1, %al
; X86-O0-NEXT: movb %al, {{[0-9]+}}(%esp)
; X86-O0-NEXT: cmpb $0, c
; X86-O0-NEXT: setne %al
; X86-O0-NEXT: xorb $-1, %al
; X86-O0-NEXT: xorb $-1, %al
; X86-O0-NEXT: andb $1, %al
; X86-O0-NEXT: movzbl %al, %eax
; X86-O0-NEXT: movzbl c, %ecx
; X86-O0-NEXT: cmpl %ecx, %eax
; X86-O0-NEXT: setle %al
; X86-O0-NEXT: andb $1, %al
; X86-O0-NEXT: movzbl %al, %eax
; X86-O0-NEXT: movl %eax, (%esp)
; X86-O0-NEXT: addl $8, %esp
; X86-O0-NEXT: .cfi_def_cfa_offset 4
; X86-O0-NEXT: retl
;
; X86-LABEL: foo:
; X86: # %bb.0: # %entry
; X86-NEXT: subl $8, %esp
; X86-NEXT: .cfi_def_cfa_offset 12
; X86-NEXT: movzbl c, %eax
; X86-NEXT: xorl %ecx, %ecx
; X86-NEXT: testl %eax, %eax
; X86-NEXT: setne {{[0-9]+}}(%esp)
; X86-NEXT: setne %cl
; X86-NEXT: xorl %edx, %edx
; X86-NEXT: cmpl %eax, %ecx
; X86-NEXT: setle %dl
; X86-NEXT: movl %edx, {{[0-9]+}}(%esp)
; X86-NEXT: addl $8, %esp
; X86-NEXT: .cfi_def_cfa_offset 4
; X86-NEXT: retl
entry:
%a = alloca i8, align 1
%b = alloca i32, align 4
%0 = load i8, ptr @c, align 1
%conv = zext i8 %0 to i32
%sub = sub nsw i32 0, %conv
%conv1 = sext i32 %sub to i64
%sub2 = sub nsw i64 0, %conv1
%conv3 = trunc i64 %sub2 to i8
%tobool = icmp ne i8 %conv3, 0
%frombool = zext i1 %tobool to i8
store i8 %frombool, ptr %a, align 1
%1 = load i8, ptr @c, align 1
%tobool4 = icmp ne i8 %1, 0
%lnot = xor i1 %tobool4, true
%lnot5 = xor i1 %lnot, true
%conv6 = zext i1 %lnot5 to i32
%2 = load i8, ptr @c, align 1
%conv7 = zext i8 %2 to i32
%cmp = icmp sle i32 %conv6, %conv7
%conv8 = zext i1 %cmp to i32
store i32 %conv8, ptr %b, align 4
ret void
}
@var_5 = external dso_local global i32, align 4
@var_57 = external dso_local global i64, align 8
@_ZN8struct_210member_2_0E = external dso_local global i64, align 8
define void @f1() {
; X64-O0-LABEL: f1:
; X64-O0: # %bb.0: # %entry
; X64-O0-NEXT: movslq var_5, %rax
; X64-O0-NEXT: movabsq $8381627093, %rcx # imm = 0x1F3957AD5
; X64-O0-NEXT: addq %rcx, %rax
; X64-O0-NEXT: cmpq $0, %rax
; X64-O0-NEXT: setne %al
; X64-O0-NEXT: andb $1, %al
; X64-O0-NEXT: movb %al, -{{[0-9]+}}(%rsp)
; X64-O0-NEXT: movl var_5, %eax
; X64-O0-NEXT: xorl $-1, %eax
; X64-O0-NEXT: cmpl $0, %eax
; X64-O0-NEXT: setne %al
; X64-O0-NEXT: xorb $-1, %al
; X64-O0-NEXT: andb $1, %al
; X64-O0-NEXT: movzbl %al, %eax
; X64-O0-NEXT: # kill: def $rax killed $eax
; X64-O0-NEXT: movslq var_5, %rcx
; X64-O0-NEXT: addq $7093, %rcx # imm = 0x1BB5
; X64-O0-NEXT: cmpq %rcx, %rax
; X64-O0-NEXT: setg %al
; X64-O0-NEXT: andb $1, %al
; X64-O0-NEXT: movzbl %al, %eax
; X64-O0-NEXT: # kill: def $rax killed $eax
; X64-O0-NEXT: movq %rax, var_57
; X64-O0-NEXT: movl var_5, %eax
; X64-O0-NEXT: xorl $-1, %eax
; X64-O0-NEXT: cmpl $0, %eax
; X64-O0-NEXT: setne %al
; X64-O0-NEXT: xorb $-1, %al
; X64-O0-NEXT: andb $1, %al
; X64-O0-NEXT: movzbl %al, %eax
; X64-O0-NEXT: # kill: def $rax killed $eax
; X64-O0-NEXT: movq %rax, _ZN8struct_210member_2_0E
; X64-O0-NEXT: retq
;
; X64-LABEL: f1:
; X64: # %bb.0: # %entry
; X64-NEXT: movslq var_5(%rip), %rax
; X64-NEXT: movabsq $-8381627093, %rcx # imm = 0xFFFFFFFE0C6A852B
; X64-NEXT: cmpq %rcx, %rax
; X64-NEXT: setne -{{[0-9]+}}(%rsp)
; X64-NEXT: xorl %ecx, %ecx
; X64-NEXT: cmpq $-1, %rax
; X64-NEXT: sete %cl
; X64-NEXT: xorl %edx, %edx
; X64-NEXT: cmpl $-1, %eax
; X64-NEXT: sete %dl
; X64-NEXT: addq $7093, %rax # imm = 0x1BB5
; X64-NEXT: xorl %esi, %esi
; X64-NEXT: cmpq %rax, %rdx
; X64-NEXT: setg %sil
; X64-NEXT: movq %rsi, var_57(%rip)
; X64-NEXT: movq %rcx, _ZN8struct_210member_2_0E(%rip)
; X64-NEXT: retq
;
; X86-O0-LABEL: f1:
; X86-O0: # %bb.0: # %entry
; X86-O0-NEXT: subl $1, %esp
; X86-O0-NEXT: .cfi_def_cfa_offset 5
; X86-O0-NEXT: movl var_5, %eax
; X86-O0-NEXT: movl %eax, %ecx
; X86-O0-NEXT: sarl $31, %ecx
; X86-O0-NEXT: xorl $208307499, %eax # imm = 0xC6A852B
; X86-O0-NEXT: xorl $-2, %ecx
; X86-O0-NEXT: orl %ecx, %eax
; X86-O0-NEXT: setne (%esp)
; X86-O0-NEXT: movl var_5, %ecx
; X86-O0-NEXT: movl %ecx, %eax
; X86-O0-NEXT: sarl $31, %eax
; X86-O0-NEXT: movl %ecx, %edx
; X86-O0-NEXT: subl $-1, %edx
; X86-O0-NEXT: sete %dl
; X86-O0-NEXT: movzbl %dl, %edx
; X86-O0-NEXT: addl $7093, %ecx # imm = 0x1BB5
; X86-O0-NEXT: adcl $0, %eax
; X86-O0-NEXT: subl %edx, %ecx
; X86-O0-NEXT: sbbl $0, %eax
; X86-O0-NEXT: setl %al
; X86-O0-NEXT: movzbl %al, %eax
; X86-O0-NEXT: movl %eax, var_57
; X86-O0-NEXT: movl $0, var_57+4
; X86-O0-NEXT: movl var_5, %eax
; X86-O0-NEXT: subl $-1, %eax
; X86-O0-NEXT: sete %al
; X86-O0-NEXT: movzbl %al, %eax
; X86-O0-NEXT: movl %eax, _ZN8struct_210member_2_0E
; X86-O0-NEXT: movl $0, _ZN8struct_210member_2_0E+4
; X86-O0-NEXT: addl $1, %esp
; X86-O0-NEXT: .cfi_def_cfa_offset 4
; X86-O0-NEXT: retl
;
; X86-LABEL: f1:
; X86: # %bb.0: # %entry
; X86-NEXT: pushl %ebx
; X86-NEXT: .cfi_def_cfa_offset 8
; X86-NEXT: pushl %esi
; X86-NEXT: .cfi_def_cfa_offset 12
; X86-NEXT: subl $1, %esp
; X86-NEXT: .cfi_def_cfa_offset 13
; X86-NEXT: .cfi_offset %esi, -12
; X86-NEXT: .cfi_offset %ebx, -8
; X86-NEXT: movl var_5, %eax
; X86-NEXT: movl %eax, %edx
; X86-NEXT: xorl $208307499, %edx # imm = 0xC6A852B
; X86-NEXT: movl %eax, %ecx
; X86-NEXT: sarl $31, %ecx
; X86-NEXT: movl %ecx, %esi
; X86-NEXT: xorl $-2, %esi
; X86-NEXT: orl %edx, %esi
; X86-NEXT: setne (%esp)
; X86-NEXT: movl %eax, %esi
; X86-NEXT: andl %ecx, %esi
; X86-NEXT: xorl %edx, %edx
; X86-NEXT: cmpl $-1, %esi
; X86-NEXT: sete %dl
; X86-NEXT: xorl %ebx, %ebx
; X86-NEXT: cmpl $-1, %eax
; X86-NEXT: sete %bl
; X86-NEXT: addl $7093, %eax # imm = 0x1BB5
; X86-NEXT: adcl $0, %ecx
; X86-NEXT: cmpl %ebx, %eax
; X86-NEXT: sbbl $0, %ecx
; X86-NEXT: setl %al
; X86-NEXT: movzbl %al, %eax
; X86-NEXT: movl %eax, var_57
; X86-NEXT: movl $0, var_57+4
; X86-NEXT: movl %edx, _ZN8struct_210member_2_0E
; X86-NEXT: movl $0, _ZN8struct_210member_2_0E+4
; X86-NEXT: addl $1, %esp
; X86-NEXT: .cfi_def_cfa_offset 12
; X86-NEXT: popl %esi
; X86-NEXT: .cfi_def_cfa_offset 8
; X86-NEXT: popl %ebx
; X86-NEXT: .cfi_def_cfa_offset 4
; X86-NEXT: retl
entry:
%a = alloca i8, align 1
%0 = load i32, ptr @var_5, align 4
%conv = sext i32 %0 to i64
%add = add nsw i64 %conv, 8381627093
%tobool = icmp ne i64 %add, 0
%frombool = zext i1 %tobool to i8
store i8 %frombool, ptr %a, align 1
%1 = load i32, ptr @var_5, align 4
%neg = xor i32 %1, -1
%tobool1 = icmp ne i32 %neg, 0
%lnot = xor i1 %tobool1, true
%conv2 = zext i1 %lnot to i64
%2 = load i32, ptr @var_5, align 4
%conv3 = sext i32 %2 to i64
%add4 = add nsw i64 %conv3, 7093
%cmp = icmp sgt i64 %conv2, %add4
%conv5 = zext i1 %cmp to i64
store i64 %conv5, ptr @var_57, align 8
%3 = load i32, ptr @var_5, align 4
%neg6 = xor i32 %3, -1
%tobool7 = icmp ne i32 %neg6, 0
%lnot8 = xor i1 %tobool7, true
%conv9 = zext i1 %lnot8 to i64
store i64 %conv9, ptr @_ZN8struct_210member_2_0E, align 8
ret void
}
@var_7 = external dso_local global i8, align 1
define void @f2() {
; X64-O0-LABEL: f2:
; X64-O0: # %bb.0: # %entry
; X64-O0-NEXT: movzbl var_7, %eax
; X64-O0-NEXT: cmpb $0, var_7
; X64-O0-NEXT: setne %cl
; X64-O0-NEXT: xorb $-1, %cl
; X64-O0-NEXT: andb $1, %cl
; X64-O0-NEXT: movzbl %cl, %ecx
; X64-O0-NEXT: xorl %ecx, %eax
; X64-O0-NEXT: # kill: def $ax killed $ax killed $eax
; X64-O0-NEXT: movw %ax, -{{[0-9]+}}(%rsp)
; X64-O0-NEXT: movzbl var_7, %eax
; X64-O0-NEXT: # kill: def $ax killed $ax killed $eax
; X64-O0-NEXT: cmpw $0, %ax
; X64-O0-NEXT: setne %al
; X64-O0-NEXT: xorb $-1, %al
; X64-O0-NEXT: andb $1, %al
; X64-O0-NEXT: movzbl %al, %eax
; X64-O0-NEXT: movzbl var_7, %ecx
; X64-O0-NEXT: cmpl %ecx, %eax
; X64-O0-NEXT: sete %al
; X64-O0-NEXT: andb $1, %al
; X64-O0-NEXT: movzbl %al, %eax
; X64-O0-NEXT: movw %ax, %cx
; X64-O0-NEXT: # implicit-def: $rax
; X64-O0-NEXT: movw %cx, (%rax)
; X64-O0-NEXT: retq
;
; X64-LABEL: f2:
; X64: # %bb.0: # %entry
; X64-NEXT: movzbl var_7(%rip), %eax
; X64-NEXT: xorl %ecx, %ecx
; X64-NEXT: testl %eax, %eax
; X64-NEXT: sete %cl
; X64-NEXT: movl %eax, %edx
; X64-NEXT: xorl %ecx, %edx
; X64-NEXT: movw %dx, -{{[0-9]+}}(%rsp)
; X64-NEXT: xorl %edx, %edx
; X64-NEXT: cmpl %eax, %ecx
; X64-NEXT: sete %dl
; X64-NEXT: movw %dx, (%rax)
; X64-NEXT: retq
;
; X86-O0-LABEL: f2:
; X86-O0: # %bb.0: # %entry
; X86-O0-NEXT: subl $2, %esp
; X86-O0-NEXT: .cfi_def_cfa_offset 6
; X86-O0-NEXT: movzbl var_7, %eax
; X86-O0-NEXT: cmpb $0, var_7
; X86-O0-NEXT: setne %cl
; X86-O0-NEXT: xorb $-1, %cl
; X86-O0-NEXT: andb $1, %cl
; X86-O0-NEXT: movzbl %cl, %ecx
; X86-O0-NEXT: xorl %ecx, %eax
; X86-O0-NEXT: # kill: def $ax killed $ax killed $eax
; X86-O0-NEXT: movw %ax, (%esp)
; X86-O0-NEXT: movzbl var_7, %eax
; X86-O0-NEXT: # kill: def $ax killed $ax killed $eax
; X86-O0-NEXT: cmpw $0, %ax
; X86-O0-NEXT: setne %al
; X86-O0-NEXT: xorb $-1, %al
; X86-O0-NEXT: andb $1, %al
; X86-O0-NEXT: movzbl %al, %eax
; X86-O0-NEXT: movzbl var_7, %ecx
; X86-O0-NEXT: cmpl %ecx, %eax
; X86-O0-NEXT: sete %al
; X86-O0-NEXT: andb $1, %al
; X86-O0-NEXT: movzbl %al, %eax
; X86-O0-NEXT: movw %ax, %cx
; X86-O0-NEXT: # implicit-def: $eax
; X86-O0-NEXT: movw %cx, (%eax)
; X86-O0-NEXT: addl $2, %esp
; X86-O0-NEXT: .cfi_def_cfa_offset 4
; X86-O0-NEXT: retl
;
; X86-LABEL: f2:
; X86: # %bb.0: # %entry
; X86-NEXT: subl $2, %esp
; X86-NEXT: .cfi_def_cfa_offset 6
; X86-NEXT: movzbl var_7, %ecx
; X86-NEXT: xorl %eax, %eax
; X86-NEXT: testl %ecx, %ecx
; X86-NEXT: sete %al
; X86-NEXT: movl %ecx, %edx
; X86-NEXT: xorl %eax, %edx
; X86-NEXT: movw %dx, (%esp)
; X86-NEXT: xorl %edx, %edx
; X86-NEXT: cmpl %ecx, %eax
; X86-NEXT: sete %dl
; X86-NEXT: movw %dx, (%eax)
; X86-NEXT: addl $2, %esp
; X86-NEXT: .cfi_def_cfa_offset 4
; X86-NEXT: retl
entry:
%a = alloca i16, align 2
%0 = load i8, ptr @var_7, align 1
%conv = zext i8 %0 to i32
%1 = load i8, ptr @var_7, align 1
%tobool = icmp ne i8 %1, 0
%lnot = xor i1 %tobool, true
%conv1 = zext i1 %lnot to i32
%xor = xor i32 %conv, %conv1
%conv2 = trunc i32 %xor to i16
store i16 %conv2, ptr %a, align 2
%2 = load i8, ptr @var_7, align 1
%conv3 = zext i8 %2 to i16
%tobool4 = icmp ne i16 %conv3, 0
%lnot5 = xor i1 %tobool4, true
%conv6 = zext i1 %lnot5 to i32
%3 = load i8, ptr @var_7, align 1
%conv7 = zext i8 %3 to i32
%cmp = icmp eq i32 %conv6, %conv7
%conv8 = zext i1 %cmp to i32
%conv9 = trunc i32 %conv8 to i16
store i16 %conv9, ptr undef, align 2
ret void
}
@var_13 = external dso_local global i32, align 4
@var_16 = external dso_local global i32, align 4
@var_46 = external dso_local global i32, align 4
define void @f3() #0 {
; X64-O0-LABEL: f3:
; X64-O0: # %bb.0: # %entry
; X64-O0-NEXT: movl var_13, %eax
; X64-O0-NEXT: xorl $-1, %eax
; X64-O0-NEXT: movl %eax, %eax
; X64-O0-NEXT: # kill: def $rax killed $eax
; X64-O0-NEXT: cmpl $0, var_13
; X64-O0-NEXT: setne %cl
; X64-O0-NEXT: xorb $-1, %cl
; X64-O0-NEXT: andb $1, %cl
; X64-O0-NEXT: movzbl %cl, %ecx
; X64-O0-NEXT: # kill: def $rcx killed $ecx
; X64-O0-NEXT: movl var_13, %edx
; X64-O0-NEXT: xorl $-1, %edx
; X64-O0-NEXT: xorl var_16, %edx
; X64-O0-NEXT: movl %edx, %edx
; X64-O0-NEXT: # kill: def $rdx killed $edx
; X64-O0-NEXT: andq %rdx, %rcx
; X64-O0-NEXT: orq %rcx, %rax
; X64-O0-NEXT: movq %rax, -{{[0-9]+}}(%rsp)
; X64-O0-NEXT: movl var_13, %eax
; X64-O0-NEXT: xorl $-1, %eax
; X64-O0-NEXT: movl %eax, %eax
; X64-O0-NEXT: # kill: def $rax killed $eax
; X64-O0-NEXT: cmpl $0, var_13
; X64-O0-NEXT: setne %cl
; X64-O0-NEXT: xorb $-1, %cl
; X64-O0-NEXT: andb $1, %cl
; X64-O0-NEXT: movzbl %cl, %ecx
; X64-O0-NEXT: # kill: def $rcx killed $ecx
; X64-O0-NEXT: andq $0, %rcx
; X64-O0-NEXT: orq %rcx, %rax
; X64-O0-NEXT: # kill: def $eax killed $eax killed $rax
; X64-O0-NEXT: movl %eax, var_46
; X64-O0-NEXT: retq
;
; X64-LABEL: f3:
; X64: # %bb.0: # %entry
; X64-NEXT: movl var_13(%rip), %eax
; X64-NEXT: xorl %ecx, %ecx
; X64-NEXT: testl %eax, %eax
; X64-NEXT: notl %eax
; X64-NEXT: sete %cl
; X64-NEXT: movl var_16(%rip), %edx
; X64-NEXT: xorl %eax, %edx
; X64-NEXT: andl %edx, %ecx
; X64-NEXT: orl %eax, %ecx
; X64-NEXT: movq %rcx, -{{[0-9]+}}(%rsp)
; X64-NEXT: movl %eax, var_46(%rip)
; X64-NEXT: retq
;
; X86-O0-LABEL: f3:
; X86-O0: # %bb.0: # %entry
; X86-O0-NEXT: pushl %ebp
; X86-O0-NEXT: .cfi_def_cfa_offset 8
; X86-O0-NEXT: .cfi_offset %ebp, -8
; X86-O0-NEXT: movl %esp, %ebp
; X86-O0-NEXT: .cfi_def_cfa_register %ebp
; X86-O0-NEXT: pushl %esi
; X86-O0-NEXT: andl $-8, %esp
; X86-O0-NEXT: subl $16, %esp
; X86-O0-NEXT: .cfi_offset %esi, -12
; X86-O0-NEXT: movl var_13, %ecx
; X86-O0-NEXT: movl %ecx, %eax
; X86-O0-NEXT: notl %eax
; X86-O0-NEXT: testl %ecx, %ecx
; X86-O0-NEXT: sete %cl
; X86-O0-NEXT: movzbl %cl, %ecx
; X86-O0-NEXT: movl var_16, %esi
; X86-O0-NEXT: movl %eax, %edx
; X86-O0-NEXT: xorl %esi, %edx
; X86-O0-NEXT: andl %edx, %ecx
; X86-O0-NEXT: orl %ecx, %eax
; X86-O0-NEXT: movl %eax, (%esp)
; X86-O0-NEXT: movl $0, {{[0-9]+}}(%esp)
; X86-O0-NEXT: movl var_13, %eax
; X86-O0-NEXT: notl %eax
; X86-O0-NEXT: movl %eax, var_46
; X86-O0-NEXT: leal -4(%ebp), %esp
; X86-O0-NEXT: popl %esi
; X86-O0-NEXT: popl %ebp
; X86-O0-NEXT: .cfi_def_cfa %esp, 4
; X86-O0-NEXT: retl
;
; X86-LABEL: f3:
; X86: # %bb.0: # %entry
; X86-NEXT: pushl %ebp
; X86-NEXT: .cfi_def_cfa_offset 8
; X86-NEXT: .cfi_offset %ebp, -8
; X86-NEXT: movl %esp, %ebp
; X86-NEXT: .cfi_def_cfa_register %ebp
; X86-NEXT: andl $-8, %esp
; X86-NEXT: subl $8, %esp
; X86-NEXT: movl var_13, %eax
; X86-NEXT: xorl %ecx, %ecx
; X86-NEXT: testl %eax, %eax
; X86-NEXT: notl %eax
; X86-NEXT: sete %cl
; X86-NEXT: movl var_16, %edx
; X86-NEXT: xorl %eax, %edx
; X86-NEXT: andl %ecx, %edx
; X86-NEXT: orl %eax, %edx
; X86-NEXT: movl %edx, (%esp)
; X86-NEXT: movl $0, {{[0-9]+}}(%esp)
; X86-NEXT: movl %eax, var_46
; X86-NEXT: movl %ebp, %esp
; X86-NEXT: popl %ebp
; X86-NEXT: .cfi_def_cfa %esp, 4
; X86-NEXT: retl
entry:
%a = alloca i64, align 8
%0 = load i32, ptr @var_13, align 4
%neg = xor i32 %0, -1
%conv = zext i32 %neg to i64
%1 = load i32, ptr @var_13, align 4
%tobool = icmp ne i32 %1, 0
%lnot = xor i1 %tobool, true
%conv1 = zext i1 %lnot to i64
%2 = load i32, ptr @var_13, align 4
%neg2 = xor i32 %2, -1
%3 = load i32, ptr @var_16, align 4
%xor = xor i32 %neg2, %3
%conv3 = zext i32 %xor to i64
%and = and i64 %conv1, %conv3
%or = or i64 %conv, %and
store i64 %or, ptr %a, align 8
%4 = load i32, ptr @var_13, align 4
%neg4 = xor i32 %4, -1
%conv5 = zext i32 %neg4 to i64
%5 = load i32, ptr @var_13, align 4
%tobool6 = icmp ne i32 %5, 0
%lnot7 = xor i1 %tobool6, true
%conv8 = zext i1 %lnot7 to i64
%and9 = and i64 %conv8, 0
%or10 = or i64 %conv5, %and9
%conv11 = trunc i64 %or10 to i32
store i32 %conv11, ptr @var_46, align 4
ret void
}