; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py ; RUN: llc -O0 -mtriple=x86_64-unknown -mcpu=skx -o - %s | FileCheck %s --check-prefix=X64-O0 ; RUN: llc -mtriple=x86_64-unknown -mcpu=skx -o - %s | FileCheck %s --check-prefix=X64 ; RUN: llc -O0 -mtriple=i686-unknown -mcpu=skx -o - %s | FileCheck %s --check-prefix=X86-O0 ; RUN: llc -mtriple=i686-unknown -mcpu=skx -o - %s | FileCheck %s --check-prefix=X86 @c = external dso_local constant i8, align 1 define void @foo() { ; X64-O0-LABEL: foo: ; X64-O0: # %bb.0: # %entry ; X64-O0-NEXT: movzbl c, %ecx ; X64-O0-NEXT: xorl %eax, %eax ; X64-O0-NEXT: subl %ecx, %eax ; X64-O0-NEXT: movslq %eax, %rcx ; X64-O0-NEXT: xorl %eax, %eax ; X64-O0-NEXT: # kill: def $rax killed $eax ; X64-O0-NEXT: subq %rcx, %rax ; X64-O0-NEXT: # kill: def $al killed $al killed $rax ; X64-O0-NEXT: cmpb $0, %al ; X64-O0-NEXT: setne %al ; X64-O0-NEXT: andb $1, %al ; X64-O0-NEXT: movb %al, -{{[0-9]+}}(%rsp) ; X64-O0-NEXT: cmpb $0, c ; X64-O0-NEXT: setne %al ; X64-O0-NEXT: xorb $-1, %al ; X64-O0-NEXT: xorb $-1, %al ; X64-O0-NEXT: andb $1, %al ; X64-O0-NEXT: movzbl %al, %eax ; X64-O0-NEXT: movzbl c, %ecx ; X64-O0-NEXT: cmpl %ecx, %eax ; X64-O0-NEXT: setle %al ; X64-O0-NEXT: andb $1, %al ; X64-O0-NEXT: movzbl %al, %eax ; X64-O0-NEXT: movl %eax, -{{[0-9]+}}(%rsp) ; X64-O0-NEXT: retq ; ; X64-LABEL: foo: ; X64: # %bb.0: # %entry ; X64-NEXT: movzbl c(%rip), %eax ; X64-NEXT: xorl %ecx, %ecx ; X64-NEXT: testl %eax, %eax ; X64-NEXT: setne -{{[0-9]+}}(%rsp) ; X64-NEXT: setne %cl ; X64-NEXT: xorl %edx, %edx ; X64-NEXT: cmpl %eax, %ecx ; X64-NEXT: setle %dl ; X64-NEXT: movl %edx, -{{[0-9]+}}(%rsp) ; X64-NEXT: retq ; ; X86-O0-LABEL: foo: ; X86-O0: # %bb.0: # %entry ; X86-O0-NEXT: subl $8, %esp ; X86-O0-NEXT: .cfi_def_cfa_offset 12 ; X86-O0-NEXT: movb c, %al ; X86-O0-NEXT: cmpb $0, %al ; X86-O0-NEXT: setne %al ; X86-O0-NEXT: andb $1, %al ; X86-O0-NEXT: movb %al, {{[0-9]+}}(%esp) ; X86-O0-NEXT: cmpb $0, c ; X86-O0-NEXT: setne %al ; X86-O0-NEXT: xorb $-1, %al ; X86-O0-NEXT: xorb $-1, %al ; X86-O0-NEXT: andb $1, %al ; X86-O0-NEXT: movzbl %al, %eax ; X86-O0-NEXT: movzbl c, %ecx ; X86-O0-NEXT: cmpl %ecx, %eax ; X86-O0-NEXT: setle %al ; X86-O0-NEXT: andb $1, %al ; X86-O0-NEXT: movzbl %al, %eax ; X86-O0-NEXT: movl %eax, (%esp) ; X86-O0-NEXT: addl $8, %esp ; X86-O0-NEXT: .cfi_def_cfa_offset 4 ; X86-O0-NEXT: retl ; ; X86-LABEL: foo: ; X86: # %bb.0: # %entry ; X86-NEXT: subl $8, %esp ; X86-NEXT: .cfi_def_cfa_offset 12 ; X86-NEXT: movzbl c, %eax ; X86-NEXT: xorl %ecx, %ecx ; X86-NEXT: testl %eax, %eax ; X86-NEXT: setne {{[0-9]+}}(%esp) ; X86-NEXT: setne %cl ; X86-NEXT: xorl %edx, %edx ; X86-NEXT: cmpl %eax, %ecx ; X86-NEXT: setle %dl ; X86-NEXT: movl %edx, {{[0-9]+}}(%esp) ; X86-NEXT: addl $8, %esp ; X86-NEXT: .cfi_def_cfa_offset 4 ; X86-NEXT: retl entry: %a = alloca i8, align 1 %b = alloca i32, align 4 %0 = load i8, ptr @c, align 1 %conv = zext i8 %0 to i32 %sub = sub nsw i32 0, %conv %conv1 = sext i32 %sub to i64 %sub2 = sub nsw i64 0, %conv1 %conv3 = trunc i64 %sub2 to i8 %tobool = icmp ne i8 %conv3, 0 %frombool = zext i1 %tobool to i8 store i8 %frombool, ptr %a, align 1 %1 = load i8, ptr @c, align 1 %tobool4 = icmp ne i8 %1, 0 %lnot = xor i1 %tobool4, true %lnot5 = xor i1 %lnot, true %conv6 = zext i1 %lnot5 to i32 %2 = load i8, ptr @c, align 1 %conv7 = zext i8 %2 to i32 %cmp = icmp sle i32 %conv6, %conv7 %conv8 = zext i1 %cmp to i32 store i32 %conv8, ptr %b, align 4 ret void } @var_5 = external dso_local global i32, align 4 @var_57 = external dso_local global i64, align 8 @_ZN8struct_210member_2_0E = external dso_local global i64, align 8 define void @f1() { ; X64-O0-LABEL: f1: ; X64-O0: # %bb.0: # %entry ; X64-O0-NEXT: movslq var_5, %rax ; X64-O0-NEXT: movabsq $8381627093, %rcx # imm = 0x1F3957AD5 ; X64-O0-NEXT: addq %rcx, %rax ; X64-O0-NEXT: cmpq $0, %rax ; X64-O0-NEXT: setne %al ; X64-O0-NEXT: andb $1, %al ; X64-O0-NEXT: movb %al, -{{[0-9]+}}(%rsp) ; X64-O0-NEXT: movl var_5, %eax ; X64-O0-NEXT: xorl $-1, %eax ; X64-O0-NEXT: cmpl $0, %eax ; X64-O0-NEXT: setne %al ; X64-O0-NEXT: xorb $-1, %al ; X64-O0-NEXT: andb $1, %al ; X64-O0-NEXT: movzbl %al, %eax ; X64-O0-NEXT: # kill: def $rax killed $eax ; X64-O0-NEXT: movslq var_5, %rcx ; X64-O0-NEXT: addq $7093, %rcx # imm = 0x1BB5 ; X64-O0-NEXT: cmpq %rcx, %rax ; X64-O0-NEXT: setg %al ; X64-O0-NEXT: andb $1, %al ; X64-O0-NEXT: movzbl %al, %eax ; X64-O0-NEXT: # kill: def $rax killed $eax ; X64-O0-NEXT: movq %rax, var_57 ; X64-O0-NEXT: movl var_5, %eax ; X64-O0-NEXT: xorl $-1, %eax ; X64-O0-NEXT: cmpl $0, %eax ; X64-O0-NEXT: setne %al ; X64-O0-NEXT: xorb $-1, %al ; X64-O0-NEXT: andb $1, %al ; X64-O0-NEXT: movzbl %al, %eax ; X64-O0-NEXT: # kill: def $rax killed $eax ; X64-O0-NEXT: movq %rax, _ZN8struct_210member_2_0E ; X64-O0-NEXT: retq ; ; X64-LABEL: f1: ; X64: # %bb.0: # %entry ; X64-NEXT: movslq var_5(%rip), %rax ; X64-NEXT: movabsq $-8381627093, %rcx # imm = 0xFFFFFFFE0C6A852B ; X64-NEXT: cmpq %rcx, %rax ; X64-NEXT: setne -{{[0-9]+}}(%rsp) ; X64-NEXT: xorl %ecx, %ecx ; X64-NEXT: cmpq $-1, %rax ; X64-NEXT: sete %cl ; X64-NEXT: xorl %edx, %edx ; X64-NEXT: cmpl $-1, %eax ; X64-NEXT: sete %dl ; X64-NEXT: addq $7093, %rax # imm = 0x1BB5 ; X64-NEXT: xorl %esi, %esi ; X64-NEXT: cmpq %rax, %rdx ; X64-NEXT: setg %sil ; X64-NEXT: movq %rsi, var_57(%rip) ; X64-NEXT: movq %rcx, _ZN8struct_210member_2_0E(%rip) ; X64-NEXT: retq ; ; X86-O0-LABEL: f1: ; X86-O0: # %bb.0: # %entry ; X86-O0-NEXT: subl $1, %esp ; X86-O0-NEXT: .cfi_def_cfa_offset 5 ; X86-O0-NEXT: movl var_5, %eax ; X86-O0-NEXT: movl %eax, %ecx ; X86-O0-NEXT: sarl $31, %ecx ; X86-O0-NEXT: xorl $208307499, %eax # imm = 0xC6A852B ; X86-O0-NEXT: xorl $-2, %ecx ; X86-O0-NEXT: orl %ecx, %eax ; X86-O0-NEXT: setne (%esp) ; X86-O0-NEXT: movl var_5, %ecx ; X86-O0-NEXT: movl %ecx, %eax ; X86-O0-NEXT: sarl $31, %eax ; X86-O0-NEXT: movl %ecx, %edx ; X86-O0-NEXT: subl $-1, %edx ; X86-O0-NEXT: sete %dl ; X86-O0-NEXT: movzbl %dl, %edx ; X86-O0-NEXT: addl $7093, %ecx # imm = 0x1BB5 ; X86-O0-NEXT: adcl $0, %eax ; X86-O0-NEXT: subl %edx, %ecx ; X86-O0-NEXT: sbbl $0, %eax ; X86-O0-NEXT: setl %al ; X86-O0-NEXT: movzbl %al, %eax ; X86-O0-NEXT: movl %eax, var_57 ; X86-O0-NEXT: movl $0, var_57+4 ; X86-O0-NEXT: movl var_5, %eax ; X86-O0-NEXT: subl $-1, %eax ; X86-O0-NEXT: sete %al ; X86-O0-NEXT: movzbl %al, %eax ; X86-O0-NEXT: movl %eax, _ZN8struct_210member_2_0E ; X86-O0-NEXT: movl $0, _ZN8struct_210member_2_0E+4 ; X86-O0-NEXT: addl $1, %esp ; X86-O0-NEXT: .cfi_def_cfa_offset 4 ; X86-O0-NEXT: retl ; ; X86-LABEL: f1: ; X86: # %bb.0: # %entry ; X86-NEXT: pushl %ebx ; X86-NEXT: .cfi_def_cfa_offset 8 ; X86-NEXT: pushl %esi ; X86-NEXT: .cfi_def_cfa_offset 12 ; X86-NEXT: subl $1, %esp ; X86-NEXT: .cfi_def_cfa_offset 13 ; X86-NEXT: .cfi_offset %esi, -12 ; X86-NEXT: .cfi_offset %ebx, -8 ; X86-NEXT: movl var_5, %eax ; X86-NEXT: movl %eax, %edx ; X86-NEXT: xorl $208307499, %edx # imm = 0xC6A852B ; X86-NEXT: movl %eax, %ecx ; X86-NEXT: sarl $31, %ecx ; X86-NEXT: movl %ecx, %esi ; X86-NEXT: xorl $-2, %esi ; X86-NEXT: orl %edx, %esi ; X86-NEXT: setne (%esp) ; X86-NEXT: movl %eax, %esi ; X86-NEXT: andl %ecx, %esi ; X86-NEXT: xorl %edx, %edx ; X86-NEXT: cmpl $-1, %esi ; X86-NEXT: sete %dl ; X86-NEXT: xorl %ebx, %ebx ; X86-NEXT: cmpl $-1, %eax ; X86-NEXT: sete %bl ; X86-NEXT: addl $7093, %eax # imm = 0x1BB5 ; X86-NEXT: adcl $0, %ecx ; X86-NEXT: cmpl %ebx, %eax ; X86-NEXT: sbbl $0, %ecx ; X86-NEXT: setl %al ; X86-NEXT: movzbl %al, %eax ; X86-NEXT: movl %eax, var_57 ; X86-NEXT: movl $0, var_57+4 ; X86-NEXT: movl %edx, _ZN8struct_210member_2_0E ; X86-NEXT: movl $0, _ZN8struct_210member_2_0E+4 ; X86-NEXT: addl $1, %esp ; X86-NEXT: .cfi_def_cfa_offset 12 ; X86-NEXT: popl %esi ; X86-NEXT: .cfi_def_cfa_offset 8 ; X86-NEXT: popl %ebx ; X86-NEXT: .cfi_def_cfa_offset 4 ; X86-NEXT: retl entry: %a = alloca i8, align 1 %0 = load i32, ptr @var_5, align 4 %conv = sext i32 %0 to i64 %add = add nsw i64 %conv, 8381627093 %tobool = icmp ne i64 %add, 0 %frombool = zext i1 %tobool to i8 store i8 %frombool, ptr %a, align 1 %1 = load i32, ptr @var_5, align 4 %neg = xor i32 %1, -1 %tobool1 = icmp ne i32 %neg, 0 %lnot = xor i1 %tobool1, true %conv2 = zext i1 %lnot to i64 %2 = load i32, ptr @var_5, align 4 %conv3 = sext i32 %2 to i64 %add4 = add nsw i64 %conv3, 7093 %cmp = icmp sgt i64 %conv2, %add4 %conv5 = zext i1 %cmp to i64 store i64 %conv5, ptr @var_57, align 8 %3 = load i32, ptr @var_5, align 4 %neg6 = xor i32 %3, -1 %tobool7 = icmp ne i32 %neg6, 0 %lnot8 = xor i1 %tobool7, true %conv9 = zext i1 %lnot8 to i64 store i64 %conv9, ptr @_ZN8struct_210member_2_0E, align 8 ret void } @var_7 = external dso_local global i8, align 1 define void @f2() { ; X64-O0-LABEL: f2: ; X64-O0: # %bb.0: # %entry ; X64-O0-NEXT: movzbl var_7, %eax ; X64-O0-NEXT: cmpb $0, var_7 ; X64-O0-NEXT: setne %cl ; X64-O0-NEXT: xorb $-1, %cl ; X64-O0-NEXT: andb $1, %cl ; X64-O0-NEXT: movzbl %cl, %ecx ; X64-O0-NEXT: xorl %ecx, %eax ; X64-O0-NEXT: # kill: def $ax killed $ax killed $eax ; X64-O0-NEXT: movw %ax, -{{[0-9]+}}(%rsp) ; X64-O0-NEXT: movzbl var_7, %eax ; X64-O0-NEXT: # kill: def $ax killed $ax killed $eax ; X64-O0-NEXT: cmpw $0, %ax ; X64-O0-NEXT: setne %al ; X64-O0-NEXT: xorb $-1, %al ; X64-O0-NEXT: andb $1, %al ; X64-O0-NEXT: movzbl %al, %eax ; X64-O0-NEXT: movzbl var_7, %ecx ; X64-O0-NEXT: cmpl %ecx, %eax ; X64-O0-NEXT: sete %al ; X64-O0-NEXT: andb $1, %al ; X64-O0-NEXT: movzbl %al, %eax ; X64-O0-NEXT: movw %ax, %cx ; X64-O0-NEXT: # implicit-def: $rax ; X64-O0-NEXT: movw %cx, (%rax) ; X64-O0-NEXT: retq ; ; X64-LABEL: f2: ; X64: # %bb.0: # %entry ; X64-NEXT: movzbl var_7(%rip), %eax ; X64-NEXT: xorl %ecx, %ecx ; X64-NEXT: testl %eax, %eax ; X64-NEXT: sete %cl ; X64-NEXT: movl %eax, %edx ; X64-NEXT: xorl %ecx, %edx ; X64-NEXT: movw %dx, -{{[0-9]+}}(%rsp) ; X64-NEXT: xorl %edx, %edx ; X64-NEXT: cmpl %eax, %ecx ; X64-NEXT: sete %dl ; X64-NEXT: movw %dx, (%rax) ; X64-NEXT: retq ; ; X86-O0-LABEL: f2: ; X86-O0: # %bb.0: # %entry ; X86-O0-NEXT: subl $2, %esp ; X86-O0-NEXT: .cfi_def_cfa_offset 6 ; X86-O0-NEXT: movzbl var_7, %eax ; X86-O0-NEXT: cmpb $0, var_7 ; X86-O0-NEXT: setne %cl ; X86-O0-NEXT: xorb $-1, %cl ; X86-O0-NEXT: andb $1, %cl ; X86-O0-NEXT: movzbl %cl, %ecx ; X86-O0-NEXT: xorl %ecx, %eax ; X86-O0-NEXT: # kill: def $ax killed $ax killed $eax ; X86-O0-NEXT: movw %ax, (%esp) ; X86-O0-NEXT: movzbl var_7, %eax ; X86-O0-NEXT: # kill: def $ax killed $ax killed $eax ; X86-O0-NEXT: cmpw $0, %ax ; X86-O0-NEXT: setne %al ; X86-O0-NEXT: xorb $-1, %al ; X86-O0-NEXT: andb $1, %al ; X86-O0-NEXT: movzbl %al, %eax ; X86-O0-NEXT: movzbl var_7, %ecx ; X86-O0-NEXT: cmpl %ecx, %eax ; X86-O0-NEXT: sete %al ; X86-O0-NEXT: andb $1, %al ; X86-O0-NEXT: movzbl %al, %eax ; X86-O0-NEXT: movw %ax, %cx ; X86-O0-NEXT: # implicit-def: $eax ; X86-O0-NEXT: movw %cx, (%eax) ; X86-O0-NEXT: addl $2, %esp ; X86-O0-NEXT: .cfi_def_cfa_offset 4 ; X86-O0-NEXT: retl ; ; X86-LABEL: f2: ; X86: # %bb.0: # %entry ; X86-NEXT: subl $2, %esp ; X86-NEXT: .cfi_def_cfa_offset 6 ; X86-NEXT: movzbl var_7, %ecx ; X86-NEXT: xorl %eax, %eax ; X86-NEXT: testl %ecx, %ecx ; X86-NEXT: sete %al ; X86-NEXT: movl %ecx, %edx ; X86-NEXT: xorl %eax, %edx ; X86-NEXT: movw %dx, (%esp) ; X86-NEXT: xorl %edx, %edx ; X86-NEXT: cmpl %ecx, %eax ; X86-NEXT: sete %dl ; X86-NEXT: movw %dx, (%eax) ; X86-NEXT: addl $2, %esp ; X86-NEXT: .cfi_def_cfa_offset 4 ; X86-NEXT: retl entry: %a = alloca i16, align 2 %0 = load i8, ptr @var_7, align 1 %conv = zext i8 %0 to i32 %1 = load i8, ptr @var_7, align 1 %tobool = icmp ne i8 %1, 0 %lnot = xor i1 %tobool, true %conv1 = zext i1 %lnot to i32 %xor = xor i32 %conv, %conv1 %conv2 = trunc i32 %xor to i16 store i16 %conv2, ptr %a, align 2 %2 = load i8, ptr @var_7, align 1 %conv3 = zext i8 %2 to i16 %tobool4 = icmp ne i16 %conv3, 0 %lnot5 = xor i1 %tobool4, true %conv6 = zext i1 %lnot5 to i32 %3 = load i8, ptr @var_7, align 1 %conv7 = zext i8 %3 to i32 %cmp = icmp eq i32 %conv6, %conv7 %conv8 = zext i1 %cmp to i32 %conv9 = trunc i32 %conv8 to i16 store i16 %conv9, ptr undef, align 2 ret void } @var_13 = external dso_local global i32, align 4 @var_16 = external dso_local global i32, align 4 @var_46 = external dso_local global i32, align 4 define void @f3() #0 { ; X64-O0-LABEL: f3: ; X64-O0: # %bb.0: # %entry ; X64-O0-NEXT: movl var_13, %eax ; X64-O0-NEXT: xorl $-1, %eax ; X64-O0-NEXT: movl %eax, %eax ; X64-O0-NEXT: # kill: def $rax killed $eax ; X64-O0-NEXT: cmpl $0, var_13 ; X64-O0-NEXT: setne %cl ; X64-O0-NEXT: xorb $-1, %cl ; X64-O0-NEXT: andb $1, %cl ; X64-O0-NEXT: movzbl %cl, %ecx ; X64-O0-NEXT: # kill: def $rcx killed $ecx ; X64-O0-NEXT: movl var_13, %edx ; X64-O0-NEXT: xorl $-1, %edx ; X64-O0-NEXT: xorl var_16, %edx ; X64-O0-NEXT: movl %edx, %edx ; X64-O0-NEXT: # kill: def $rdx killed $edx ; X64-O0-NEXT: andq %rdx, %rcx ; X64-O0-NEXT: orq %rcx, %rax ; X64-O0-NEXT: movq %rax, -{{[0-9]+}}(%rsp) ; X64-O0-NEXT: movl var_13, %eax ; X64-O0-NEXT: xorl $-1, %eax ; X64-O0-NEXT: movl %eax, %eax ; X64-O0-NEXT: # kill: def $rax killed $eax ; X64-O0-NEXT: cmpl $0, var_13 ; X64-O0-NEXT: setne %cl ; X64-O0-NEXT: xorb $-1, %cl ; X64-O0-NEXT: andb $1, %cl ; X64-O0-NEXT: movzbl %cl, %ecx ; X64-O0-NEXT: # kill: def $rcx killed $ecx ; X64-O0-NEXT: andq $0, %rcx ; X64-O0-NEXT: orq %rcx, %rax ; X64-O0-NEXT: # kill: def $eax killed $eax killed $rax ; X64-O0-NEXT: movl %eax, var_46 ; X64-O0-NEXT: retq ; ; X64-LABEL: f3: ; X64: # %bb.0: # %entry ; X64-NEXT: movl var_13(%rip), %eax ; X64-NEXT: xorl %ecx, %ecx ; X64-NEXT: testl %eax, %eax ; X64-NEXT: notl %eax ; X64-NEXT: sete %cl ; X64-NEXT: movl var_16(%rip), %edx ; X64-NEXT: xorl %eax, %edx ; X64-NEXT: andl %edx, %ecx ; X64-NEXT: orl %eax, %ecx ; X64-NEXT: movq %rcx, -{{[0-9]+}}(%rsp) ; X64-NEXT: movl %eax, var_46(%rip) ; X64-NEXT: retq ; ; X86-O0-LABEL: f3: ; X86-O0: # %bb.0: # %entry ; X86-O0-NEXT: pushl %ebp ; X86-O0-NEXT: .cfi_def_cfa_offset 8 ; X86-O0-NEXT: .cfi_offset %ebp, -8 ; X86-O0-NEXT: movl %esp, %ebp ; X86-O0-NEXT: .cfi_def_cfa_register %ebp ; X86-O0-NEXT: pushl %esi ; X86-O0-NEXT: andl $-8, %esp ; X86-O0-NEXT: subl $16, %esp ; X86-O0-NEXT: .cfi_offset %esi, -12 ; X86-O0-NEXT: movl var_13, %ecx ; X86-O0-NEXT: movl %ecx, %eax ; X86-O0-NEXT: notl %eax ; X86-O0-NEXT: testl %ecx, %ecx ; X86-O0-NEXT: sete %cl ; X86-O0-NEXT: movzbl %cl, %ecx ; X86-O0-NEXT: movl var_16, %esi ; X86-O0-NEXT: movl %eax, %edx ; X86-O0-NEXT: xorl %esi, %edx ; X86-O0-NEXT: andl %edx, %ecx ; X86-O0-NEXT: orl %ecx, %eax ; X86-O0-NEXT: movl %eax, (%esp) ; X86-O0-NEXT: movl $0, {{[0-9]+}}(%esp) ; X86-O0-NEXT: movl var_13, %eax ; X86-O0-NEXT: notl %eax ; X86-O0-NEXT: movl %eax, var_46 ; X86-O0-NEXT: leal -4(%ebp), %esp ; X86-O0-NEXT: popl %esi ; X86-O0-NEXT: popl %ebp ; X86-O0-NEXT: .cfi_def_cfa %esp, 4 ; X86-O0-NEXT: retl ; ; X86-LABEL: f3: ; X86: # %bb.0: # %entry ; X86-NEXT: pushl %ebp ; X86-NEXT: .cfi_def_cfa_offset 8 ; X86-NEXT: .cfi_offset %ebp, -8 ; X86-NEXT: movl %esp, %ebp ; X86-NEXT: .cfi_def_cfa_register %ebp ; X86-NEXT: andl $-8, %esp ; X86-NEXT: subl $8, %esp ; X86-NEXT: movl var_13, %eax ; X86-NEXT: xorl %ecx, %ecx ; X86-NEXT: testl %eax, %eax ; X86-NEXT: notl %eax ; X86-NEXT: sete %cl ; X86-NEXT: movl var_16, %edx ; X86-NEXT: xorl %eax, %edx ; X86-NEXT: andl %ecx, %edx ; X86-NEXT: orl %eax, %edx ; X86-NEXT: movl %edx, (%esp) ; X86-NEXT: movl $0, {{[0-9]+}}(%esp) ; X86-NEXT: movl %eax, var_46 ; X86-NEXT: movl %ebp, %esp ; X86-NEXT: popl %ebp ; X86-NEXT: .cfi_def_cfa %esp, 4 ; X86-NEXT: retl entry: %a = alloca i64, align 8 %0 = load i32, ptr @var_13, align 4 %neg = xor i32 %0, -1 %conv = zext i32 %neg to i64 %1 = load i32, ptr @var_13, align 4 %tobool = icmp ne i32 %1, 0 %lnot = xor i1 %tobool, true %conv1 = zext i1 %lnot to i64 %2 = load i32, ptr @var_13, align 4 %neg2 = xor i32 %2, -1 %3 = load i32, ptr @var_16, align 4 %xor = xor i32 %neg2, %3 %conv3 = zext i32 %xor to i64 %and = and i64 %conv1, %conv3 %or = or i64 %conv, %and store i64 %or, ptr %a, align 8 %4 = load i32, ptr @var_13, align 4 %neg4 = xor i32 %4, -1 %conv5 = zext i32 %neg4 to i64 %5 = load i32, ptr @var_13, align 4 %tobool6 = icmp ne i32 %5, 0 %lnot7 = xor i1 %tobool6, true %conv8 = zext i1 %lnot7 to i64 %and9 = and i64 %conv8, 0 %or10 = or i64 %conv5, %and9 %conv11 = trunc i64 %or10 to i32 store i32 %conv11, ptr @var_46, align 4 ret void }