; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py ; RUN: llc < %s -mtriple=i686-unknown-unknown | FileCheck %s --check-prefixes=X86 ; RUN: llc < %s -mtriple=x86_64-unknown-unknown | FileCheck %s --check-prefixes=X64,NOBMI ; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+bmi | FileCheck %s --check-prefixes=X64,BMI ; InstCombine and DAGCombiner transform an 'add' into an 'or' ; if there are no common bits from the incoming operands. ; LEA instruction selection should be able to see through that ; transform and reduce add/shift/or instruction counts. define i32 @or_shift1_and1(i32 %x, i32 %y) { ; X86-LABEL: or_shift1_and1: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: andl $1, %ecx ; X86-NEXT: leal (%ecx,%eax,2), %eax ; X86-NEXT: retl ; ; X64-LABEL: or_shift1_and1: ; X64: # %bb.0: ; X64-NEXT: # kill: def $esi killed $esi def $rsi ; X64-NEXT: # kill: def $edi killed $edi def $rdi ; X64-NEXT: andl $1, %esi ; X64-NEXT: leal (%rsi,%rdi,2), %eax ; X64-NEXT: retq %shl = shl i32 %x, 1 %and = and i32 %y, 1 %or = or i32 %and, %shl ret i32 %or } define i32 @or_shift1_and1_swapped(i32 %x, i32 %y) { ; X86-LABEL: or_shift1_and1_swapped: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: andl $1, %ecx ; X86-NEXT: leal (%ecx,%eax,2), %eax ; X86-NEXT: retl ; ; X64-LABEL: or_shift1_and1_swapped: ; X64: # %bb.0: ; X64-NEXT: # kill: def $esi killed $esi def $rsi ; X64-NEXT: # kill: def $edi killed $edi def $rdi ; X64-NEXT: andl $1, %esi ; X64-NEXT: leal (%rsi,%rdi,2), %eax ; X64-NEXT: retq %shl = shl i32 %x, 1 %and = and i32 %y, 1 %or = or i32 %shl, %and ret i32 %or } define i32 @or_shift2_and1(i32 %x, i32 %y) { ; X86-LABEL: or_shift2_and1: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: andl $1, %ecx ; X86-NEXT: leal (%ecx,%eax,4), %eax ; X86-NEXT: retl ; ; X64-LABEL: or_shift2_and1: ; X64: # %bb.0: ; X64-NEXT: # kill: def $esi killed $esi def $rsi ; X64-NEXT: # kill: def $edi killed $edi def $rdi ; X64-NEXT: andl $1, %esi ; X64-NEXT: leal (%rsi,%rdi,4), %eax ; X64-NEXT: retq %shl = shl i32 %x, 2 %and = and i32 %y, 1 %or = or i32 %shl, %and ret i32 %or } define i32 @or_shift3_and1(i32 %x, i32 %y) { ; X86-LABEL: or_shift3_and1: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: andl $1, %ecx ; X86-NEXT: leal (%ecx,%eax,8), %eax ; X86-NEXT: retl ; ; X64-LABEL: or_shift3_and1: ; X64: # %bb.0: ; X64-NEXT: # kill: def $esi killed $esi def $rsi ; X64-NEXT: # kill: def $edi killed $edi def $rdi ; X64-NEXT: andl $1, %esi ; X64-NEXT: leal (%rsi,%rdi,8), %eax ; X64-NEXT: retq %shl = shl i32 %x, 3 %and = and i32 %y, 1 %or = or i32 %shl, %and ret i32 %or } define i32 @or_shift3_and7(i32 %x, i32 %y) { ; X86-LABEL: or_shift3_and7: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: andl $7, %ecx ; X86-NEXT: leal (%ecx,%eax,8), %eax ; X86-NEXT: retl ; ; X64-LABEL: or_shift3_and7: ; X64: # %bb.0: ; X64-NEXT: # kill: def $esi killed $esi def $rsi ; X64-NEXT: # kill: def $edi killed $edi def $rdi ; X64-NEXT: andl $7, %esi ; X64-NEXT: leal (%rsi,%rdi,8), %eax ; X64-NEXT: retq %shl = shl i32 %x, 3 %and = and i32 %y, 7 %or = or i32 %shl, %and ret i32 %or } ; The shift is too big for an LEA. define i32 @or_shift4_and1(i32 %x, i32 %y) { ; X86-LABEL: or_shift4_and1: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: shll $4, %ecx ; X86-NEXT: andl $1, %eax ; X86-NEXT: orl %ecx, %eax ; X86-NEXT: retl ; ; X64-LABEL: or_shift4_and1: ; X64: # %bb.0: ; X64-NEXT: # kill: def $esi killed $esi def $rsi ; X64-NEXT: # kill: def $edi killed $edi def $rdi ; X64-NEXT: shll $4, %edi ; X64-NEXT: andl $1, %esi ; X64-NEXT: leal (%rsi,%rdi), %eax ; X64-NEXT: retq %shl = shl i32 %x, 4 %and = and i32 %y, 1 %or = or i32 %shl, %and ret i32 %or } ; The mask is too big for the shift, so the 'or' isn't equivalent to an 'add'. define i32 @or_shift3_and8(i32 %x, i32 %y) { ; X86-LABEL: or_shift3_and8: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: shll $3, %ecx ; X86-NEXT: andl $8, %eax ; X86-NEXT: orl %ecx, %eax ; X86-NEXT: retl ; ; X64-LABEL: or_shift3_and8: ; X64: # %bb.0: ; X64-NEXT: # kill: def $edi killed $edi def $rdi ; X64-NEXT: leal (,%rdi,8), %eax ; X64-NEXT: andl $8, %esi ; X64-NEXT: orl %esi, %eax ; X64-NEXT: retq %shl = shl i32 %x, 3 %and = and i32 %y, 8 %or = or i32 %shl, %and ret i32 %or } ; 64-bit operands should work too. define i64 @or_shift1_and1_64(i64 %x, i64 %y) { ; X86-LABEL: or_shift1_and1_64: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: movl {{[0-9]+}}(%esp), %edx ; X86-NEXT: shldl $1, %ecx, %edx ; X86-NEXT: andl $1, %eax ; X86-NEXT: leal (%eax,%ecx,2), %eax ; X86-NEXT: retl ; ; X64-LABEL: or_shift1_and1_64: ; X64: # %bb.0: ; X64-NEXT: andl $1, %esi ; X64-NEXT: leaq (%rsi,%rdi,2), %rax ; X64-NEXT: retq %shl = shl i64 %x, 1 %and = and i64 %y, 1 %or = or i64 %and, %shl ret i64 %or } ; In the following patterns, lhs and rhs of the or instruction have no common bits. define i32 @or_and_and_rhs_neg_i32(i32 %x, i32 %y, i32 %z) { ; X86-LABEL: or_and_and_rhs_neg_i32: ; X86: # %bb.0: # %entry ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: xorl %ecx, %eax ; X86-NEXT: andl {{[0-9]+}}(%esp), %eax ; X86-NEXT: xorl %ecx, %eax ; X86-NEXT: incl %eax ; X86-NEXT: retl ; ; NOBMI-LABEL: or_and_and_rhs_neg_i32: ; NOBMI: # %bb.0: # %entry ; NOBMI-NEXT: # kill: def $edx killed $edx def $rdx ; NOBMI-NEXT: xorl %edi, %edx ; NOBMI-NEXT: andl %esi, %edx ; NOBMI-NEXT: xorl %edi, %edx ; NOBMI-NEXT: leal 1(%rdx), %eax ; NOBMI-NEXT: retq ; ; BMI-LABEL: or_and_and_rhs_neg_i32: ; BMI: # %bb.0: # %entry ; BMI-NEXT: # kill: def $edx killed $edx def $rdx ; BMI-NEXT: andl %esi, %edx ; BMI-NEXT: andnl %edi, %esi, %eax ; BMI-NEXT: leal 1(%rdx,%rax), %eax ; BMI-NEXT: retq entry: %and1 = and i32 %z, %y %xor = xor i32 %y, -1 %and2 = and i32 %x, %xor %or = or i32 %and1, %and2 %inc = add i32 %or, 1 ret i32 %inc } define i32 @or_and_and_lhs_neg_i32(i32 %x, i32 %y, i32 %z) { ; X86-LABEL: or_and_and_lhs_neg_i32: ; X86: # %bb.0: # %entry ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: xorl %ecx, %eax ; X86-NEXT: andl {{[0-9]+}}(%esp), %eax ; X86-NEXT: xorl %ecx, %eax ; X86-NEXT: incl %eax ; X86-NEXT: retl ; ; NOBMI-LABEL: or_and_and_lhs_neg_i32: ; NOBMI: # %bb.0: # %entry ; NOBMI-NEXT: # kill: def $edx killed $edx def $rdx ; NOBMI-NEXT: xorl %edi, %edx ; NOBMI-NEXT: andl %esi, %edx ; NOBMI-NEXT: xorl %edi, %edx ; NOBMI-NEXT: leal 1(%rdx), %eax ; NOBMI-NEXT: retq ; ; BMI-LABEL: or_and_and_lhs_neg_i32: ; BMI: # %bb.0: # %entry ; BMI-NEXT: # kill: def $edx killed $edx def $rdx ; BMI-NEXT: andl %esi, %edx ; BMI-NEXT: andnl %edi, %esi, %eax ; BMI-NEXT: leal 1(%rdx,%rax), %eax ; BMI-NEXT: retq entry: %and1 = and i32 %z, %y %xor = xor i32 %y, -1 %and2 = and i32 %xor, %x %or = or i32 %and1, %and2 %inc = add i32 %or, 1 ret i32 %inc } define i32 @or_and_rhs_neg_and_i32(i32 %x, i32 %y, i32 %z) { ; X86-LABEL: or_and_rhs_neg_and_i32: ; X86: # %bb.0: # %entry ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: xorl %ecx, %eax ; X86-NEXT: andl {{[0-9]+}}(%esp), %eax ; X86-NEXT: xorl %ecx, %eax ; X86-NEXT: incl %eax ; X86-NEXT: retl ; ; NOBMI-LABEL: or_and_rhs_neg_and_i32: ; NOBMI: # %bb.0: # %entry ; NOBMI-NEXT: # kill: def $edi killed $edi def $rdi ; NOBMI-NEXT: xorl %edx, %edi ; NOBMI-NEXT: andl %esi, %edi ; NOBMI-NEXT: xorl %edx, %edi ; NOBMI-NEXT: leal 1(%rdi), %eax ; NOBMI-NEXT: retq ; ; BMI-LABEL: or_and_rhs_neg_and_i32: ; BMI: # %bb.0: # %entry ; BMI-NEXT: # kill: def $edi killed $edi def $rdi ; BMI-NEXT: andnl %edx, %esi, %eax ; BMI-NEXT: andl %esi, %edi ; BMI-NEXT: leal 1(%rax,%rdi), %eax ; BMI-NEXT: retq entry: %xor = xor i32 %y, -1 %and1 = and i32 %z, %xor %and2 = and i32 %x, %y %or = or i32 %and1, %and2 %inc = add i32 %or, 1 ret i32 %inc } define i32 @or_and_lhs_neg_and_i32(i32 %x, i32 %y, i32 %z) { ; X86-LABEL: or_and_lhs_neg_and_i32: ; X86: # %bb.0: # %entry ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: xorl %ecx, %eax ; X86-NEXT: andl {{[0-9]+}}(%esp), %eax ; X86-NEXT: xorl %ecx, %eax ; X86-NEXT: incl %eax ; X86-NEXT: retl ; ; NOBMI-LABEL: or_and_lhs_neg_and_i32: ; NOBMI: # %bb.0: # %entry ; NOBMI-NEXT: # kill: def $edi killed $edi def $rdi ; NOBMI-NEXT: xorl %edx, %edi ; NOBMI-NEXT: andl %esi, %edi ; NOBMI-NEXT: xorl %edx, %edi ; NOBMI-NEXT: leal 1(%rdi), %eax ; NOBMI-NEXT: retq ; ; BMI-LABEL: or_and_lhs_neg_and_i32: ; BMI: # %bb.0: # %entry ; BMI-NEXT: # kill: def $edi killed $edi def $rdi ; BMI-NEXT: andnl %edx, %esi, %eax ; BMI-NEXT: andl %esi, %edi ; BMI-NEXT: leal 1(%rax,%rdi), %eax ; BMI-NEXT: retq entry: %xor = xor i32 %y, -1 %and1 = and i32 %xor, %z %and2 = and i32 %x, %y %or = or i32 %and1, %and2 %inc = add i32 %or, 1 ret i32 %inc } define i64 @or_and_and_rhs_neg_i64(i64 %x, i64 %y, i64 %z) { ; X86-LABEL: or_and_and_rhs_neg_i64: ; X86: # %bb.0: # %entry ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: movl {{[0-9]+}}(%esp), %edx ; X86-NEXT: xorl %eax, %edx ; X86-NEXT: andl {{[0-9]+}}(%esp), %edx ; X86-NEXT: xorl %eax, %edx ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: xorl %ecx, %eax ; X86-NEXT: andl {{[0-9]+}}(%esp), %eax ; X86-NEXT: xorl %ecx, %eax ; X86-NEXT: addl $1, %eax ; X86-NEXT: adcl $0, %edx ; X86-NEXT: retl ; ; NOBMI-LABEL: or_and_and_rhs_neg_i64: ; NOBMI: # %bb.0: # %entry ; NOBMI-NEXT: xorq %rdi, %rdx ; NOBMI-NEXT: andq %rsi, %rdx ; NOBMI-NEXT: xorq %rdi, %rdx ; NOBMI-NEXT: leaq 1(%rdx), %rax ; NOBMI-NEXT: retq ; ; BMI-LABEL: or_and_and_rhs_neg_i64: ; BMI: # %bb.0: # %entry ; BMI-NEXT: andq %rsi, %rdx ; BMI-NEXT: andnq %rdi, %rsi, %rax ; BMI-NEXT: leaq 1(%rdx,%rax), %rax ; BMI-NEXT: retq entry: %and1 = and i64 %z, %y %xor = xor i64 %y, -1 %and2 = and i64 %x, %xor %or = or i64 %and1, %and2 %inc = add i64 %or, 1 ret i64 %inc } define i64 @or_and_and_lhs_neg_i64(i64 %x, i64 %y, i64 %z) { ; X86-LABEL: or_and_and_lhs_neg_i64: ; X86: # %bb.0: # %entry ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: movl {{[0-9]+}}(%esp), %edx ; X86-NEXT: xorl %eax, %edx ; X86-NEXT: andl {{[0-9]+}}(%esp), %edx ; X86-NEXT: xorl %eax, %edx ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: xorl %ecx, %eax ; X86-NEXT: andl {{[0-9]+}}(%esp), %eax ; X86-NEXT: xorl %ecx, %eax ; X86-NEXT: addl $1, %eax ; X86-NEXT: adcl $0, %edx ; X86-NEXT: retl ; ; NOBMI-LABEL: or_and_and_lhs_neg_i64: ; NOBMI: # %bb.0: # %entry ; NOBMI-NEXT: xorq %rdi, %rdx ; NOBMI-NEXT: andq %rsi, %rdx ; NOBMI-NEXT: xorq %rdi, %rdx ; NOBMI-NEXT: leaq 1(%rdx), %rax ; NOBMI-NEXT: retq ; ; BMI-LABEL: or_and_and_lhs_neg_i64: ; BMI: # %bb.0: # %entry ; BMI-NEXT: andq %rsi, %rdx ; BMI-NEXT: andnq %rdi, %rsi, %rax ; BMI-NEXT: leaq 1(%rdx,%rax), %rax ; BMI-NEXT: retq entry: %and1 = and i64 %z, %y %xor = xor i64 %y, -1 %and2 = and i64 %xor, %x %or = or i64 %and1, %and2 %inc = add i64 %or, 1 ret i64 %inc } define i64 @or_and_rhs_neg_and_i64(i64 %x, i64 %y, i64 %z) { ; X86-LABEL: or_and_rhs_neg_and_i64: ; X86: # %bb.0: # %entry ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: movl {{[0-9]+}}(%esp), %edx ; X86-NEXT: xorl %eax, %edx ; X86-NEXT: andl {{[0-9]+}}(%esp), %edx ; X86-NEXT: xorl %eax, %edx ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: xorl %ecx, %eax ; X86-NEXT: andl {{[0-9]+}}(%esp), %eax ; X86-NEXT: xorl %ecx, %eax ; X86-NEXT: addl $1, %eax ; X86-NEXT: adcl $0, %edx ; X86-NEXT: retl ; ; NOBMI-LABEL: or_and_rhs_neg_and_i64: ; NOBMI: # %bb.0: # %entry ; NOBMI-NEXT: xorq %rdx, %rdi ; NOBMI-NEXT: andq %rsi, %rdi ; NOBMI-NEXT: xorq %rdx, %rdi ; NOBMI-NEXT: leaq 1(%rdi), %rax ; NOBMI-NEXT: retq ; ; BMI-LABEL: or_and_rhs_neg_and_i64: ; BMI: # %bb.0: # %entry ; BMI-NEXT: andnq %rdx, %rsi, %rax ; BMI-NEXT: andq %rsi, %rdi ; BMI-NEXT: leaq 1(%rax,%rdi), %rax ; BMI-NEXT: retq entry: %xor = xor i64 %y, -1 %and1 = and i64 %z, %xor %and2 = and i64 %x, %y %or = or i64 %and1, %and2 %inc = add i64 %or, 1 ret i64 %inc } define i64 @or_and_lhs_neg_and_i64(i64 %x, i64 %y, i64 %z) { ; X86-LABEL: or_and_lhs_neg_and_i64: ; X86: # %bb.0: # %entry ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: movl {{[0-9]+}}(%esp), %edx ; X86-NEXT: xorl %eax, %edx ; X86-NEXT: andl {{[0-9]+}}(%esp), %edx ; X86-NEXT: xorl %eax, %edx ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: xorl %ecx, %eax ; X86-NEXT: andl {{[0-9]+}}(%esp), %eax ; X86-NEXT: xorl %ecx, %eax ; X86-NEXT: addl $1, %eax ; X86-NEXT: adcl $0, %edx ; X86-NEXT: retl ; ; NOBMI-LABEL: or_and_lhs_neg_and_i64: ; NOBMI: # %bb.0: # %entry ; NOBMI-NEXT: xorq %rdx, %rdi ; NOBMI-NEXT: andq %rsi, %rdi ; NOBMI-NEXT: xorq %rdx, %rdi ; NOBMI-NEXT: leaq 1(%rdi), %rax ; NOBMI-NEXT: retq ; ; BMI-LABEL: or_and_lhs_neg_and_i64: ; BMI: # %bb.0: # %entry ; BMI-NEXT: andnq %rdx, %rsi, %rax ; BMI-NEXT: andq %rsi, %rdi ; BMI-NEXT: leaq 1(%rax,%rdi), %rax ; BMI-NEXT: retq entry: %xor = xor i64 %y, -1 %and1 = and i64 %xor, %z %and2 = and i64 %x, %y %or = or i64 %and1, %and2 %inc = add i64 %or, 1 ret i64 %inc } define i32 @or_sext1(i32 %x) { ; X86-LABEL: or_sext1: ; X86: # %bb.0: ; X86-NEXT: xorl %eax, %eax ; X86-NEXT: cmpl $43, {{[0-9]+}}(%esp) ; X86-NEXT: setl %al ; X86-NEXT: leal -1(%eax,%eax), %eax ; X86-NEXT: retl ; ; X64-LABEL: or_sext1: ; X64: # %bb.0: ; X64-NEXT: xorl %eax, %eax ; X64-NEXT: cmpl $43, %edi ; X64-NEXT: setl %al ; X64-NEXT: leal -1(%rax,%rax), %eax ; X64-NEXT: retq %cmp = icmp sgt i32 %x, 42 %sext = sext i1 %cmp to i32 %or = or i32 %sext, 1 ret i32 %or } define i64 @or_sext1_64(i64 %x) { ; X86-LABEL: or_sext1_64: ; X86: # %bb.0: ; X86-NEXT: xorl %eax, %eax ; X86-NEXT: movl $42, %ecx ; X86-NEXT: cmpl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: sbbl {{[0-9]+}}(%esp), %eax ; X86-NEXT: setl %al ; X86-NEXT: movzbl %al, %edx ; X86-NEXT: negl %edx ; X86-NEXT: movl %edx, %eax ; X86-NEXT: orl $1, %eax ; X86-NEXT: retl ; ; X64-LABEL: or_sext1_64: ; X64: # %bb.0: ; X64-NEXT: xorl %eax, %eax ; X64-NEXT: cmpq $43, %rdi ; X64-NEXT: setl %al ; X64-NEXT: leaq -1(%rax,%rax), %rax ; X64-NEXT: retq %cmp = icmp sgt i64 %x, 42 %sext = sext i1 %cmp to i64 %or = or i64 %sext, 1 ret i64 %or } define i32 @or_sext2(i32 %x) { ; X86-LABEL: or_sext2: ; X86: # %bb.0: ; X86-NEXT: xorl %eax, %eax ; X86-NEXT: cmpl $43, {{[0-9]+}}(%esp) ; X86-NEXT: setl %al ; X86-NEXT: leal -1(%eax,%eax,2), %eax ; X86-NEXT: retl ; ; X64-LABEL: or_sext2: ; X64: # %bb.0: ; X64-NEXT: xorl %eax, %eax ; X64-NEXT: cmpl $43, %edi ; X64-NEXT: setl %al ; X64-NEXT: leal -1(%rax,%rax,2), %eax ; X64-NEXT: retq %cmp = icmp sgt i32 %x, 42 %sext = sext i1 %cmp to i32 %or = or i32 %sext, 2 ret i32 %or } define i64 @or_sext2_64(i64 %x) { ; X86-LABEL: or_sext2_64: ; X86: # %bb.0: ; X86-NEXT: xorl %eax, %eax ; X86-NEXT: movl $42, %ecx ; X86-NEXT: cmpl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: sbbl {{[0-9]+}}(%esp), %eax ; X86-NEXT: setl %al ; X86-NEXT: movzbl %al, %edx ; X86-NEXT: negl %edx ; X86-NEXT: movl %edx, %eax ; X86-NEXT: orl $2, %eax ; X86-NEXT: retl ; ; X64-LABEL: or_sext2_64: ; X64: # %bb.0: ; X64-NEXT: xorl %eax, %eax ; X64-NEXT: cmpq $43, %rdi ; X64-NEXT: setl %al ; X64-NEXT: leaq -1(%rax,%rax,2), %rax ; X64-NEXT: retq %cmp = icmp sgt i64 %x, 42 %sext = sext i1 %cmp to i64 %or = or i64 %sext, 2 ret i64 %or } define i32 @or_sext3(i32 %x) { ; X86-LABEL: or_sext3: ; X86: # %bb.0: ; X86-NEXT: xorl %eax, %eax ; X86-NEXT: cmpl $43, {{[0-9]+}}(%esp) ; X86-NEXT: setl %al ; X86-NEXT: leal -1(,%eax,4), %eax ; X86-NEXT: retl ; ; X64-LABEL: or_sext3: ; X64: # %bb.0: ; X64-NEXT: xorl %eax, %eax ; X64-NEXT: cmpl $43, %edi ; X64-NEXT: setl %al ; X64-NEXT: leal -1(,%rax,4), %eax ; X64-NEXT: retq %cmp = icmp sgt i32 %x, 42 %sext = sext i1 %cmp to i32 %or = or i32 %sext, 3 ret i32 %or } define i64 @or_sext3_64(i64 %x) { ; X86-LABEL: or_sext3_64: ; X86: # %bb.0: ; X86-NEXT: xorl %eax, %eax ; X86-NEXT: movl $42, %ecx ; X86-NEXT: cmpl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: sbbl {{[0-9]+}}(%esp), %eax ; X86-NEXT: setl %al ; X86-NEXT: movzbl %al, %edx ; X86-NEXT: negl %edx ; X86-NEXT: movl %edx, %eax ; X86-NEXT: orl $3, %eax ; X86-NEXT: retl ; ; X64-LABEL: or_sext3_64: ; X64: # %bb.0: ; X64-NEXT: xorl %eax, %eax ; X64-NEXT: cmpq $43, %rdi ; X64-NEXT: setl %al ; X64-NEXT: leaq -1(,%rax,4), %rax ; X64-NEXT: retq %cmp = icmp sgt i64 %x, 42 %sext = sext i1 %cmp to i64 %or = or i64 %sext, 3 ret i64 %or } define i32 @or_sext4(i32 %x) { ; X86-LABEL: or_sext4: ; X86: # %bb.0: ; X86-NEXT: xorl %eax, %eax ; X86-NEXT: cmpl $43, {{[0-9]+}}(%esp) ; X86-NEXT: setl %al ; X86-NEXT: leal -1(%eax,%eax,4), %eax ; X86-NEXT: retl ; ; X64-LABEL: or_sext4: ; X64: # %bb.0: ; X64-NEXT: xorl %eax, %eax ; X64-NEXT: cmpl $43, %edi ; X64-NEXT: setl %al ; X64-NEXT: leal -1(%rax,%rax,4), %eax ; X64-NEXT: retq %cmp = icmp sgt i32 %x, 42 %sext = sext i1 %cmp to i32 %or = or i32 %sext, 4 ret i32 %or } define i64 @or_sext4_64(i64 %x) { ; X86-LABEL: or_sext4_64: ; X86: # %bb.0: ; X86-NEXT: xorl %eax, %eax ; X86-NEXT: movl $42, %ecx ; X86-NEXT: cmpl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: sbbl {{[0-9]+}}(%esp), %eax ; X86-NEXT: setl %al ; X86-NEXT: movzbl %al, %edx ; X86-NEXT: negl %edx ; X86-NEXT: movl %edx, %eax ; X86-NEXT: orl $4, %eax ; X86-NEXT: retl ; ; X64-LABEL: or_sext4_64: ; X64: # %bb.0: ; X64-NEXT: xorl %eax, %eax ; X64-NEXT: cmpq $43, %rdi ; X64-NEXT: setl %al ; X64-NEXT: leaq -1(%rax,%rax,4), %rax ; X64-NEXT: retq %cmp = icmp sgt i64 %x, 42 %sext = sext i1 %cmp to i64 %or = or i64 %sext, 4 ret i64 %or } define i32 @or_sext7(i32 %x) { ; X86-LABEL: or_sext7: ; X86: # %bb.0: ; X86-NEXT: xorl %eax, %eax ; X86-NEXT: cmpl $43, {{[0-9]+}}(%esp) ; X86-NEXT: setl %al ; X86-NEXT: leal -1(,%eax,8), %eax ; X86-NEXT: retl ; ; X64-LABEL: or_sext7: ; X64: # %bb.0: ; X64-NEXT: xorl %eax, %eax ; X64-NEXT: cmpl $43, %edi ; X64-NEXT: setl %al ; X64-NEXT: leal -1(,%rax,8), %eax ; X64-NEXT: retq %cmp = icmp sgt i32 %x, 42 %sext = sext i1 %cmp to i32 %or = or i32 %sext, 7 ret i32 %or } define i64 @or_sext7_64(i64 %x) { ; X86-LABEL: or_sext7_64: ; X86: # %bb.0: ; X86-NEXT: xorl %eax, %eax ; X86-NEXT: movl $42, %ecx ; X86-NEXT: cmpl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: sbbl {{[0-9]+}}(%esp), %eax ; X86-NEXT: setl %al ; X86-NEXT: movzbl %al, %edx ; X86-NEXT: negl %edx ; X86-NEXT: movl %edx, %eax ; X86-NEXT: orl $7, %eax ; X86-NEXT: retl ; ; X64-LABEL: or_sext7_64: ; X64: # %bb.0: ; X64-NEXT: xorl %eax, %eax ; X64-NEXT: cmpq $43, %rdi ; X64-NEXT: setl %al ; X64-NEXT: leaq -1(,%rax,8), %rax ; X64-NEXT: retq %cmp = icmp sgt i64 %x, 42 %sext = sext i1 %cmp to i64 %or = or i64 %sext, 7 ret i64 %or } define i32 @or_sext8(i32 %x) { ; X86-LABEL: or_sext8: ; X86: # %bb.0: ; X86-NEXT: xorl %eax, %eax ; X86-NEXT: cmpl $43, {{[0-9]+}}(%esp) ; X86-NEXT: setl %al ; X86-NEXT: leal -1(%eax,%eax,8), %eax ; X86-NEXT: retl ; ; X64-LABEL: or_sext8: ; X64: # %bb.0: ; X64-NEXT: xorl %eax, %eax ; X64-NEXT: cmpl $43, %edi ; X64-NEXT: setl %al ; X64-NEXT: leal -1(%rax,%rax,8), %eax ; X64-NEXT: retq %cmp = icmp sgt i32 %x, 42 %sext = sext i1 %cmp to i32 %or = or i32 %sext, 8 ret i32 %or } define i64 @or_sext8_64(i64 %x) { ; X86-LABEL: or_sext8_64: ; X86: # %bb.0: ; X86-NEXT: xorl %eax, %eax ; X86-NEXT: movl $42, %ecx ; X86-NEXT: cmpl {{[0-9]+}}(%esp), %ecx ; X86-NEXT: sbbl {{[0-9]+}}(%esp), %eax ; X86-NEXT: setl %al ; X86-NEXT: movzbl %al, %edx ; X86-NEXT: negl %edx ; X86-NEXT: movl %edx, %eax ; X86-NEXT: orl $8, %eax ; X86-NEXT: retl ; ; X64-LABEL: or_sext8_64: ; X64: # %bb.0: ; X64-NEXT: xorl %eax, %eax ; X64-NEXT: cmpq $43, %rdi ; X64-NEXT: setl %al ; X64-NEXT: leaq -1(%rax,%rax,8), %rax ; X64-NEXT: retq %cmp = icmp sgt i64 %x, 42 %sext = sext i1 %cmp to i64 %or = or i64 %sext, 8 ret i64 %or } define i64 @or_large_constant(i64 %x) { ; X86-LABEL: or_large_constant: ; X86: # %bb.0: # %entry ; X86-NEXT: xorl %edx, %edx ; X86-NEXT: movl $1, %eax ; X86-NEXT: cmpl {{[0-9]+}}(%esp), %eax ; X86-NEXT: movl $0, %eax ; X86-NEXT: sbbl {{[0-9]+}}(%esp), %eax ; X86-NEXT: setl %al ; X86-NEXT: movzbl %al, %eax ; X86-NEXT: negl %eax ; X86-NEXT: sbbl %edx, %edx ; X86-NEXT: orl $1, %eax ; X86-NEXT: orl $128, %edx ; X86-NEXT: retl ; ; X64-LABEL: or_large_constant: ; X64: # %bb.0: # %entry ; X64-NEXT: xorl %ecx, %ecx ; X64-NEXT: cmpq $2, %rdi ; X64-NEXT: setge %cl ; X64-NEXT: negq %rcx ; X64-NEXT: movabsq $549755813889, %rax # imm = 0x8000000001 ; X64-NEXT: orq %rcx, %rax ; X64-NEXT: retq entry: %cmp = icmp sgt i64 %x, 1 %zext = zext i1 %cmp to i64 %sub = sub i64 0, %zext %or = or i64 %sub, 549755813889 ; 0x8000000001 ret i64 %or } define i32 @or_shift1_disjoint(i32 %x, i32 %y) { ; X86-LABEL: or_shift1_disjoint: ; X86: # %bb.0: ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax ; X86-NEXT: addl %eax, %eax ; X86-NEXT: orl {{[0-9]+}}(%esp), %eax ; X86-NEXT: retl ; ; X64-LABEL: or_shift1_disjoint: ; X64: # %bb.0: ; X64-NEXT: # kill: def $esi killed $esi def $rsi ; X64-NEXT: # kill: def $edi killed $edi def $rdi ; X64-NEXT: leal (%rsi,%rdi,2), %eax ; X64-NEXT: retq %shl = shl i32 %x, 1 %or = or disjoint i32 %y, %shl ret i32 %or }