; NOTE: Assertions have been autogenerated by utils/update_test_checks.py ; This test makes sure that these instructions are properly eliminated. ; ; RUN: opt < %s -passes=instcombine -S | FileCheck %s declare void @use(i8) define i32 @shl_C1_add_A_C2_i32(i16 %A) { ; CHECK-LABEL: @shl_C1_add_A_C2_i32( ; CHECK-NEXT: [[B:%.*]] = zext nneg i16 [[A:%.*]] to i32 ; CHECK-NEXT: [[D:%.*]] = shl i32 192, [[B]] ; CHECK-NEXT: ret i32 [[D]] ; %B = zext i16 %A to i32 %C = add i32 %B, 5 %D = shl i32 6, %C ret i32 %D } define i32 @ashr_C1_add_A_C2_i32(i32 %A) { ; CHECK-LABEL: @ashr_C1_add_A_C2_i32( ; CHECK-NEXT: ret i32 0 ; %B = and i32 %A, 65535 %C = add i32 %B, 5 %D = ashr i32 6, %C ret i32 %D } define i32 @lshr_C1_add_A_C2_i32(i32 %A) { ; CHECK-LABEL: @lshr_C1_add_A_C2_i32( ; CHECK-NEXT: [[B:%.*]] = and i32 [[A:%.*]], 65535 ; CHECK-NEXT: [[D:%.*]] = shl i32 192, [[B]] ; CHECK-NEXT: ret i32 [[D]] ; %B = and i32 %A, 65535 %C = add i32 %B, 5 %D = shl i32 6, %C ret i32 %D } define <4 x i32> @shl_C1_add_A_C2_v4i32(<4 x i16> %A) { ; CHECK-LABEL: @shl_C1_add_A_C2_v4i32( ; CHECK-NEXT: [[B:%.*]] = zext nneg <4 x i16> [[A:%.*]] to <4 x i32> ; CHECK-NEXT: [[D:%.*]] = shl <4 x i32> , [[B]] ; CHECK-NEXT: ret <4 x i32> [[D]] ; %B = zext <4 x i16> %A to <4 x i32> %C = add <4 x i32> %B, %D = shl <4 x i32> , %C ret <4 x i32> %D } define <4 x i32> @ashr_C1_add_A_C2_v4i32(<4 x i32> %A) { ; CHECK-LABEL: @ashr_C1_add_A_C2_v4i32( ; CHECK-NEXT: [[B:%.*]] = and <4 x i32> [[A:%.*]], ; CHECK-NEXT: [[D:%.*]] = ashr <4 x i32> , [[B]] ; CHECK-NEXT: ret <4 x i32> [[D]] ; %B = and <4 x i32> %A, %C = add <4 x i32> %B, %D = ashr <4 x i32> , %C ret <4 x i32> %D } define <4 x i32> @lshr_C1_add_A_C2_v4i32(<4 x i32> %A) { ; CHECK-LABEL: @lshr_C1_add_A_C2_v4i32( ; CHECK-NEXT: [[B:%.*]] = and <4 x i32> [[A:%.*]], ; CHECK-NEXT: [[D:%.*]] = lshr <4 x i32> , [[B]] ; CHECK-NEXT: ret <4 x i32> [[D]] ; %B = and <4 x i32> %A, %C = add <4 x i32> %B, %D = lshr <4 x i32> , %C ret <4 x i32> %D } define <4 x i32> @shl_C1_add_A_C2_v4i32_splat(i16 %I) { ; CHECK-LABEL: @shl_C1_add_A_C2_v4i32_splat( ; CHECK-NEXT: [[A:%.*]] = zext i16 [[I:%.*]] to i32 ; CHECK-NEXT: [[B:%.*]] = insertelement <4 x i32> poison, i32 [[A]], i64 0 ; CHECK-NEXT: [[C:%.*]] = shufflevector <4 x i32> [[B]], <4 x i32> poison, <4 x i32> zeroinitializer ; CHECK-NEXT: [[E:%.*]] = shl <4 x i32> , [[C]] ; CHECK-NEXT: ret <4 x i32> [[E]] ; %A = zext i16 %I to i32 %B = insertelement <4 x i32> undef, i32 %A, i32 0 %C = shufflevector <4 x i32> %B, <4 x i32> undef, <4 x i32> zeroinitializer %D = add <4 x i32> %C, %E = shl <4 x i32> , %D ret <4 x i32> %E } define <4 x i32> @ashr_C1_add_A_C2_v4i32_splat(i16 %I) { ; CHECK-LABEL: @ashr_C1_add_A_C2_v4i32_splat( ; CHECK-NEXT: [[A:%.*]] = zext i16 [[I:%.*]] to i32 ; CHECK-NEXT: [[B:%.*]] = insertelement <4 x i32> poison, i32 [[A]], i64 0 ; CHECK-NEXT: [[C:%.*]] = shufflevector <4 x i32> [[B]], <4 x i32> poison, <4 x i32> zeroinitializer ; CHECK-NEXT: [[E:%.*]] = ashr <4 x i32> , [[C]] ; CHECK-NEXT: ret <4 x i32> [[E]] ; %A = zext i16 %I to i32 %B = insertelement <4 x i32> undef, i32 %A, i32 0 %C = shufflevector <4 x i32> %B, <4 x i32> undef, <4 x i32> zeroinitializer %D = add <4 x i32> %C, %E = ashr <4 x i32> , %D ret <4 x i32> %E } define <4 x i32> @lshr_C1_add_A_C2_v4i32_splat(i16 %I) { ; CHECK-LABEL: @lshr_C1_add_A_C2_v4i32_splat( ; CHECK-NEXT: [[A:%.*]] = zext i16 [[I:%.*]] to i32 ; CHECK-NEXT: [[B:%.*]] = insertelement <4 x i32> poison, i32 [[A]], i64 0 ; CHECK-NEXT: [[C:%.*]] = shufflevector <4 x i32> [[B]], <4 x i32> poison, <4 x i32> zeroinitializer ; CHECK-NEXT: [[E:%.*]] = lshr <4 x i32> , [[C]] ; CHECK-NEXT: ret <4 x i32> [[E]] ; %A = zext i16 %I to i32 %B = insertelement <4 x i32> undef, i32 %A, i32 0 %C = shufflevector <4 x i32> %B, <4 x i32> undef, <4 x i32> zeroinitializer %D = add <4 x i32> %C, %E = lshr <4 x i32> , %D ret <4 x i32> %E } define i32 @shl_add_nuw(i32 %x) { ; CHECK-LABEL: @shl_add_nuw( ; CHECK-NEXT: [[R:%.*]] = shl i32 192, [[X:%.*]] ; CHECK-NEXT: ret i32 [[R]] ; %a = add nuw i32 %x, 5 %r = shl i32 6, %a ret i32 %r } ; vectors with arbitrary constants work too define <2 x i12> @lshr_add_nuw(<2 x i12> %x) { ; CHECK-LABEL: @lshr_add_nuw( ; CHECK-NEXT: [[R:%.*]] = lshr <2 x i12> , [[X:%.*]] ; CHECK-NEXT: ret <2 x i12> [[R]] ; %a = add nuw <2 x i12> %x, %r = lshr <2 x i12> , %a ret <2 x i12> %r } ; extra use is ok and in this case the result can be simplified to a constant define i32 @ashr_add_nuw(i32 %x, ptr %p) { ; CHECK-LABEL: @ashr_add_nuw( ; CHECK-NEXT: [[A:%.*]] = add nuw i32 [[X:%.*]], 5 ; CHECK-NEXT: store i32 [[A]], ptr [[P:%.*]], align 4 ; CHECK-NEXT: ret i32 -1 ; %a = add nuw i32 %x, 5 store i32 %a, ptr %p %r = ashr i32 -6, %a ret i32 %r } ; negative test - must have 'nuw' define i32 @shl_add_nsw(i32 %x) { ; CHECK-LABEL: @shl_add_nsw( ; CHECK-NEXT: [[A:%.*]] = add nsw i32 [[X:%.*]], 5 ; CHECK-NEXT: [[R:%.*]] = shl i32 6, [[A]] ; CHECK-NEXT: ret i32 [[R]] ; %a = add nsw i32 %x, 5 %r = shl i32 6, %a ret i32 %r } ; offset precondition check (must be negative constant) for lshr_exact_add_negative_shift_positive define i32 @lshr_exact_add_positive_shift_positive(i32 %x) { ; CHECK-LABEL: @lshr_exact_add_positive_shift_positive( ; CHECK-NEXT: [[A:%.*]] = add i32 [[X:%.*]], 1 ; CHECK-NEXT: [[R:%.*]] = lshr exact i32 2, [[A]] ; CHECK-NEXT: ret i32 [[R]] ; %a = add i32 %x, 1 %r = lshr exact i32 2, %a ret i32 %r } define i32 @lshr_exact_add_big_negative_offset(i32 %x) { ; CHECK-LABEL: @lshr_exact_add_big_negative_offset( ; CHECK-NEXT: [[A:%.*]] = add i32 [[X:%.*]], -33 ; CHECK-NEXT: [[R:%.*]] = lshr exact i32 2, [[A]] ; CHECK-NEXT: ret i32 [[R]] ; %a = add i32 %x, -33 %r = lshr exact i32 2, %a ret i32 %r } ; leading zeros for shifted constant precondition check for lshr_exact_add_negative_shift_positive define i32 @lshr_exact_add_negative_shift_negative(i32 %x) { ; CHECK-LABEL: @lshr_exact_add_negative_shift_negative( ; CHECK-NEXT: [[A:%.*]] = add i32 [[X:%.*]], -1 ; CHECK-NEXT: [[R:%.*]] = lshr exact i32 -2, [[A]] ; CHECK-NEXT: ret i32 [[R]] ; %a = add i32 %x, -1 %r = lshr exact i32 -2, %a ret i32 %r } ; exact precondition check for lshr_exact_add_negative_shift_positive define i32 @lshr_add_negative_shift_no_exact(i32 %x) { ; CHECK-LABEL: @lshr_add_negative_shift_no_exact( ; CHECK-NEXT: [[A:%.*]] = add i32 [[X:%.*]], -1 ; CHECK-NEXT: [[R:%.*]] = lshr i32 2, [[A]] ; CHECK-NEXT: ret i32 [[R]] ; %a = add i32 %x, -1 %r = lshr i32 2, %a ret i32 %r } define i32 @lshr_exact_add_negative_shift_positive(i32 %x) { ; CHECK-LABEL: @lshr_exact_add_negative_shift_positive( ; CHECK-NEXT: [[R:%.*]] = lshr exact i32 4, [[X:%.*]] ; CHECK-NEXT: ret i32 [[R]] ; %a = add i32 %x, -1 %r = lshr exact i32 2, %a ret i32 %r } define i8 @lshr_exact_add_negative_shift_positive_extra_use(i8 %x) { ; CHECK-LABEL: @lshr_exact_add_negative_shift_positive_extra_use( ; CHECK-NEXT: [[A:%.*]] = add i8 [[X:%.*]], -1 ; CHECK-NEXT: call void @use(i8 [[A]]) ; CHECK-NEXT: [[R:%.*]] = lshr exact i8 -128, [[X]] ; CHECK-NEXT: ret i8 [[R]] ; %a = add i8 %x, -1 call void @use(i8 %a) %r = lshr exact i8 64, %a ret i8 %r } define <2 x i9> @lshr_exact_add_negative_shift_positive_vec(<2 x i9> %x) { ; CHECK-LABEL: @lshr_exact_add_negative_shift_positive_vec( ; CHECK-NEXT: [[R:%.*]] = lshr exact <2 x i9> , [[X:%.*]] ; CHECK-NEXT: ret <2 x i9> [[R]] ; %a = add <2 x i9> %x, %r = lshr exact <2 x i9> , %a ret <2 x i9> %r } ; not enough leading zeros in shift constant define <2 x i9> @lshr_exact_add_negative_shift_lzcnt(<2 x i9> %x) { ; CHECK-LABEL: @lshr_exact_add_negative_shift_lzcnt( ; CHECK-NEXT: [[A:%.*]] = add <2 x i9> [[X:%.*]], ; CHECK-NEXT: [[R:%.*]] = lshr exact <2 x i9> , [[A]] ; CHECK-NEXT: ret <2 x i9> [[R]] ; %a = add <2 x i9> %x, %r = lshr exact <2 x i9> , %a ret <2 x i9> %r } ; leading ones precondition check for ashr_exact_add_negative_shift_[positive,negative] define i8 @ashr_exact_add_negative_shift_no_trailing_zeros(i8 %x) { ; CHECK-LABEL: @ashr_exact_add_negative_shift_no_trailing_zeros( ; CHECK-NEXT: [[A:%.*]] = add i8 [[X:%.*]], -4 ; CHECK-NEXT: [[R:%.*]] = ashr exact i8 -112, [[A]] ; CHECK-NEXT: ret i8 [[R]] ; %a = add i8 %x, -4 %r = ashr exact i8 -112, %a ; 0b1001_0000 ret i8 %r } define i32 @ashr_exact_add_big_negative_offset(i32 %x) { ; CHECK-LABEL: @ashr_exact_add_big_negative_offset( ; CHECK-NEXT: [[A:%.*]] = add i32 [[X:%.*]], -33 ; CHECK-NEXT: [[R:%.*]] = ashr exact i32 -2, [[A]] ; CHECK-NEXT: ret i32 [[R]] ; %a = add i32 %x, -33 %r = ashr exact i32 -2, %a ret i32 %r } ; exact precondition check for ashr_exact_add_negative_shift_[positive,negative] define i32 @ashr_add_negative_shift_no_exact(i32 %x) { ; CHECK-LABEL: @ashr_add_negative_shift_no_exact( ; CHECK-NEXT: [[A:%.*]] = add i32 [[X:%.*]], -1 ; CHECK-NEXT: [[R:%.*]] = ashr i32 -2, [[A]] ; CHECK-NEXT: ret i32 [[R]] ; %a = add i32 %x, -1 %r = ashr i32 -2, %a ret i32 %r } define i32 @ashr_exact_add_negative_shift_negative(i32 %x) { ; CHECK-LABEL: @ashr_exact_add_negative_shift_negative( ; CHECK-NEXT: [[R:%.*]] = ashr exact i32 -4, [[X:%.*]] ; CHECK-NEXT: ret i32 [[R]] ; %a = add i32 %x, -1 %r = ashr exact i32 -2, %a ret i32 %r } define i8 @ashr_exact_add_negative_shift_negative_extra_use(i8 %x) { ; CHECK-LABEL: @ashr_exact_add_negative_shift_negative_extra_use( ; CHECK-NEXT: [[A:%.*]] = add i8 [[X:%.*]], -2 ; CHECK-NEXT: call void @use(i8 [[A]]) ; CHECK-NEXT: [[R:%.*]] = ashr exact i8 -128, [[X]] ; CHECK-NEXT: ret i8 [[R]] ; %a = add i8 %x, -2 call void @use(i8 %a) %r = ashr exact i8 -32, %a ret i8 %r } define <2 x i7> @ashr_exact_add_negative_shift_negative_vec(<2 x i7> %x) { ; CHECK-LABEL: @ashr_exact_add_negative_shift_negative_vec( ; CHECK-NEXT: [[R:%.*]] = ashr exact <2 x i7> , [[X:%.*]] ; CHECK-NEXT: ret <2 x i7> [[R]] ; %a = add <2 x i7> %x, %r = ashr exact <2 x i7> , %a ret <2 x i7> %r } ; not enough leading ones in shift constant define <2 x i7> @ashr_exact_add_negative_leading_ones_vec(<2 x i7> %x) { ; CHECK-LABEL: @ashr_exact_add_negative_leading_ones_vec( ; CHECK-NEXT: [[A:%.*]] = add <2 x i7> [[X:%.*]], ; CHECK-NEXT: [[R:%.*]] = ashr exact <2 x i7> , [[A]] ; CHECK-NEXT: ret <2 x i7> [[R]] ; %a = add <2 x i7> %x, %r = ashr exact <2 x i7> , %a ret <2 x i7> %r } ; PR54890 define i32 @shl_nsw_add_negative(i32 %x) { ; CHECK-LABEL: @shl_nsw_add_negative( ; CHECK-NEXT: [[R:%.*]] = shl nuw i32 1, [[X:%.*]] ; CHECK-NEXT: ret i32 [[R]] ; %a = add i32 %x, -1 %r = shl nsw i32 2, %a ret i32 %r } ; vectors and extra uses are allowed ; nuw propagates to the new shift define <2 x i8> @shl_nuw_add_negative_splat_uses(<2 x i8> %x, ptr %p) { ; CHECK-LABEL: @shl_nuw_add_negative_splat_uses( ; CHECK-NEXT: [[A:%.*]] = add <2 x i8> [[X:%.*]], ; CHECK-NEXT: store <2 x i8> [[A]], ptr [[P:%.*]], align 2 ; CHECK-NEXT: [[R:%.*]] = shl nuw <2 x i8> , [[X]] ; CHECK-NEXT: ret <2 x i8> [[R]] ; %a = add <2 x i8> %x, store <2 x i8> %a, ptr %p %r = shl nuw <2 x i8> , %a ret <2 x i8> %r } ; negative test - shift constant must have enough trailing zeros to allow the pre-shift define i32 @shl_nsw_add_negative_invalid_constant(i32 %x) { ; CHECK-LABEL: @shl_nsw_add_negative_invalid_constant( ; CHECK-NEXT: [[A:%.*]] = add i32 [[X:%.*]], -2 ; CHECK-NEXT: [[R:%.*]] = shl nsw i32 2, [[A]] ; CHECK-NEXT: ret i32 [[R]] ; %a = add i32 %x, -2 %r = shl nsw i32 2, %a ret i32 %r } ; negative test - the offset constant must be negative define i32 @shl_nsw_add_positive_invalid_constant(i32 %x) { ; CHECK-LABEL: @shl_nsw_add_positive_invalid_constant( ; CHECK-NEXT: [[A:%.*]] = add i32 [[X:%.*]], 2 ; CHECK-NEXT: [[R:%.*]] = shl nsw i32 4, [[A]] ; CHECK-NEXT: ret i32 [[R]] ; %a = add i32 %x, 2 %r = shl nsw i32 4, %a ret i32 %r } ; negative test - a large shift must be detected without crashing define i32 @shl_nsw_add_negative_invalid_constant2(i32 %x) { ; CHECK-LABEL: @shl_nsw_add_negative_invalid_constant2( ; CHECK-NEXT: [[A:%.*]] = add i32 [[X:%.*]], -33 ; CHECK-NEXT: [[R:%.*]] = shl nsw i32 2, [[A]] ; CHECK-NEXT: ret i32 [[R]] ; %a = add i32 %x, -33 %r = shl nsw i32 2, %a ret i32 %r } ; negative test - currently transformed to 'xor' before we see it, ; but INT_MIN should be handled too define i4 @shl_nsw_add_negative_invalid_constant3(i4 %x) { ; CHECK-LABEL: @shl_nsw_add_negative_invalid_constant3( ; CHECK-NEXT: [[A:%.*]] = xor i4 [[X:%.*]], -8 ; CHECK-NEXT: [[R:%.*]] = shl nsw i4 2, [[A]] ; CHECK-NEXT: ret i4 [[R]] ; %a = add i4 %x, 8 %r = shl nsw i4 2, %a ret i4 %r } define i2 @lshr_2_add_zext_basic(i1 %a, i1 %b) { ; CHECK-LABEL: @lshr_2_add_zext_basic( ; CHECK-NEXT: [[TMP1:%.*]] = and i1 [[A:%.*]], [[B:%.*]] ; CHECK-NEXT: [[LSHR:%.*]] = zext i1 [[TMP1]] to i2 ; CHECK-NEXT: ret i2 [[LSHR]] ; %zext.a = zext i1 %a to i2 %zext.b = zext i1 %b to i2 %add = add i2 %zext.a, %zext.b %lshr = lshr i2 %add, 1 ret i2 %lshr } define i2 @ashr_2_add_zext_basic(i1 %a, i1 %b) { ; CHECK-LABEL: @ashr_2_add_zext_basic( ; CHECK-NEXT: [[ZEXT_A:%.*]] = zext i1 [[A:%.*]] to i2 ; CHECK-NEXT: [[ZEXT_B:%.*]] = zext i1 [[B:%.*]] to i2 ; CHECK-NEXT: [[ADD:%.*]] = add nuw i2 [[ZEXT_A]], [[ZEXT_B]] ; CHECK-NEXT: [[LSHR:%.*]] = ashr i2 [[ADD]], 1 ; CHECK-NEXT: ret i2 [[LSHR]] ; %zext.a = zext i1 %a to i2 %zext.b = zext i1 %b to i2 %add = add i2 %zext.a, %zext.b %lshr = ashr i2 %add, 1 ret i2 %lshr } define i32 @lshr_16_add_zext_basic(i16 %a, i16 %b) { ; CHECK-LABEL: @lshr_16_add_zext_basic( ; CHECK-NEXT: [[TMP1:%.*]] = xor i16 [[A:%.*]], -1 ; CHECK-NEXT: [[ADD_NARROWED_OVERFLOW:%.*]] = icmp ult i16 [[TMP1]], [[B:%.*]] ; CHECK-NEXT: [[LSHR:%.*]] = zext i1 [[ADD_NARROWED_OVERFLOW]] to i32 ; CHECK-NEXT: ret i32 [[LSHR]] ; %zext.a = zext i16 %a to i32 %zext.b = zext i16 %b to i32 %add = add i32 %zext.a, %zext.b %lshr = lshr i32 %add, 16 ret i32 %lshr } define i32 @lshr_16_add_zext_basic_multiuse(i16 %a, i16 %b) { ; CHECK-LABEL: @lshr_16_add_zext_basic_multiuse( ; CHECK-NEXT: [[ZEXT_A:%.*]] = zext i16 [[A:%.*]] to i32 ; CHECK-NEXT: [[ZEXT_B:%.*]] = zext i16 [[B:%.*]] to i32 ; CHECK-NEXT: [[ADD:%.*]] = add nuw nsw i32 [[ZEXT_A]], [[ZEXT_B]] ; CHECK-NEXT: [[LSHR:%.*]] = lshr i32 [[ADD]], 16 ; CHECK-NEXT: [[OTHERUSE:%.*]] = or i32 [[LSHR]], [[ZEXT_A]] ; CHECK-NEXT: ret i32 [[OTHERUSE]] ; %zext.a = zext i16 %a to i32 %zext.b = zext i16 %b to i32 %add = add i32 %zext.a, %zext.b %lshr = lshr i32 %add, 16 %otheruse = or i32 %lshr, %zext.a ret i32 %otheruse } define i32 @lshr_16_add_known_16_leading_zeroes(i32 %a, i32 %b) { ; CHECK-LABEL: @lshr_16_add_known_16_leading_zeroes( ; CHECK-NEXT: [[A16:%.*]] = and i32 [[A:%.*]], 65535 ; CHECK-NEXT: [[B16:%.*]] = and i32 [[B:%.*]], 65535 ; CHECK-NEXT: [[ADD:%.*]] = add nuw nsw i32 [[A16]], [[B16]] ; CHECK-NEXT: [[LSHR:%.*]] = lshr i32 [[ADD]], 16 ; CHECK-NEXT: ret i32 [[LSHR]] ; %a16 = and i32 %a, 65535 ; 0x65535 %b16 = and i32 %b, 65535 ; 0x65535 %add = add i32 %a16, %b16 %lshr = lshr i32 %add, 16 ret i32 %lshr } define i32 @lshr_16_add_not_known_16_leading_zeroes(i32 %a, i32 %b) { ; CHECK-LABEL: @lshr_16_add_not_known_16_leading_zeroes( ; CHECK-NEXT: [[A16:%.*]] = and i32 [[A:%.*]], 131071 ; CHECK-NEXT: [[B16:%.*]] = and i32 [[B:%.*]], 65535 ; CHECK-NEXT: [[ADD:%.*]] = add nuw nsw i32 [[A16]], [[B16]] ; CHECK-NEXT: [[LSHR:%.*]] = lshr i32 [[ADD]], 16 ; CHECK-NEXT: ret i32 [[LSHR]] ; %a16 = and i32 %a, 131071 ; 0x1FFFF %b16 = and i32 %b, 65535 ; 0x65535 %add = add i32 %a16, %b16 %lshr = lshr i32 %add, 16 ret i32 %lshr } define i64 @lshr_32_add_zext_basic(i32 %a, i32 %b) { ; CHECK-LABEL: @lshr_32_add_zext_basic( ; CHECK-NEXT: [[TMP1:%.*]] = xor i32 [[A:%.*]], -1 ; CHECK-NEXT: [[ADD_NARROWED_OVERFLOW:%.*]] = icmp ult i32 [[TMP1]], [[B:%.*]] ; CHECK-NEXT: [[LSHR:%.*]] = zext i1 [[ADD_NARROWED_OVERFLOW]] to i64 ; CHECK-NEXT: ret i64 [[LSHR]] ; %zext.a = zext i32 %a to i64 %zext.b = zext i32 %b to i64 %add = add i64 %zext.a, %zext.b %lshr = lshr i64 %add, 32 ret i64 %lshr } define i64 @lshr_32_add_zext_basic_multiuse(i32 %a, i32 %b) { ; CHECK-LABEL: @lshr_32_add_zext_basic_multiuse( ; CHECK-NEXT: [[ZEXT_A:%.*]] = zext i32 [[A:%.*]] to i64 ; CHECK-NEXT: [[ZEXT_B:%.*]] = zext i32 [[B:%.*]] to i64 ; CHECK-NEXT: [[ADD:%.*]] = add nuw nsw i64 [[ZEXT_A]], [[ZEXT_B]] ; CHECK-NEXT: [[LSHR:%.*]] = lshr i64 [[ADD]], 32 ; CHECK-NEXT: [[OTHERUSE:%.*]] = or i64 [[LSHR]], [[ZEXT_B]] ; CHECK-NEXT: ret i64 [[OTHERUSE]] ; %zext.a = zext i32 %a to i64 %zext.b = zext i32 %b to i64 %add = add i64 %zext.a, %zext.b %lshr = lshr i64 %add, 32 %otheruse = or i64 %lshr, %zext.b ret i64 %otheruse } define i64 @lshr_31_i32_add_zext_basic(i32 %a, i32 %b) { ; CHECK-LABEL: @lshr_31_i32_add_zext_basic( ; CHECK-NEXT: [[ZEXT_A:%.*]] = zext i32 [[A:%.*]] to i64 ; CHECK-NEXT: [[ZEXT_B:%.*]] = zext i32 [[B:%.*]] to i64 ; CHECK-NEXT: [[ADD:%.*]] = add nuw nsw i64 [[ZEXT_A]], [[ZEXT_B]] ; CHECK-NEXT: [[LSHR:%.*]] = lshr i64 [[ADD]], 31 ; CHECK-NEXT: ret i64 [[LSHR]] ; %zext.a = zext i32 %a to i64 %zext.b = zext i32 %b to i64 %add = add i64 %zext.a, %zext.b %lshr = lshr i64 %add, 31 ret i64 %lshr } define i64 @lshr_33_i32_add_zext_basic(i32 %a, i32 %b) { ; CHECK-LABEL: @lshr_33_i32_add_zext_basic( ; CHECK-NEXT: ret i64 0 ; %zext.a = zext i32 %a to i64 %zext.b = zext i32 %b to i64 %add = add i64 %zext.a, %zext.b %lshr = lshr i64 %add, 33 ret i64 %lshr } define i64 @lshr_16_to_64_add_zext_basic(i16 %a, i16 %b) { ; CHECK-LABEL: @lshr_16_to_64_add_zext_basic( ; CHECK-NEXT: [[TMP1:%.*]] = xor i16 [[A:%.*]], -1 ; CHECK-NEXT: [[ADD_NARROWED_OVERFLOW:%.*]] = icmp ult i16 [[TMP1]], [[B:%.*]] ; CHECK-NEXT: [[LSHR:%.*]] = zext i1 [[ADD_NARROWED_OVERFLOW]] to i64 ; CHECK-NEXT: ret i64 [[LSHR]] ; %zext.a = zext i16 %a to i64 %zext.b = zext i16 %b to i64 %add = add i64 %zext.a, %zext.b %lshr = lshr i64 %add, 16 ret i64 %lshr } define i64 @lshr_32_add_known_32_leading_zeroes(i64 %a, i64 %b) { ; CHECK-LABEL: @lshr_32_add_known_32_leading_zeroes( ; CHECK-NEXT: [[A32:%.*]] = and i64 [[A:%.*]], 4294967295 ; CHECK-NEXT: [[B32:%.*]] = and i64 [[B:%.*]], 4294967295 ; CHECK-NEXT: [[ADD:%.*]] = add nuw nsw i64 [[A32]], [[B32]] ; CHECK-NEXT: [[LSHR:%.*]] = lshr i64 [[ADD]], 32 ; CHECK-NEXT: ret i64 [[LSHR]] ; %a32 = and i64 %a, 4294967295 ; 0xFFFFFFFF %b32 = and i64 %b, 4294967295 ; 0xFFFFFFFF %add = add i64 %a32, %b32 %lshr = lshr i64 %add, 32 ret i64 %lshr } define i64 @lshr_32_add_not_known_32_leading_zeroes(i64 %a, i64 %b) { ; ; CHECK-LABEL: @lshr_32_add_not_known_32_leading_zeroes( ; CHECK-NEXT: [[A32:%.*]] = and i64 [[A:%.*]], 8589934591 ; CHECK-NEXT: [[B32:%.*]] = and i64 [[B:%.*]], 4294967295 ; CHECK-NEXT: [[ADD:%.*]] = add nuw nsw i64 [[A32]], [[B32]] ; CHECK-NEXT: [[LSHR:%.*]] = lshr i64 [[ADD]], 32 ; CHECK-NEXT: ret i64 [[LSHR]] ; %a32 = and i64 %a, 8589934591 ; 0x1FFFFFFFF %b32 = and i64 %b, 4294967295 ; 0xFFFFFFFF %add = add i64 %a32, %b32 %lshr = lshr i64 %add, 32 ret i64 %lshr } define i32 @ashr_16_add_zext_basic(i16 %a, i16 %b) { ; CHECK-LABEL: @ashr_16_add_zext_basic( ; CHECK-NEXT: [[TMP1:%.*]] = xor i16 [[A:%.*]], -1 ; CHECK-NEXT: [[ADD_NARROWED_OVERFLOW:%.*]] = icmp ult i16 [[TMP1]], [[B:%.*]] ; CHECK-NEXT: [[LSHR:%.*]] = zext i1 [[ADD_NARROWED_OVERFLOW]] to i32 ; CHECK-NEXT: ret i32 [[LSHR]] ; %zext.a = zext i16 %a to i32 %zext.b = zext i16 %b to i32 %add = add i32 %zext.a, %zext.b %lshr = lshr i32 %add, 16 ret i32 %lshr } define i64 @ashr_32_add_zext_basic(i32 %a, i32 %b) { ; CHECK-LABEL: @ashr_32_add_zext_basic( ; CHECK-NEXT: [[TMP1:%.*]] = xor i32 [[A:%.*]], -1 ; CHECK-NEXT: [[ADD_NARROWED_OVERFLOW:%.*]] = icmp ult i32 [[TMP1]], [[B:%.*]] ; CHECK-NEXT: [[LSHR:%.*]] = zext i1 [[ADD_NARROWED_OVERFLOW]] to i64 ; CHECK-NEXT: ret i64 [[LSHR]] ; %zext.a = zext i32 %a to i64 %zext.b = zext i32 %b to i64 %add = add i64 %zext.a, %zext.b %lshr = ashr i64 %add, 32 ret i64 %lshr } define i64 @ashr_16_to_64_add_zext_basic(i16 %a, i16 %b) { ; CHECK-LABEL: @ashr_16_to_64_add_zext_basic( ; CHECK-NEXT: [[TMP1:%.*]] = xor i16 [[A:%.*]], -1 ; CHECK-NEXT: [[ADD_NARROWED_OVERFLOW:%.*]] = icmp ult i16 [[TMP1]], [[B:%.*]] ; CHECK-NEXT: [[LSHR:%.*]] = zext i1 [[ADD_NARROWED_OVERFLOW]] to i64 ; CHECK-NEXT: ret i64 [[LSHR]] ; %zext.a = zext i16 %a to i64 %zext.b = zext i16 %b to i64 %add = add i64 %zext.a, %zext.b %lshr = ashr i64 %add, 16 ret i64 %lshr } define i32 @lshr_32_add_zext_trunc(i32 %a, i32 %b) { ; CHECK-LABEL: @lshr_32_add_zext_trunc( ; CHECK-NEXT: [[ADD_NARROWED:%.*]] = add i32 [[A:%.*]], [[B:%.*]] ; CHECK-NEXT: [[ADD_NARROWED_OVERFLOW:%.*]] = icmp ult i32 [[ADD_NARROWED]], [[A]] ; CHECK-NEXT: [[TRUNC_SHR:%.*]] = zext i1 [[ADD_NARROWED_OVERFLOW]] to i32 ; CHECK-NEXT: [[RET:%.*]] = add i32 [[ADD_NARROWED]], [[TRUNC_SHR]] ; CHECK-NEXT: ret i32 [[RET]] ; %zext.a = zext i32 %a to i64 %zext.b = zext i32 %b to i64 %add = add i64 %zext.a, %zext.b %trunc.add = trunc i64 %add to i32 %shr = lshr i64 %add, 32 %trunc.shr = trunc i64 %shr to i32 %ret = add i32 %trunc.add, %trunc.shr ret i32 %ret } define <3 x i32> @add3_i96(<3 x i32> %0, <3 x i32> %1) { ; CHECK-LABEL: @add3_i96( ; CHECK-NEXT: [[TMP3:%.*]] = extractelement <3 x i32> [[TMP0:%.*]], i64 0 ; CHECK-NEXT: [[TMP4:%.*]] = extractelement <3 x i32> [[TMP1:%.*]], i64 0 ; CHECK-NEXT: [[ADD_NARROWED:%.*]] = add i32 [[TMP4]], [[TMP3]] ; CHECK-NEXT: [[ADD_NARROWED_OVERFLOW:%.*]] = icmp ult i32 [[ADD_NARROWED]], [[TMP4]] ; CHECK-NEXT: [[TMP5:%.*]] = extractelement <3 x i32> [[TMP0]], i64 1 ; CHECK-NEXT: [[TMP6:%.*]] = zext i32 [[TMP5]] to i64 ; CHECK-NEXT: [[TMP7:%.*]] = extractelement <3 x i32> [[TMP1]], i64 1 ; CHECK-NEXT: [[TMP8:%.*]] = zext i32 [[TMP7]] to i64 ; CHECK-NEXT: [[TMP9:%.*]] = add nuw nsw i64 [[TMP8]], [[TMP6]] ; CHECK-NEXT: [[TMP10:%.*]] = zext i1 [[ADD_NARROWED_OVERFLOW]] to i64 ; CHECK-NEXT: [[TMP11:%.*]] = add nuw nsw i64 [[TMP9]], [[TMP10]] ; CHECK-NEXT: [[TMP12:%.*]] = extractelement <3 x i32> [[TMP0]], i64 2 ; CHECK-NEXT: [[TMP13:%.*]] = extractelement <3 x i32> [[TMP1]], i64 2 ; CHECK-NEXT: [[TMP14:%.*]] = add i32 [[TMP13]], [[TMP12]] ; CHECK-NEXT: [[TMP15:%.*]] = lshr i64 [[TMP11]], 32 ; CHECK-NEXT: [[TMP16:%.*]] = trunc i64 [[TMP15]] to i32 ; CHECK-NEXT: [[TMP17:%.*]] = add i32 [[TMP14]], [[TMP16]] ; CHECK-NEXT: [[TMP18:%.*]] = insertelement <3 x i32> poison, i32 [[ADD_NARROWED]], i64 0 ; CHECK-NEXT: [[TMP19:%.*]] = trunc i64 [[TMP11]] to i32 ; CHECK-NEXT: [[TMP20:%.*]] = insertelement <3 x i32> [[TMP18]], i32 [[TMP19]], i64 1 ; CHECK-NEXT: [[TMP21:%.*]] = insertelement <3 x i32> [[TMP20]], i32 [[TMP17]], i64 2 ; CHECK-NEXT: ret <3 x i32> [[TMP21]] ; %3 = extractelement <3 x i32> %0, i64 0 %4 = zext i32 %3 to i64 %5 = extractelement <3 x i32> %1, i64 0 %6 = zext i32 %5 to i64 %7 = add nuw nsw i64 %6, %4 %8 = extractelement <3 x i32> %0, i64 1 %9 = zext i32 %8 to i64 %10 = extractelement <3 x i32> %1, i64 1 %11 = zext i32 %10 to i64 %12 = add nuw nsw i64 %11, %9 %13 = lshr i64 %7, 32 %14 = add nuw nsw i64 %12, %13 %15 = extractelement <3 x i32> %0, i64 2 %16 = extractelement <3 x i32> %1, i64 2 %17 = add i32 %16, %15 %18 = lshr i64 %14, 32 %19 = trunc i64 %18 to i32 %20 = add i32 %17, %19 %21 = trunc i64 %7 to i32 %22 = insertelement <3 x i32> undef, i32 %21, i32 0 %23 = trunc i64 %14 to i32 %24 = insertelement <3 x i32> %22, i32 %23, i32 1 %25 = insertelement <3 x i32> %24, i32 %20, i32 2 ret <3 x i32> %25 }