| ; NOTE: Assertions have been autogenerated by utils/update_test_checks.py |
| ; RUN: opt < %s -passes=instcombine -S | FileCheck %s |
| |
| declare void @use(i64) |
| |
| define i8 @shl_and(i8 %x, i8 %y) { |
| ; CHECK-LABEL: @shl_and( |
| ; CHECK-NEXT: [[TMP1:%.*]] = shl i8 [[X:%.*]], 5 |
| ; CHECK-NEXT: [[TMP2:%.*]] = shl i8 [[Y:%.*]], 2 |
| ; CHECK-NEXT: [[SH1:%.*]] = and i8 [[TMP1]], [[TMP2]] |
| ; CHECK-NEXT: ret i8 [[SH1]] |
| ; |
| %sh0 = shl i8 %x, 3 |
| %r = and i8 %sh0, %y |
| %sh1 = shl i8 %r, 2 |
| ret i8 %sh1 |
| } |
| |
| define <2 x i8> @shl_and_nonuniform(<2 x i8> %x, <2 x i8> %y) { |
| ; CHECK-LABEL: @shl_and_nonuniform( |
| ; CHECK-NEXT: [[TMP1:%.*]] = shl <2 x i8> [[X:%.*]], <i8 5, i8 4> |
| ; CHECK-NEXT: [[TMP2:%.*]] = shl <2 x i8> [[Y:%.*]], <i8 2, i8 0> |
| ; CHECK-NEXT: [[SH1:%.*]] = and <2 x i8> [[TMP1]], [[TMP2]] |
| ; CHECK-NEXT: ret <2 x i8> [[SH1]] |
| ; |
| %sh0 = shl <2 x i8> %x, <i8 3, i8 4> |
| %r = and <2 x i8> %sh0, %y |
| %sh1 = shl <2 x i8> %r, <i8 2, i8 0> |
| ret <2 x i8> %sh1 |
| } |
| |
| define i16 @shl_or(i16 %x, i16 %py) { |
| ; CHECK-LABEL: @shl_or( |
| ; CHECK-NEXT: [[Y:%.*]] = srem i16 [[PY:%.*]], 42 |
| ; CHECK-NEXT: [[TMP1:%.*]] = shl i16 [[X:%.*]], 12 |
| ; CHECK-NEXT: [[TMP2:%.*]] = shl nsw i16 [[Y]], 7 |
| ; CHECK-NEXT: [[SH1:%.*]] = or i16 [[TMP1]], [[TMP2]] |
| ; CHECK-NEXT: ret i16 [[SH1]] |
| ; |
| %y = srem i16 %py, 42 ; thwart complexity-based canonicalization |
| %sh0 = shl i16 %x, 5 |
| %r = or i16 %y, %sh0 |
| %sh1 = shl i16 %r, 7 |
| ret i16 %sh1 |
| } |
| |
| define <2 x i16> @shl_or_undef(<2 x i16> %x, <2 x i16> %py) { |
| ; CHECK-LABEL: @shl_or_undef( |
| ; CHECK-NEXT: [[Y:%.*]] = srem <2 x i16> [[PY:%.*]], <i16 42, i16 42> |
| ; CHECK-NEXT: [[TMP1:%.*]] = shl <2 x i16> [[X:%.*]], <i16 12, i16 undef> |
| ; CHECK-NEXT: [[TMP2:%.*]] = shl <2 x i16> [[Y]], <i16 7, i16 undef> |
| ; CHECK-NEXT: [[SH1:%.*]] = or <2 x i16> [[TMP1]], [[TMP2]] |
| ; CHECK-NEXT: ret <2 x i16> [[SH1]] |
| ; |
| %y = srem <2 x i16> %py, <i16 42, i16 42> ; thwart complexity-based canonicalization |
| %sh0 = shl <2 x i16> %x, <i16 5, i16 undef> |
| %r = or <2 x i16> %y, %sh0 |
| %sh1 = shl <2 x i16> %r, <i16 7, i16 undef> |
| ret <2 x i16> %sh1 |
| } |
| |
| define i32 @shl_xor(i32 %x, i32 %y) { |
| ; CHECK-LABEL: @shl_xor( |
| ; CHECK-NEXT: [[TMP1:%.*]] = shl i32 [[X:%.*]], 12 |
| ; CHECK-NEXT: [[TMP2:%.*]] = shl i32 [[Y:%.*]], 7 |
| ; CHECK-NEXT: [[SH1:%.*]] = xor i32 [[TMP1]], [[TMP2]] |
| ; CHECK-NEXT: ret i32 [[SH1]] |
| ; |
| %sh0 = shl i32 %x, 5 |
| %r = xor i32 %sh0, %y |
| %sh1 = shl i32 %r, 7 |
| ret i32 %sh1 |
| } |
| |
| define <2 x i32> @shl_xor_nonuniform(<2 x i32> %x, <2 x i32> %y) { |
| ; CHECK-LABEL: @shl_xor_nonuniform( |
| ; CHECK-NEXT: [[TMP1:%.*]] = shl <2 x i32> [[X:%.*]], <i32 12, i32 14> |
| ; CHECK-NEXT: [[TMP2:%.*]] = shl <2 x i32> [[Y:%.*]], <i32 7, i32 8> |
| ; CHECK-NEXT: [[SH1:%.*]] = xor <2 x i32> [[TMP1]], [[TMP2]] |
| ; CHECK-NEXT: ret <2 x i32> [[SH1]] |
| ; |
| %sh0 = shl <2 x i32> %x, <i32 5, i32 6> |
| %r = xor <2 x i32> %sh0, %y |
| %sh1 = shl <2 x i32> %r, <i32 7, i32 8> |
| ret <2 x i32> %sh1 |
| } |
| |
| define i64 @lshr_and(i64 %x, i64 %py) { |
| ; CHECK-LABEL: @lshr_and( |
| ; CHECK-NEXT: [[Y:%.*]] = srem i64 [[PY:%.*]], 42 |
| ; CHECK-NEXT: [[TMP1:%.*]] = lshr i64 [[X:%.*]], 12 |
| ; CHECK-NEXT: [[TMP2:%.*]] = lshr i64 [[Y]], 7 |
| ; CHECK-NEXT: [[SH1:%.*]] = and i64 [[TMP1]], [[TMP2]] |
| ; CHECK-NEXT: ret i64 [[SH1]] |
| ; |
| %y = srem i64 %py, 42 ; thwart complexity-based canonicalization |
| %sh0 = lshr i64 %x, 5 |
| %r = and i64 %y, %sh0 |
| %sh1 = lshr i64 %r, 7 |
| ret i64 %sh1 |
| } |
| |
| define <2 x i64> @lshr_and_undef(<2 x i64> %x, <2 x i64> %py) { |
| ; CHECK-LABEL: @lshr_and_undef( |
| ; CHECK-NEXT: [[Y:%.*]] = srem <2 x i64> [[PY:%.*]], <i64 42, i64 42> |
| ; CHECK-NEXT: [[TMP1:%.*]] = lshr <2 x i64> [[X:%.*]], <i64 12, i64 undef> |
| ; CHECK-NEXT: [[TMP2:%.*]] = lshr <2 x i64> [[Y]], <i64 7, i64 undef> |
| ; CHECK-NEXT: [[SH1:%.*]] = and <2 x i64> [[TMP1]], [[TMP2]] |
| ; CHECK-NEXT: ret <2 x i64> [[SH1]] |
| ; |
| %y = srem <2 x i64> %py, <i64 42, i64 42> ; thwart complexity-based canonicalization |
| %sh0 = lshr <2 x i64> %x, <i64 5, i64 undef> |
| %r = and <2 x i64> %y, %sh0 |
| %sh1 = lshr <2 x i64> %r, <i64 7, i64 undef> |
| ret <2 x i64> %sh1 |
| } |
| |
| define <4 x i32> @lshr_or(<4 x i32> %x, <4 x i32> %y) { |
| ; CHECK-LABEL: @lshr_or( |
| ; CHECK-NEXT: [[TMP1:%.*]] = lshr <4 x i32> [[X:%.*]], <i32 12, i32 12, i32 12, i32 12> |
| ; CHECK-NEXT: [[TMP2:%.*]] = lshr <4 x i32> [[Y:%.*]], <i32 7, i32 7, i32 7, i32 7> |
| ; CHECK-NEXT: [[SH1:%.*]] = or <4 x i32> [[TMP1]], [[TMP2]] |
| ; CHECK-NEXT: ret <4 x i32> [[SH1]] |
| ; |
| %sh0 = lshr <4 x i32> %x, <i32 5, i32 5, i32 5, i32 5> |
| %r = or <4 x i32> %sh0, %y |
| %sh1 = lshr <4 x i32> %r, <i32 7, i32 7, i32 7, i32 7> |
| ret <4 x i32> %sh1 |
| } |
| |
| define <8 x i16> @lshr_xor(<8 x i16> %x, <8 x i16> %py) { |
| ; CHECK-LABEL: @lshr_xor( |
| ; CHECK-NEXT: [[Y:%.*]] = srem <8 x i16> [[PY:%.*]], <i16 42, i16 42, i16 42, i16 42, i16 42, i16 42, i16 42, i16 42> |
| ; CHECK-NEXT: [[TMP1:%.*]] = lshr <8 x i16> [[X:%.*]], <i16 12, i16 12, i16 12, i16 12, i16 12, i16 12, i16 12, i16 12> |
| ; CHECK-NEXT: [[TMP2:%.*]] = lshr <8 x i16> [[Y]], <i16 7, i16 7, i16 7, i16 7, i16 7, i16 7, i16 7, i16 7> |
| ; CHECK-NEXT: [[SH1:%.*]] = xor <8 x i16> [[TMP1]], [[TMP2]] |
| ; CHECK-NEXT: ret <8 x i16> [[SH1]] |
| ; |
| %y = srem <8 x i16> %py, <i16 42, i16 42, i16 42, i16 42, i16 42, i16 42, i16 42, i16 -42> ; thwart complexity-based canonicalization |
| %sh0 = lshr <8 x i16> %x, <i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5, i16 5> |
| %r = xor <8 x i16> %y, %sh0 |
| %sh1 = lshr <8 x i16> %r, <i16 7, i16 7, i16 7, i16 7, i16 7, i16 7, i16 7, i16 7> |
| ret <8 x i16> %sh1 |
| } |
| |
| define <16 x i8> @ashr_and(<16 x i8> %x, <16 x i8> %py, <16 x i8> %pz) { |
| ; CHECK-LABEL: @ashr_and( |
| ; CHECK-NEXT: [[Y:%.*]] = srem <16 x i8> [[PY:%.*]], [[PZ:%.*]] |
| ; CHECK-NEXT: [[TMP1:%.*]] = ashr <16 x i8> [[X:%.*]], <i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5, i8 5> |
| ; CHECK-NEXT: [[TMP2:%.*]] = ashr <16 x i8> [[Y]], <i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2> |
| ; CHECK-NEXT: [[SH1:%.*]] = and <16 x i8> [[TMP1]], [[TMP2]] |
| ; CHECK-NEXT: ret <16 x i8> [[SH1]] |
| ; |
| %y = srem <16 x i8> %py, %pz ; thwart complexity-based canonicalization |
| %sh0 = ashr <16 x i8> %x, <i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3, i8 3> |
| %r = and <16 x i8> %y, %sh0 |
| %sh1 = ashr <16 x i8> %r, <i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2, i8 2> |
| ret <16 x i8> %sh1 |
| } |
| |
| define <2 x i64> @ashr_or(<2 x i64> %x, <2 x i64> %y) { |
| ; CHECK-LABEL: @ashr_or( |
| ; CHECK-NEXT: [[TMP1:%.*]] = ashr <2 x i64> [[X:%.*]], <i64 12, i64 12> |
| ; CHECK-NEXT: [[TMP2:%.*]] = ashr <2 x i64> [[Y:%.*]], <i64 7, i64 7> |
| ; CHECK-NEXT: [[SH1:%.*]] = or <2 x i64> [[TMP1]], [[TMP2]] |
| ; CHECK-NEXT: ret <2 x i64> [[SH1]] |
| ; |
| %sh0 = ashr <2 x i64> %x, <i64 5, i64 5> |
| %r = or <2 x i64> %sh0, %y |
| %sh1 = ashr <2 x i64> %r, <i64 7, i64 7> |
| ret <2 x i64> %sh1 |
| } |
| |
| define i32 @ashr_xor(i32 %x, i32 %py) { |
| ; CHECK-LABEL: @ashr_xor( |
| ; CHECK-NEXT: [[Y:%.*]] = srem i32 [[PY:%.*]], 42 |
| ; CHECK-NEXT: [[TMP1:%.*]] = ashr i32 [[X:%.*]], 12 |
| ; CHECK-NEXT: [[TMP2:%.*]] = ashr i32 [[Y]], 7 |
| ; CHECK-NEXT: [[SH1:%.*]] = xor i32 [[TMP1]], [[TMP2]] |
| ; CHECK-NEXT: ret i32 [[SH1]] |
| ; |
| %y = srem i32 %py, 42 ; thwart complexity-based canonicalization |
| %sh0 = ashr i32 %x, 5 |
| %r = xor i32 %y, %sh0 |
| %sh1 = ashr i32 %r, 7 |
| ret i32 %sh1 |
| } |
| |
| define i32 @shr_mismatch_xor(i32 %x, i32 %y) { |
| ; CHECK-LABEL: @shr_mismatch_xor( |
| ; CHECK-NEXT: [[SH0:%.*]] = ashr i32 [[X:%.*]], 5 |
| ; CHECK-NEXT: [[R:%.*]] = xor i32 [[SH0]], [[Y:%.*]] |
| ; CHECK-NEXT: [[SH1:%.*]] = lshr i32 [[R]], 7 |
| ; CHECK-NEXT: ret i32 [[SH1]] |
| ; |
| %sh0 = ashr i32 %x, 5 |
| %r = xor i32 %y, %sh0 |
| %sh1 = lshr i32 %r, 7 |
| ret i32 %sh1 |
| } |
| |
| define i32 @ashr_overshift_xor(i32 %x, i32 %y) { |
| ; CHECK-LABEL: @ashr_overshift_xor( |
| ; CHECK-NEXT: [[SH0:%.*]] = ashr i32 [[X:%.*]], 15 |
| ; CHECK-NEXT: [[R:%.*]] = xor i32 [[SH0]], [[Y:%.*]] |
| ; CHECK-NEXT: [[SH1:%.*]] = ashr i32 [[R]], 17 |
| ; CHECK-NEXT: ret i32 [[SH1]] |
| ; |
| %sh0 = ashr i32 %x, 15 |
| %r = xor i32 %y, %sh0 |
| %sh1 = ashr i32 %r, 17 |
| ret i32 %sh1 |
| } |
| |
| define <2 x i32> @ashr_undef_undef_xor(<2 x i32> %x, <2 x i32> %y) { |
| ; CHECK-LABEL: @ashr_undef_undef_xor( |
| ; CHECK-NEXT: [[SH0:%.*]] = ashr <2 x i32> [[X:%.*]], <i32 15, i32 undef> |
| ; CHECK-NEXT: [[R:%.*]] = xor <2 x i32> [[SH0]], [[Y:%.*]] |
| ; CHECK-NEXT: [[SH1:%.*]] = ashr <2 x i32> [[R]], <i32 undef, i32 17> |
| ; CHECK-NEXT: ret <2 x i32> [[SH1]] |
| ; |
| %sh0 = ashr <2 x i32> %x, <i32 15, i32 undef> |
| %r = xor <2 x i32> %y, %sh0 |
| %sh1 = ashr <2 x i32> %r, <i32 undef, i32 17> |
| ret <2 x i32> %sh1 |
| } |
| |
| define i32 @lshr_or_extra_use(i32 %x, i32 %y, ptr %p) { |
| ; CHECK-LABEL: @lshr_or_extra_use( |
| ; CHECK-NEXT: [[SH0:%.*]] = lshr i32 [[X:%.*]], 5 |
| ; CHECK-NEXT: [[R:%.*]] = or i32 [[SH0]], [[Y:%.*]] |
| ; CHECK-NEXT: store i32 [[R]], ptr [[P:%.*]], align 4 |
| ; CHECK-NEXT: [[SH1:%.*]] = lshr i32 [[R]], 7 |
| ; CHECK-NEXT: ret i32 [[SH1]] |
| ; |
| %sh0 = lshr i32 %x, 5 |
| %r = or i32 %sh0, %y |
| store i32 %r, ptr %p |
| %sh1 = lshr i32 %r, 7 |
| ret i32 %sh1 |
| } |
| |
| ; Avoid crashing on constant expressions. |
| |
| @g = external global i32 |
| |
| define i32 @PR44028(i32 %x) { |
| ; CHECK-LABEL: @PR44028( |
| ; CHECK-NEXT: [[SH1:%.*]] = ashr exact i32 [[X:%.*]], 16 |
| ; CHECK-NEXT: [[T0:%.*]] = xor i32 [[SH1]], shl (i32 ptrtoint (ptr @g to i32), i32 16) |
| ; CHECK-NEXT: [[T27:%.*]] = ashr exact i32 [[T0]], 16 |
| ; CHECK-NEXT: ret i32 [[T27]] |
| ; |
| %sh1 = ashr exact i32 %x, 16 |
| %t0 = xor i32 %sh1, shl (i32 ptrtoint (ptr @g to i32), i32 16) |
| %t27 = ashr exact i32 %t0, 16 |
| ret i32 %t27 |
| } |
| |
| define i64 @lshr_mul(i64 %0) { |
| ; CHECK-LABEL: @lshr_mul( |
| ; CHECK-NEXT: [[TMP2:%.*]] = mul nuw i64 [[TMP0:%.*]], 13 |
| ; CHECK-NEXT: ret i64 [[TMP2]] |
| ; |
| %2 = mul nuw i64 %0, 52 |
| %3 = lshr i64 %2, 2 |
| ret i64 %3 |
| } |
| |
| define i64 @lshr_mul_nuw_nsw(i64 %0) { |
| ; CHECK-LABEL: @lshr_mul_nuw_nsw( |
| ; CHECK-NEXT: [[TMP2:%.*]] = mul nuw nsw i64 [[TMP0:%.*]], 13 |
| ; CHECK-NEXT: ret i64 [[TMP2]] |
| ; |
| %2 = mul nuw nsw i64 %0, 52 |
| %3 = lshr i64 %2, 2 |
| ret i64 %3 |
| } |
| |
| define <4 x i32> @lshr_mul_vector(<4 x i32> %0) { |
| ; CHECK-LABEL: @lshr_mul_vector( |
| ; CHECK-NEXT: [[TMP2:%.*]] = mul nuw <4 x i32> [[TMP0:%.*]], <i32 13, i32 13, i32 13, i32 13> |
| ; CHECK-NEXT: ret <4 x i32> [[TMP2]] |
| ; |
| %2 = mul nuw <4 x i32> %0, <i32 52, i32 52, i32 52, i32 52> |
| %3 = lshr <4 x i32> %2, <i32 2, i32 2, i32 2, i32 2> |
| ret <4 x i32> %3 |
| } |
| |
| define i64 @lshr_mul_negative_noexact(i64 %0) { |
| ; CHECK-LABEL: @lshr_mul_negative_noexact( |
| ; CHECK-NEXT: [[TMP2:%.*]] = mul nuw i64 [[TMP0:%.*]], 53 |
| ; CHECK-NEXT: [[TMP3:%.*]] = lshr i64 [[TMP2]], 2 |
| ; CHECK-NEXT: ret i64 [[TMP3]] |
| ; |
| %2 = mul nuw i64 %0, 53 |
| %3 = lshr i64 %2, 2 |
| ret i64 %3 |
| } |
| |
| define i64 @lshr_mul_negative_oneuse(i64 %0) { |
| ; CHECK-LABEL: @lshr_mul_negative_oneuse( |
| ; CHECK-NEXT: [[TMP2:%.*]] = mul nuw i64 [[TMP0:%.*]], 52 |
| ; CHECK-NEXT: call void @use(i64 [[TMP2]]) |
| ; CHECK-NEXT: [[TMP3:%.*]] = lshr exact i64 [[TMP2]], 2 |
| ; CHECK-NEXT: ret i64 [[TMP3]] |
| ; |
| %2 = mul nuw i64 %0, 52 |
| call void @use(i64 %2) |
| %3 = lshr i64 %2, 2 |
| ret i64 %3 |
| } |
| |
| define i64 @lshr_mul_negative_nonuw(i64 %0) { |
| ; CHECK-LABEL: @lshr_mul_negative_nonuw( |
| ; CHECK-NEXT: [[TMP2:%.*]] = mul i64 [[TMP0:%.*]], 52 |
| ; CHECK-NEXT: [[TMP3:%.*]] = lshr exact i64 [[TMP2]], 2 |
| ; CHECK-NEXT: ret i64 [[TMP3]] |
| ; |
| %2 = mul i64 %0, 52 |
| %3 = lshr i64 %2, 2 |
| ret i64 %3 |
| } |
| |
| define i64 @lshr_mul_negative_nsw(i64 %0) { |
| ; CHECK-LABEL: @lshr_mul_negative_nsw( |
| ; CHECK-NEXT: [[TMP2:%.*]] = mul nsw i64 [[TMP0:%.*]], 52 |
| ; CHECK-NEXT: [[TMP3:%.*]] = lshr exact i64 [[TMP2]], 2 |
| ; CHECK-NEXT: ret i64 [[TMP3]] |
| ; |
| %2 = mul nsw i64 %0, 52 |
| %3 = lshr i64 %2, 2 |
| ret i64 %3 |
| } |
| |
| define i8 @shl_add(i8 %x, i8 %y) { |
| ; CHECK-LABEL: @shl_add( |
| ; CHECK-NEXT: [[TMP1:%.*]] = shl i8 [[X:%.*]], 5 |
| ; CHECK-NEXT: [[TMP2:%.*]] = shl i8 [[Y:%.*]], 2 |
| ; CHECK-NEXT: [[SH1:%.*]] = add i8 [[TMP1]], [[TMP2]] |
| ; CHECK-NEXT: ret i8 [[SH1]] |
| ; |
| %sh0 = shl i8 %x, 3 |
| %r = add i8 %sh0, %y |
| %sh1 = shl i8 %r, 2 |
| ret i8 %sh1 |
| } |
| |
| define <2 x i8> @shl_add_nonuniform(<2 x i8> %x, <2 x i8> %y) { |
| ; CHECK-LABEL: @shl_add_nonuniform( |
| ; CHECK-NEXT: [[TMP1:%.*]] = shl <2 x i8> [[X:%.*]], <i8 5, i8 4> |
| ; CHECK-NEXT: [[TMP2:%.*]] = shl <2 x i8> [[Y:%.*]], <i8 2, i8 0> |
| ; CHECK-NEXT: [[SH1:%.*]] = add <2 x i8> [[TMP1]], [[TMP2]] |
| ; CHECK-NEXT: ret <2 x i8> [[SH1]] |
| ; |
| %sh0 = shl <2 x i8> %x, <i8 3, i8 4> |
| %r = add <2 x i8> %sh0, %y |
| %sh1 = shl <2 x i8> %r, <i8 2, i8 0> |
| ret <2 x i8> %sh1 |
| } |
| |
| |
| define <2 x i64> @shl_add_undef(<2 x i64> %x, <2 x i64> %py) { |
| ; CHECK-LABEL: @shl_add_undef( |
| ; CHECK-NEXT: [[Y:%.*]] = srem <2 x i64> [[PY:%.*]], <i64 42, i64 42> |
| ; CHECK-NEXT: [[TMP1:%.*]] = shl <2 x i64> [[X:%.*]], <i64 12, i64 undef> |
| ; CHECK-NEXT: [[TMP2:%.*]] = shl <2 x i64> [[Y]], <i64 7, i64 undef> |
| ; CHECK-NEXT: [[SH1:%.*]] = add <2 x i64> [[TMP1]], [[TMP2]] |
| ; CHECK-NEXT: ret <2 x i64> [[SH1]] |
| ; |
| %y = srem <2 x i64> %py, <i64 42, i64 42> ; thwart complexity-based canonicalization |
| %sh0 = shl <2 x i64> %x, <i64 5, i64 undef> |
| %r = add <2 x i64> %y, %sh0 |
| %sh1 = shl <2 x i64> %r, <i64 7, i64 undef> |
| ret <2 x i64> %sh1 |
| } |
| |
| |
| define i8 @lshr_add(i8 %x, i8 %y) { |
| ; CHECK-LABEL: @lshr_add( |
| ; CHECK-NEXT: [[SH0:%.*]] = lshr i8 [[X:%.*]], 3 |
| ; CHECK-NEXT: [[R:%.*]] = add i8 [[SH0]], [[Y:%.*]] |
| ; CHECK-NEXT: [[SH1:%.*]] = lshr i8 [[R]], 2 |
| ; CHECK-NEXT: ret i8 [[SH1]] |
| ; |
| %sh0 = lshr i8 %x, 3 |
| %r = add i8 %sh0, %y |
| %sh1 = lshr i8 %r, 2 |
| ret i8 %sh1 |
| } |
| |
| define <2 x i8> @lshr_add_nonuniform(<2 x i8> %x, <2 x i8> %y) { |
| ; CHECK-LABEL: @lshr_add_nonuniform( |
| ; CHECK-NEXT: [[SH0:%.*]] = lshr <2 x i8> [[X:%.*]], <i8 3, i8 4> |
| ; CHECK-NEXT: [[R:%.*]] = add <2 x i8> [[SH0]], [[Y:%.*]] |
| ; CHECK-NEXT: [[SH1:%.*]] = lshr <2 x i8> [[R]], <i8 2, i8 0> |
| ; CHECK-NEXT: ret <2 x i8> [[SH1]] |
| ; |
| %sh0 = lshr <2 x i8> %x, <i8 3, i8 4> |
| %r = add <2 x i8> %sh0, %y |
| %sh1 = lshr <2 x i8> %r, <i8 2, i8 0> |
| ret <2 x i8> %sh1 |
| } |
| |
| define <2 x i64> @lshr_add_undef(<2 x i64> %x, <2 x i64> %py) { |
| ; CHECK-LABEL: @lshr_add_undef( |
| ; CHECK-NEXT: [[Y:%.*]] = srem <2 x i64> [[PY:%.*]], <i64 42, i64 42> |
| ; CHECK-NEXT: [[SH0:%.*]] = lshr <2 x i64> [[X:%.*]], <i64 5, i64 undef> |
| ; CHECK-NEXT: [[R:%.*]] = add <2 x i64> [[Y]], [[SH0]] |
| ; CHECK-NEXT: [[SH1:%.*]] = lshr <2 x i64> [[R]], <i64 7, i64 undef> |
| ; CHECK-NEXT: ret <2 x i64> [[SH1]] |
| ; |
| %y = srem <2 x i64> %py, <i64 42, i64 42> ; thwart complexity-based canonicalization |
| %sh0 = lshr <2 x i64> %x, <i64 5, i64 undef> |
| %r = add <2 x i64> %y, %sh0 |
| %sh1 = lshr <2 x i64> %r, <i64 7, i64 undef> |
| ret <2 x i64> %sh1 |
| } |
| |
| define i8 @shl_sub(i8 %x, i8 %y) { |
| ; CHECK-LABEL: @shl_sub( |
| ; CHECK-NEXT: [[TMP1:%.*]] = shl i8 [[X:%.*]], 5 |
| ; CHECK-NEXT: [[TMP2:%.*]] = shl i8 [[Y:%.*]], 2 |
| ; CHECK-NEXT: [[SH1:%.*]] = sub i8 [[TMP1]], [[TMP2]] |
| ; CHECK-NEXT: ret i8 [[SH1]] |
| ; |
| %sh0 = shl i8 %x, 3 |
| %r = sub i8 %sh0, %y |
| %sh1 = shl i8 %r, 2 |
| ret i8 %sh1 |
| } |
| |
| ; Make sure we don't commute operands for sub |
| define i8 @shl_sub_no_commute(i8 %x, i8 %y) { |
| ; CHECK-LABEL: @shl_sub_no_commute( |
| ; CHECK-NEXT: [[TMP1:%.*]] = shl i8 [[Y:%.*]], 5 |
| ; CHECK-NEXT: [[TMP2:%.*]] = shl i8 [[X:%.*]], 2 |
| ; CHECK-NEXT: [[SH1:%.*]] = sub i8 [[TMP2]], [[TMP1]] |
| ; CHECK-NEXT: ret i8 [[SH1]] |
| ; |
| %sh0 = shl i8 %y, 3 |
| %r = sub i8 %x, %sh0 |
| %sh1 = shl i8 %r, 2 |
| ret i8 %sh1 |
| } |
| |
| define <2 x i8> @shl_sub_nonuniform(<2 x i8> %x, <2 x i8> %y) { |
| ; CHECK-LABEL: @shl_sub_nonuniform( |
| ; CHECK-NEXT: [[TMP1:%.*]] = shl <2 x i8> [[X:%.*]], <i8 5, i8 4> |
| ; CHECK-NEXT: [[TMP2:%.*]] = shl <2 x i8> [[Y:%.*]], <i8 2, i8 0> |
| ; CHECK-NEXT: [[SH1:%.*]] = sub <2 x i8> [[TMP1]], [[TMP2]] |
| ; CHECK-NEXT: ret <2 x i8> [[SH1]] |
| ; |
| %sh0 = shl <2 x i8> %x, <i8 3, i8 4> |
| %r = sub <2 x i8> %sh0, %y |
| %sh1 = shl <2 x i8> %r, <i8 2, i8 0> |
| ret <2 x i8> %sh1 |
| } |
| |
| |
| define <2 x i64> @shl_sub_undef(<2 x i64> %x, <2 x i64> %py) { |
| ; CHECK-LABEL: @shl_sub_undef( |
| ; CHECK-NEXT: [[Y:%.*]] = srem <2 x i64> [[PY:%.*]], <i64 42, i64 42> |
| ; CHECK-NEXT: [[TMP1:%.*]] = shl <2 x i64> [[X:%.*]], <i64 12, i64 undef> |
| ; CHECK-NEXT: [[TMP2:%.*]] = shl <2 x i64> [[Y]], <i64 7, i64 undef> |
| ; CHECK-NEXT: [[SH1:%.*]] = sub <2 x i64> [[TMP2]], [[TMP1]] |
| ; CHECK-NEXT: ret <2 x i64> [[SH1]] |
| ; |
| %y = srem <2 x i64> %py, <i64 42, i64 42> ; thwart complexity-based canonicalization |
| %sh0 = shl <2 x i64> %x, <i64 5, i64 undef> |
| %r = sub <2 x i64> %y, %sh0 |
| %sh1 = shl <2 x i64> %r, <i64 7, i64 undef> |
| ret <2 x i64> %sh1 |
| } |
| |
| |
| define i8 @lshr_sub(i8 %x, i8 %y) { |
| ; CHECK-LABEL: @lshr_sub( |
| ; CHECK-NEXT: [[SH0:%.*]] = lshr i8 [[X:%.*]], 3 |
| ; CHECK-NEXT: [[R:%.*]] = sub i8 [[SH0]], [[Y:%.*]] |
| ; CHECK-NEXT: [[SH1:%.*]] = lshr i8 [[R]], 2 |
| ; CHECK-NEXT: ret i8 [[SH1]] |
| ; |
| %sh0 = lshr i8 %x, 3 |
| %r = sub i8 %sh0, %y |
| %sh1 = lshr i8 %r, 2 |
| ret i8 %sh1 |
| } |
| |
| define <2 x i8> @lshr_sub_nonuniform(<2 x i8> %x, <2 x i8> %y) { |
| ; CHECK-LABEL: @lshr_sub_nonuniform( |
| ; CHECK-NEXT: [[SH0:%.*]] = lshr <2 x i8> [[X:%.*]], <i8 3, i8 4> |
| ; CHECK-NEXT: [[R:%.*]] = sub <2 x i8> [[SH0]], [[Y:%.*]] |
| ; CHECK-NEXT: [[SH1:%.*]] = lshr <2 x i8> [[R]], <i8 2, i8 0> |
| ; CHECK-NEXT: ret <2 x i8> [[SH1]] |
| ; |
| %sh0 = lshr <2 x i8> %x, <i8 3, i8 4> |
| %r = sub <2 x i8> %sh0, %y |
| %sh1 = lshr <2 x i8> %r, <i8 2, i8 0> |
| ret <2 x i8> %sh1 |
| } |
| |
| define <2 x i64> @lshr_sub_undef(<2 x i64> %x, <2 x i64> %py) { |
| ; CHECK-LABEL: @lshr_sub_undef( |
| ; CHECK-NEXT: [[Y:%.*]] = srem <2 x i64> [[PY:%.*]], <i64 42, i64 42> |
| ; CHECK-NEXT: [[SH0:%.*]] = lshr <2 x i64> [[X:%.*]], <i64 5, i64 undef> |
| ; CHECK-NEXT: [[R:%.*]] = sub <2 x i64> [[Y]], [[SH0]] |
| ; CHECK-NEXT: [[SH1:%.*]] = lshr <2 x i64> [[R]], <i64 7, i64 undef> |
| ; CHECK-NEXT: ret <2 x i64> [[SH1]] |
| ; |
| %y = srem <2 x i64> %py, <i64 42, i64 42> ; thwart complexity-based canonicalization |
| %sh0 = lshr <2 x i64> %x, <i64 5, i64 undef> |
| %r = sub <2 x i64> %y, %sh0 |
| %sh1 = lshr <2 x i64> %r, <i64 7, i64 undef> |
| ret <2 x i64> %sh1 |
| } |