| ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 6 |
| ; RUN: llc -mcpu=gfx900 < %s | FileCheck %s |
| |
| ; Check for assert when trying to optimize s_and_b64 + s_cbranch_vccz |
| ; pattern. |
| |
| target triple = "amdgcn-amd-amdhsa" |
| |
| declare hidden void @func(<4 x i8>, <2 x i8>) |
| |
| define <4 x i8> @issue176578() #0 { |
| ; CHECK-LABEL: issue176578: |
| ; CHECK: ; %bb.0: ; %bb |
| ; CHECK-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) |
| ; CHECK-NEXT: s_mov_b32 s16, s33 |
| ; CHECK-NEXT: s_mov_b32 s33, s32 |
| ; CHECK-NEXT: s_or_saveexec_b64 s[18:19], -1 |
| ; CHECK-NEXT: buffer_store_dword v41, off, s[0:3], s33 offset:4 ; 4-byte Folded Spill |
| ; CHECK-NEXT: s_mov_b64 exec, s[18:19] |
| ; CHECK-NEXT: v_writelane_b32 v41, s16, 15 |
| ; CHECK-NEXT: v_writelane_b32 v41, s30, 0 |
| ; CHECK-NEXT: v_writelane_b32 v41, s31, 1 |
| ; CHECK-NEXT: v_writelane_b32 v41, s34, 2 |
| ; CHECK-NEXT: v_writelane_b32 v41, s35, 3 |
| ; CHECK-NEXT: v_writelane_b32 v41, s36, 4 |
| ; CHECK-NEXT: v_writelane_b32 v41, s37, 5 |
| ; CHECK-NEXT: v_writelane_b32 v41, s38, 6 |
| ; CHECK-NEXT: v_writelane_b32 v41, s39, 7 |
| ; CHECK-NEXT: v_writelane_b32 v41, s48, 8 |
| ; CHECK-NEXT: v_writelane_b32 v41, s49, 9 |
| ; CHECK-NEXT: v_writelane_b32 v41, s50, 10 |
| ; CHECK-NEXT: v_writelane_b32 v41, s51, 11 |
| ; CHECK-NEXT: v_writelane_b32 v41, s52, 12 |
| ; CHECK-NEXT: v_writelane_b32 v41, s53, 13 |
| ; CHECK-NEXT: buffer_store_dword v40, off, s[0:3], s33 ; 4-byte Folded Spill |
| ; CHECK-NEXT: v_writelane_b32 v41, s54, 14 |
| ; CHECK-NEXT: v_mov_b32_e32 v40, v31 |
| ; CHECK-NEXT: s_mov_b32 s50, s15 |
| ; CHECK-NEXT: s_mov_b32 s51, s14 |
| ; CHECK-NEXT: s_mov_b32 s52, s13 |
| ; CHECK-NEXT: s_mov_b32 s53, s12 |
| ; CHECK-NEXT: s_mov_b64 s[34:35], s[10:11] |
| ; CHECK-NEXT: s_mov_b64 s[36:37], s[8:9] |
| ; CHECK-NEXT: s_mov_b64 s[38:39], s[6:7] |
| ; CHECK-NEXT: s_mov_b64 s[48:49], s[4:5] |
| ; CHECK-NEXT: s_mov_b32 s54, 0 |
| ; CHECK-NEXT: s_addk_i32 s32, 0x400 |
| ; CHECK-NEXT: s_branch .LBB0_2 |
| ; CHECK-NEXT: .LBB0_1: ; %Flow |
| ; CHECK-NEXT: ; in Loop: Header=BB0_2 Depth=1 |
| ; CHECK-NEXT: s_andn2_b64 vcc, exec, s[4:5] |
| ; CHECK-NEXT: s_cbranch_vccz .LBB0_4 |
| ; CHECK-NEXT: .LBB0_2: ; %bb1 |
| ; CHECK-NEXT: ; =>This Inner Loop Header: Depth=1 |
| ; CHECK-NEXT: s_and_b64 vcc, exec, -1 |
| ; CHECK-NEXT: s_cselect_b32 s54, 0, s54 |
| ; CHECK-NEXT: s_mov_b64 s[4:5], -1 |
| ; CHECK-NEXT: s_cbranch_execnz .LBB0_1 |
| ; CHECK-NEXT: ; %bb.3: ; %bb3 |
| ; CHECK-NEXT: ; in Loop: Header=BB0_2 Depth=1 |
| ; CHECK-NEXT: s_getpc_b64 s[16:17] |
| ; CHECK-NEXT: s_add_u32 s16, s16, func@rel32@lo+4 |
| ; CHECK-NEXT: s_addc_u32 s17, s17, func@rel32@hi+12 |
| ; CHECK-NEXT: s_mov_b64 s[4:5], s[48:49] |
| ; CHECK-NEXT: s_mov_b64 s[6:7], s[38:39] |
| ; CHECK-NEXT: s_mov_b64 s[8:9], s[36:37] |
| ; CHECK-NEXT: s_mov_b64 s[10:11], s[34:35] |
| ; CHECK-NEXT: s_mov_b32 s12, s53 |
| ; CHECK-NEXT: s_mov_b32 s13, s52 |
| ; CHECK-NEXT: s_mov_b32 s14, s51 |
| ; CHECK-NEXT: s_mov_b32 s15, s50 |
| ; CHECK-NEXT: v_mov_b32_e32 v31, v40 |
| ; CHECK-NEXT: v_mov_b32_e32 v0, 0 |
| ; CHECK-NEXT: v_mov_b32_e32 v1, 0 |
| ; CHECK-NEXT: v_mov_b32_e32 v2, 0 |
| ; CHECK-NEXT: v_mov_b32_e32 v3, 0 |
| ; CHECK-NEXT: v_mov_b32_e32 v4, 0 |
| ; CHECK-NEXT: v_mov_b32_e32 v5, 0 |
| ; CHECK-NEXT: s_swappc_b64 s[30:31], s[16:17] |
| ; CHECK-NEXT: s_mov_b64 s[4:5], 0 |
| ; CHECK-NEXT: s_branch .LBB0_1 |
| ; CHECK-NEXT: .LBB0_4: ; %bb4 |
| ; CHECK-NEXT: buffer_load_dword v40, off, s[0:3], s33 ; 4-byte Folded Reload |
| ; CHECK-NEXT: v_mov_b32_e32 v0, s54 |
| ; CHECK-NEXT: v_mov_b32_e32 v1, 0 |
| ; CHECK-NEXT: v_mov_b32_e32 v2, 0 |
| ; CHECK-NEXT: v_mov_b32_e32 v3, 0 |
| ; CHECK-NEXT: v_readlane_b32 s54, v41, 14 |
| ; CHECK-NEXT: v_readlane_b32 s53, v41, 13 |
| ; CHECK-NEXT: v_readlane_b32 s52, v41, 12 |
| ; CHECK-NEXT: v_readlane_b32 s51, v41, 11 |
| ; CHECK-NEXT: v_readlane_b32 s50, v41, 10 |
| ; CHECK-NEXT: v_readlane_b32 s49, v41, 9 |
| ; CHECK-NEXT: v_readlane_b32 s48, v41, 8 |
| ; CHECK-NEXT: v_readlane_b32 s39, v41, 7 |
| ; CHECK-NEXT: v_readlane_b32 s38, v41, 6 |
| ; CHECK-NEXT: v_readlane_b32 s37, v41, 5 |
| ; CHECK-NEXT: v_readlane_b32 s36, v41, 4 |
| ; CHECK-NEXT: v_readlane_b32 s35, v41, 3 |
| ; CHECK-NEXT: v_readlane_b32 s34, v41, 2 |
| ; CHECK-NEXT: v_readlane_b32 s31, v41, 1 |
| ; CHECK-NEXT: v_readlane_b32 s30, v41, 0 |
| ; CHECK-NEXT: s_mov_b32 s32, s33 |
| ; CHECK-NEXT: v_readlane_b32 s4, v41, 15 |
| ; CHECK-NEXT: s_or_saveexec_b64 s[6:7], -1 |
| ; CHECK-NEXT: buffer_load_dword v41, off, s[0:3], s33 offset:4 ; 4-byte Folded Reload |
| ; CHECK-NEXT: s_mov_b64 exec, s[6:7] |
| ; CHECK-NEXT: s_mov_b32 s33, s4 |
| ; CHECK-NEXT: s_waitcnt vmcnt(0) |
| ; CHECK-NEXT: s_setpc_b64 s[30:31] |
| bb: |
| br label %bb1 |
| |
| bb1: ; preds = %bb3, %bb |
| %phi = phi i8 [ 0, %bb ], [ %select, %bb3 ] |
| %call = tail call i32 @llvm.amdgcn.msad.u8(i32 0, i32 0, i32 0) |
| %call2 = tail call i32 @llvm.amdgcn.perm(i32 %call, i32 0, i32 0) |
| %and = and i32 %call2, 1 |
| %icmp = icmp ne i32 %and, 1 |
| %select = select i1 %icmp, i8 0, i8 %phi |
| br i1 %icmp, label %bb4, label %bb3 |
| |
| bb3: ; preds = %bb1 |
| tail call void @func(<4 x i8> zeroinitializer, <2 x i8> zeroinitializer) |
| br label %bb1 |
| |
| bb4: ; preds = %bb1 |
| %insertelement = insertelement <4 x i8> zeroinitializer, i8 %select, i64 0 |
| ret <4 x i8> %insertelement |
| } |
| |
| declare i32 @llvm.amdgcn.perm(i32, i32, i32) #1 |
| declare i32 @llvm.amdgcn.msad.u8(i32, i32, i32) #2 |
| |
| attributes #0 = { nounwind } |
| attributes #1 = { nocallback nofree nounwind speculatable willreturn memory(none) } |
| attributes #2 = { nocallback nocreateundeforpoison nofree nosync nounwind speculatable willreturn memory(none) } |