blob: f582f984a39241164338b6373feda861ef788fea [file] [log] [blame]
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 5
; RUN: llc -mtriple=amdgcn-amd-amdhsa -mcpu=gfx942 < %s | FileCheck %s
; iglp.opt should not be flagged as clobbering the memory operand for the global_load, and we should be able to
; lower into the scalar version (i.e. should not need to lower into vector version with waterfall loop)
define amdgpu_kernel void @func(ptr addrspace(1) %in, ptr addrspace(3) %out) {
; CHECK-LABEL: func:
; CHECK: ; %bb.0: ; %.lr.ph
; CHECK-NEXT: s_load_dwordx2 s[6:7], s[4:5], 0x0
; CHECK-NEXT: s_mov_b64 s[8:9], 0
; CHECK-NEXT: s_mov_b64 s[10:11], 0
; CHECK-NEXT: s_mov_b32 s3, 32
; CHECK-NEXT: s_mov_b32 s2, 0
; CHECK-NEXT: s_mov_b64 s[12:13], 0
; CHECK-NEXT: .LBB0_1: ; %loop
; CHECK-NEXT: ; =>This Inner Loop Header: Depth=1
; CHECK-NEXT: s_mov_b64 s[0:1], s[10:11]
; CHECK-NEXT: s_waitcnt lgkmcnt(0)
; CHECK-NEXT: s_add_u32 s10, s6, s12
; CHECK-NEXT: s_addc_u32 s11, s7, s13
; CHECK-NEXT: s_load_dwordx2 s[12:13], s[8:9], 0x0
; CHECK-NEXT: s_add_i32 s3, s3, -1
; CHECK-NEXT: s_cmp_lg_u32 s3, 0
; CHECK-NEXT: ; iglp_opt mask(0x00000000)
; CHECK-NEXT: s_cbranch_scc1 .LBB0_1
; CHECK-NEXT: ; %bb.2: ; %end
; CHECK-NEXT: s_and_b32 s1, s1, 0xffff
; CHECK-NEXT: s_mov_b32 s3, s2
; CHECK-NEXT: buffer_load_dwordx2 v[0:1], off, s[0:3], 0
; CHECK-NEXT: s_load_dword s0, s[4:5], 0x8
; CHECK-NEXT: s_waitcnt vmcnt(0)
; CHECK-NEXT: v_mov_b32_e32 v1, 0
; CHECK-NEXT: s_waitcnt lgkmcnt(0)
; CHECK-NEXT: v_mov_b32_e32 v2, s0
; CHECK-NEXT: v_and_b32_e32 v0, 0xff, v0
; CHECK-NEXT: ds_write_b64 v2, v[0:1]
; CHECK-NEXT: s_endpgm
.lr.ph:
br label %loop
loop: ; preds = %1, %.lr.ph
%addr = phi ptr addrspace(1) [ null, %.lr.ph ], [ %gep, %loop ]
%offset = phi i64 [ 0, %.lr.ph ], [ %nextOff, %loop ]
%inc = phi i32 [0, %.lr.ph], [ %incCond, %loop ]
%rsrc = tail call ptr addrspace(8) @llvm.amdgcn.make.buffer.rsrc.p1(ptr addrspace(1) %addr, i16 0, i32 0, i32 0)
%load = tail call <2 x i32> @llvm.amdgcn.raw.ptr.buffer.load.v2i32(ptr addrspace(8) %rsrc, i32 0, i32 0, i32 0)
%load.bc = bitcast <2 x i32> %load to <8 x i8>
%load.elem = extractelement <8 x i8> %load.bc, i64 0
tail call void @llvm.amdgcn.iglp.opt(i32 0)
%vec = insertelement <4 x i8> zeroinitializer, i8 %load.elem, i64 0
%vec.bc = bitcast <4 x i8> %vec to <2 x half>
%shuff = shufflevector <2 x half> %vec.bc, <2 x half> zeroinitializer, <4 x i32> <i32 0, i32 1, i32 2, i32 3>
%gep = getelementptr i8, ptr addrspace(1) %in, i64 %offset
%unmaskedload49 = load <1 x i64>, ptr addrspace(1) null, align 8
%nextOff = extractelement <1 x i64> %unmaskedload49, i64 0
%incCond = add i32 %inc, 1
%cond = icmp eq i32 %incCond, 32
br i1 %cond, label %end, label %loop
end:
store <4 x half> %shuff, ptr addrspace(3) %out, align 8
ret void
}
declare ptr addrspace(8) @llvm.amdgcn.make.buffer.rsrc.p1(ptr addrspace(1) readnone, i16, i32, i32) #0
declare <2 x i32> @llvm.amdgcn.raw.ptr.buffer.load.v2i32(ptr addrspace(8) nocapture readonly, i32, i32, i32 immarg) #1