blob: c347241dff92f01fbef4d43603a81b21b0ead1c7 [file] [edit]
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 5
; RUN: llc -mtriple=riscv32 -global-isel -global-isel-abort=1 < %s 2>&1 | FileCheck %s --check-prefixes=RV32
; RUN: llc -mtriple=riscv64 -global-isel -global-isel-abort=1 < %s 2>&1 | FileCheck %s --check-prefixes=RV64
define i2 @bitreverse_i2(i2 %x) {
; RV32-LABEL: bitreverse_i2:
; RV32: # %bb.0:
; RV32-NEXT: slli a1, a0, 1
; RV32-NEXT: andi a0, a0, 3
; RV32-NEXT: andi a1, a1, 2
; RV32-NEXT: srli a0, a0, 1
; RV32-NEXT: or a0, a1, a0
; RV32-NEXT: ret
;
; RV64-LABEL: bitreverse_i2:
; RV64: # %bb.0:
; RV64-NEXT: slli a1, a0, 1
; RV64-NEXT: andi a0, a0, 3
; RV64-NEXT: andi a1, a1, 2
; RV64-NEXT: srli a0, a0, 1
; RV64-NEXT: or a0, a1, a0
; RV64-NEXT: ret
%rev = call i2 @llvm.bitreverse.i2(i2 %x)
ret i2 %rev
}
define i3 @bitreverse_i3(i3 %x) {
; RV32-LABEL: bitreverse_i3:
; RV32: # %bb.0:
; RV32-NEXT: slli a1, a0, 2
; RV32-NEXT: andi a0, a0, 7
; RV32-NEXT: andi a1, a1, 4
; RV32-NEXT: andi a2, a0, 2
; RV32-NEXT: or a1, a1, a2
; RV32-NEXT: srli a0, a0, 2
; RV32-NEXT: or a0, a1, a0
; RV32-NEXT: ret
;
; RV64-LABEL: bitreverse_i3:
; RV64: # %bb.0:
; RV64-NEXT: slli a1, a0, 2
; RV64-NEXT: andi a0, a0, 7
; RV64-NEXT: andi a1, a1, 4
; RV64-NEXT: andi a2, a0, 2
; RV64-NEXT: or a1, a1, a2
; RV64-NEXT: srli a0, a0, 2
; RV64-NEXT: or a0, a1, a0
; RV64-NEXT: ret
%rev = call i3 @llvm.bitreverse.i3(i3 %x)
ret i3 %rev
}
define i4 @bitreverse_i4(i4 %x) {
; RV32-LABEL: bitreverse_i4:
; RV32: # %bb.0:
; RV32-NEXT: slli a1, a0, 3
; RV32-NEXT: slli a2, a0, 1
; RV32-NEXT: andi a0, a0, 15
; RV32-NEXT: andi a1, a1, 8
; RV32-NEXT: andi a2, a2, 4
; RV32-NEXT: or a1, a1, a2
; RV32-NEXT: srli a2, a0, 1
; RV32-NEXT: andi a2, a2, 2
; RV32-NEXT: or a1, a1, a2
; RV32-NEXT: srli a0, a0, 3
; RV32-NEXT: or a0, a1, a0
; RV32-NEXT: ret
;
; RV64-LABEL: bitreverse_i4:
; RV64: # %bb.0:
; RV64-NEXT: slli a1, a0, 3
; RV64-NEXT: slli a2, a0, 1
; RV64-NEXT: andi a0, a0, 15
; RV64-NEXT: andi a1, a1, 8
; RV64-NEXT: andi a2, a2, 4
; RV64-NEXT: or a1, a1, a2
; RV64-NEXT: srli a2, a0, 1
; RV64-NEXT: andi a2, a2, 2
; RV64-NEXT: or a1, a1, a2
; RV64-NEXT: srli a0, a0, 3
; RV64-NEXT: or a0, a1, a0
; RV64-NEXT: ret
%rev = call i4 @llvm.bitreverse.i4(i4 %x)
ret i4 %rev
}
define i7 @bitreverse_i7(i7 %x) {
; RV32-LABEL: bitreverse_i7:
; RV32: # %bb.0:
; RV32-NEXT: slli a1, a0, 6
; RV32-NEXT: slli a2, a0, 4
; RV32-NEXT: slli a3, a0, 2
; RV32-NEXT: andi a0, a0, 127
; RV32-NEXT: andi a1, a1, 64
; RV32-NEXT: andi a2, a2, 32
; RV32-NEXT: andi a3, a3, 16
; RV32-NEXT: or a1, a1, a2
; RV32-NEXT: andi a2, a0, 8
; RV32-NEXT: or a2, a3, a2
; RV32-NEXT: srli a3, a0, 2
; RV32-NEXT: or a1, a1, a2
; RV32-NEXT: srli a2, a0, 4
; RV32-NEXT: andi a3, a3, 4
; RV32-NEXT: andi a2, a2, 2
; RV32-NEXT: or a2, a3, a2
; RV32-NEXT: or a1, a1, a2
; RV32-NEXT: srli a0, a0, 6
; RV32-NEXT: or a0, a1, a0
; RV32-NEXT: ret
;
; RV64-LABEL: bitreverse_i7:
; RV64: # %bb.0:
; RV64-NEXT: slli a1, a0, 6
; RV64-NEXT: slli a2, a0, 4
; RV64-NEXT: slli a3, a0, 2
; RV64-NEXT: andi a0, a0, 127
; RV64-NEXT: andi a1, a1, 64
; RV64-NEXT: andi a2, a2, 32
; RV64-NEXT: andi a3, a3, 16
; RV64-NEXT: or a1, a1, a2
; RV64-NEXT: andi a2, a0, 8
; RV64-NEXT: or a2, a3, a2
; RV64-NEXT: srli a3, a0, 2
; RV64-NEXT: or a1, a1, a2
; RV64-NEXT: srli a2, a0, 4
; RV64-NEXT: andi a3, a3, 4
; RV64-NEXT: andi a2, a2, 2
; RV64-NEXT: or a2, a3, a2
; RV64-NEXT: or a1, a1, a2
; RV64-NEXT: srli a0, a0, 6
; RV64-NEXT: or a0, a1, a0
; RV64-NEXT: ret
%rev = call i7 @llvm.bitreverse.i7(i7 %x)
ret i7 %rev
}
define i24 @bitreverse_i24(i24 %x) {
; RV32-LABEL: bitreverse_i24:
; RV32: # %bb.0:
; RV32-NEXT: slli a1, a0, 16
; RV32-NEXT: lui a2, 4096
; RV32-NEXT: lui a3, 1048335
; RV32-NEXT: addi a2, a2, -1
; RV32-NEXT: addi a3, a3, 240
; RV32-NEXT: and a0, a0, a2
; RV32-NEXT: srli a0, a0, 16
; RV32-NEXT: or a0, a0, a1
; RV32-NEXT: and a1, a3, a2
; RV32-NEXT: and a1, a0, a1
; RV32-NEXT: slli a0, a0, 4
; RV32-NEXT: and a0, a0, a3
; RV32-NEXT: lui a3, 1047757
; RV32-NEXT: addi a3, a3, -820
; RV32-NEXT: srli a1, a1, 4
; RV32-NEXT: or a0, a1, a0
; RV32-NEXT: and a1, a3, a2
; RV32-NEXT: and a1, a0, a1
; RV32-NEXT: slli a0, a0, 2
; RV32-NEXT: and a0, a0, a3
; RV32-NEXT: lui a3, 1047211
; RV32-NEXT: addi a3, a3, -1366
; RV32-NEXT: and a2, a3, a2
; RV32-NEXT: srli a1, a1, 2
; RV32-NEXT: or a0, a1, a0
; RV32-NEXT: and a2, a0, a2
; RV32-NEXT: slli a0, a0, 1
; RV32-NEXT: srli a2, a2, 1
; RV32-NEXT: and a0, a0, a3
; RV32-NEXT: or a0, a2, a0
; RV32-NEXT: ret
;
; RV64-LABEL: bitreverse_i24:
; RV64: # %bb.0:
; RV64-NEXT: slli a1, a0, 16
; RV64-NEXT: lui a2, 4096
; RV64-NEXT: lui a3, 1048335
; RV64-NEXT: addi a2, a2, -1
; RV64-NEXT: addi a3, a3, 240
; RV64-NEXT: and a0, a0, a2
; RV64-NEXT: srli a0, a0, 16
; RV64-NEXT: or a0, a0, a1
; RV64-NEXT: and a1, a3, a2
; RV64-NEXT: and a1, a0, a1
; RV64-NEXT: slli a0, a0, 4
; RV64-NEXT: and a0, a0, a3
; RV64-NEXT: lui a3, 1047757
; RV64-NEXT: addi a3, a3, -820
; RV64-NEXT: srli a1, a1, 4
; RV64-NEXT: or a0, a1, a0
; RV64-NEXT: and a1, a3, a2
; RV64-NEXT: and a1, a0, a1
; RV64-NEXT: slli a0, a0, 2
; RV64-NEXT: and a0, a0, a3
; RV64-NEXT: lui a3, 1047211
; RV64-NEXT: addi a3, a3, -1366
; RV64-NEXT: and a2, a3, a2
; RV64-NEXT: srli a1, a1, 2
; RV64-NEXT: or a0, a1, a0
; RV64-NEXT: and a2, a0, a2
; RV64-NEXT: slli a0, a0, 1
; RV64-NEXT: srli a2, a2, 1
; RV64-NEXT: and a0, a0, a3
; RV64-NEXT: or a0, a2, a0
; RV64-NEXT: ret
%rev = call i24 @llvm.bitreverse.i24(i24 %x)
ret i24 %rev
}
define i2 @test_ctpop_i2(i2 %a) {
; RV32-LABEL: test_ctpop_i2:
; RV32: # %bb.0:
; RV32-NEXT: addi sp, sp, -16
; RV32-NEXT: .cfi_def_cfa_offset 16
; RV32-NEXT: sw ra, 12(sp) # 4-byte Folded Spill
; RV32-NEXT: .cfi_offset ra, -4
; RV32-NEXT: andi a0, a0, 3
; RV32-NEXT: srli a1, a0, 1
; RV32-NEXT: sub a0, a0, a1
; RV32-NEXT: zext.b a1, a0
; RV32-NEXT: srli a1, a1, 2
; RV32-NEXT: andi a1, a1, 51
; RV32-NEXT: andi a0, a0, 51
; RV32-NEXT: add a0, a1, a0
; RV32-NEXT: srli a1, a0, 4
; RV32-NEXT: add a0, a1, a0
; RV32-NEXT: andi a0, a0, 15
; RV32-NEXT: li a1, 1
; RV32-NEXT: call __mulsi3
; RV32-NEXT: zext.b a0, a0
; RV32-NEXT: lw ra, 12(sp) # 4-byte Folded Reload
; RV32-NEXT: .cfi_restore ra
; RV32-NEXT: addi sp, sp, 16
; RV32-NEXT: .cfi_def_cfa_offset 0
; RV32-NEXT: ret
;
; RV64-LABEL: test_ctpop_i2:
; RV64: # %bb.0:
; RV64-NEXT: addi sp, sp, -16
; RV64-NEXT: .cfi_def_cfa_offset 16
; RV64-NEXT: sd ra, 8(sp) # 8-byte Folded Spill
; RV64-NEXT: .cfi_offset ra, -8
; RV64-NEXT: andi a0, a0, 3
; RV64-NEXT: srli a1, a0, 1
; RV64-NEXT: sub a0, a0, a1
; RV64-NEXT: zext.b a1, a0
; RV64-NEXT: srli a1, a1, 2
; RV64-NEXT: andi a1, a1, 51
; RV64-NEXT: andi a0, a0, 51
; RV64-NEXT: add a0, a1, a0
; RV64-NEXT: srli a1, a0, 4
; RV64-NEXT: add a0, a1, a0
; RV64-NEXT: andi a0, a0, 15
; RV64-NEXT: li a1, 1
; RV64-NEXT: call __muldi3
; RV64-NEXT: zext.b a0, a0
; RV64-NEXT: ld ra, 8(sp) # 8-byte Folded Reload
; RV64-NEXT: .cfi_restore ra
; RV64-NEXT: addi sp, sp, 16
; RV64-NEXT: .cfi_def_cfa_offset 0
; RV64-NEXT: ret
%1 = call i2 @llvm.ctpop.i2(i2 %a)
ret i2 %1
}
define i11 @test_ctpop_i11(i11 %a) {
; RV32-LABEL: test_ctpop_i11:
; RV32: # %bb.0:
; RV32-NEXT: andi a0, a0, 2047
; RV32-NEXT: lui a1, 5
; RV32-NEXT: srli a2, a0, 1
; RV32-NEXT: addi a1, a1, 1365
; RV32-NEXT: and a1, a2, a1
; RV32-NEXT: lui a2, 3
; RV32-NEXT: addi a2, a2, 819
; RV32-NEXT: sub a0, a0, a1
; RV32-NEXT: slli a1, a0, 16
; RV32-NEXT: and a0, a0, a2
; RV32-NEXT: srli a1, a1, 16
; RV32-NEXT: srli a1, a1, 2
; RV32-NEXT: and a1, a1, a2
; RV32-NEXT: lui a2, 1
; RV32-NEXT: add a0, a1, a0
; RV32-NEXT: srli a1, a0, 4
; RV32-NEXT: add a0, a1, a0
; RV32-NEXT: addi a1, a2, -241
; RV32-NEXT: and a0, a0, a1
; RV32-NEXT: srli a1, a0, 8
; RV32-NEXT: add a0, a0, a1
; RV32-NEXT: zext.b a0, a0
; RV32-NEXT: ret
;
; RV64-LABEL: test_ctpop_i11:
; RV64: # %bb.0:
; RV64-NEXT: andi a0, a0, 2047
; RV64-NEXT: lui a1, 5
; RV64-NEXT: srli a2, a0, 1
; RV64-NEXT: addi a1, a1, 1365
; RV64-NEXT: and a1, a2, a1
; RV64-NEXT: lui a2, 3
; RV64-NEXT: addi a2, a2, 819
; RV64-NEXT: sub a0, a0, a1
; RV64-NEXT: slli a1, a0, 48
; RV64-NEXT: and a0, a0, a2
; RV64-NEXT: srli a1, a1, 48
; RV64-NEXT: srli a1, a1, 2
; RV64-NEXT: and a1, a1, a2
; RV64-NEXT: lui a2, 1
; RV64-NEXT: add a0, a1, a0
; RV64-NEXT: srli a1, a0, 4
; RV64-NEXT: add a0, a1, a0
; RV64-NEXT: addi a1, a2, -241
; RV64-NEXT: and a0, a0, a1
; RV64-NEXT: srli a1, a0, 8
; RV64-NEXT: add a0, a0, a1
; RV64-NEXT: zext.b a0, a0
; RV64-NEXT: ret
%1 = call i11 @llvm.ctpop.i11(i11 %a)
ret i11 %1
}