[llvm] r343930 - [RISCV] Introduce alu8.ll and alu16.ll tests

Alex Bradbury via llvm-commits llvm-commits at lists.llvm.org
Sat Oct 6 23:53:46 PDT 2018


Author: asb
Date: Sat Oct  6 23:53:46 2018
New Revision: 343930

URL: http://llvm.org/viewvc/llvm-project?rev=343930&view=rev
Log:
[RISCV] Introduce alu8.ll and alu16.ll tests

These track the quality of generated code for simple arithmetic operations
that were legalised from non-native types.

Added:
    llvm/trunk/test/CodeGen/RISCV/alu16.ll
    llvm/trunk/test/CodeGen/RISCV/alu8.ll

Added: llvm/trunk/test/CodeGen/RISCV/alu16.ll
URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/test/CodeGen/RISCV/alu16.ll?rev=343930&view=auto
==============================================================================
--- llvm/trunk/test/CodeGen/RISCV/alu16.ll (added)
+++ llvm/trunk/test/CodeGen/RISCV/alu16.ll Sat Oct  6 23:53:46 2018
@@ -0,0 +1,214 @@
+; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
+; RUN: llc -mtriple=riscv32 -verify-machineinstrs < %s \
+; RUN:   | FileCheck %s -check-prefix=RV32I
+
+; These tests are identical to those in alu32.ll but operate on i16. They check
+; that legalisation of these non-native types doesn't introduce unnecessary
+; inefficiencies.
+
+; TODO: it's unnecessary to mask (zero-extend) the shift amount.
+
+define i16 @addi(i16 %a) nounwind {
+; RV32I-LABEL: addi:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    addi a0, a0, 1
+; RV32I-NEXT:    ret
+  %1 = add i16 %a, 1
+  ret i16 %1
+}
+
+define i16 @slti(i16 %a) nounwind {
+; RV32I-LABEL: slti:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    slli a0, a0, 16
+; RV32I-NEXT:    srai a0, a0, 16
+; RV32I-NEXT:    slti a0, a0, 2
+; RV32I-NEXT:    ret
+  %1 = icmp slt i16 %a, 2
+  %2 = zext i1 %1 to i16
+  ret i16 %2
+}
+
+define i16 @sltiu(i16 %a) nounwind {
+; RV32I-LABEL: sltiu:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    lui a1, 16
+; RV32I-NEXT:    addi a1, a1, -1
+; RV32I-NEXT:    and a0, a0, a1
+; RV32I-NEXT:    sltiu a0, a0, 3
+; RV32I-NEXT:    ret
+  %1 = icmp ult i16 %a, 3
+  %2 = zext i1 %1 to i16
+  ret i16 %2
+}
+
+define i16 @xori(i16 %a) nounwind {
+; RV32I-LABEL: xori:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    xori a0, a0, 4
+; RV32I-NEXT:    ret
+  %1 = xor i16 %a, 4
+  ret i16 %1
+}
+
+define i16 @ori(i16 %a) nounwind {
+; RV32I-LABEL: ori:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    ori a0, a0, 5
+; RV32I-NEXT:    ret
+  %1 = or i16 %a, 5
+  ret i16 %1
+}
+
+define i16 @andi(i16 %a) nounwind {
+; RV32I-LABEL: andi:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    andi a0, a0, 6
+; RV32I-NEXT:    ret
+  %1 = and i16 %a, 6
+  ret i16 %1
+}
+
+define i16 @slli(i16 %a) nounwind {
+; RV32I-LABEL: slli:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    slli a0, a0, 7
+; RV32I-NEXT:    ret
+  %1 = shl i16 %a, 7
+  ret i16 %1
+}
+
+define i16 @srli(i16 %a) nounwind {
+; RV32I-LABEL: srli:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    lui a1, 16
+; RV32I-NEXT:    addi a1, a1, -256
+; RV32I-NEXT:    and a0, a0, a1
+; RV32I-NEXT:    srli a0, a0, 8
+; RV32I-NEXT:    ret
+  %1 = lshr i16 %a, 8
+  ret i16 %1
+}
+
+define i16 @srai(i16 %a) nounwind {
+; RV32I-LABEL: srai:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    slli a0, a0, 16
+; RV32I-NEXT:    srai a0, a0, 25
+; RV32I-NEXT:    ret
+  %1 = ashr i16 %a, 9
+  ret i16 %1
+}
+
+
+define i16 @add(i16 %a, i16 %b) nounwind {
+; RV32I-LABEL: add:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    add a0, a0, a1
+; RV32I-NEXT:    ret
+  %1 = add i16 %a, %b
+  ret i16 %1
+}
+
+define i16 @sub(i16 %a, i16 %b) nounwind {
+; RV32I-LABEL: sub:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    sub a0, a0, a1
+; RV32I-NEXT:    ret
+  %1 = sub i16 %a, %b
+  ret i16 %1
+}
+
+define i16 @sll(i16 %a, i16 %b) nounwind {
+; RV32I-LABEL: sll:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    lui a2, 16
+; RV32I-NEXT:    addi a2, a2, -1
+; RV32I-NEXT:    and a1, a1, a2
+; RV32I-NEXT:    sll a0, a0, a1
+; RV32I-NEXT:    ret
+  %1 = shl i16 %a, %b
+  ret i16 %1
+}
+
+define i16 @slt(i16 %a, i16 %b) nounwind {
+; RV32I-LABEL: slt:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    slli a1, a1, 16
+; RV32I-NEXT:    srai a1, a1, 16
+; RV32I-NEXT:    slli a0, a0, 16
+; RV32I-NEXT:    srai a0, a0, 16
+; RV32I-NEXT:    slt a0, a0, a1
+; RV32I-NEXT:    ret
+  %1 = icmp slt i16 %a, %b
+  %2 = zext i1 %1 to i16
+  ret i16 %2
+}
+
+define i16 @sltu(i16 %a, i16 %b) nounwind {
+; RV32I-LABEL: sltu:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    lui a2, 16
+; RV32I-NEXT:    addi a2, a2, -1
+; RV32I-NEXT:    and a1, a1, a2
+; RV32I-NEXT:    and a0, a0, a2
+; RV32I-NEXT:    sltu a0, a0, a1
+; RV32I-NEXT:    ret
+  %1 = icmp ult i16 %a, %b
+  %2 = zext i1 %1 to i16
+  ret i16 %2
+}
+
+define i16 @xor(i16 %a, i16 %b) nounwind {
+; RV32I-LABEL: xor:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    xor a0, a0, a1
+; RV32I-NEXT:    ret
+  %1 = xor i16 %a, %b
+  ret i16 %1
+}
+
+define i16 @srl(i16 %a, i16 %b) nounwind {
+; RV32I-LABEL: srl:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    lui a2, 16
+; RV32I-NEXT:    addi a2, a2, -1
+; RV32I-NEXT:    and a1, a1, a2
+; RV32I-NEXT:    and a0, a0, a2
+; RV32I-NEXT:    srl a0, a0, a1
+; RV32I-NEXT:    ret
+  %1 = lshr i16 %a, %b
+  ret i16 %1
+}
+
+define i16 @sra(i16 %a, i16 %b) nounwind {
+; RV32I-LABEL: sra:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    lui a2, 16
+; RV32I-NEXT:    addi a2, a2, -1
+; RV32I-NEXT:    and a1, a1, a2
+; RV32I-NEXT:    slli a0, a0, 16
+; RV32I-NEXT:    srai a0, a0, 16
+; RV32I-NEXT:    sra a0, a0, a1
+; RV32I-NEXT:    ret
+  %1 = ashr i16 %a, %b
+  ret i16 %1
+}
+
+define i16 @or(i16 %a, i16 %b) nounwind {
+; RV32I-LABEL: or:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    or a0, a0, a1
+; RV32I-NEXT:    ret
+  %1 = or i16 %a, %b
+  ret i16 %1
+}
+
+define i16 @and(i16 %a, i16 %b) nounwind {
+; RV32I-LABEL: and:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    and a0, a0, a1
+; RV32I-NEXT:    ret
+  %1 = and i16 %a, %b
+  ret i16 %1
+}

Added: llvm/trunk/test/CodeGen/RISCV/alu8.ll
URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/test/CodeGen/RISCV/alu8.ll?rev=343930&view=auto
==============================================================================
--- llvm/trunk/test/CodeGen/RISCV/alu8.ll (added)
+++ llvm/trunk/test/CodeGen/RISCV/alu8.ll Sat Oct  6 23:53:46 2018
@@ -0,0 +1,198 @@
+; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
+; RUN: llc -mtriple=riscv32 -verify-machineinstrs < %s \
+; RUN:   | FileCheck %s -check-prefix=RV32I
+
+; These tests are identical to those in alu32.ll but operate on i8. They check
+; that legalisation of these non-native types doesn't introduce unnecessary
+; inefficiencies.
+
+; TODO: it's unnecessary to mask (zero-extend) the shift amount.
+
+define i8 @addi(i8 %a) nounwind {
+; RV32I-LABEL: addi:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    addi a0, a0, 1
+; RV32I-NEXT:    ret
+  %1 = add i8 %a, 1
+  ret i8 %1
+}
+
+define i8 @slti(i8 %a) nounwind {
+; RV32I-LABEL: slti:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    slli a0, a0, 24
+; RV32I-NEXT:    srai a0, a0, 24
+; RV32I-NEXT:    slti a0, a0, 2
+; RV32I-NEXT:    ret
+  %1 = icmp slt i8 %a, 2
+  %2 = zext i1 %1 to i8
+  ret i8 %2
+}
+
+define i8 @sltiu(i8 %a) nounwind {
+; RV32I-LABEL: sltiu:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    andi a0, a0, 255
+; RV32I-NEXT:    sltiu a0, a0, 3
+; RV32I-NEXT:    ret
+  %1 = icmp ult i8 %a, 3
+  %2 = zext i1 %1 to i8
+  ret i8 %2
+}
+
+define i8 @xori(i8 %a) nounwind {
+; RV32I-LABEL: xori:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    xori a0, a0, 4
+; RV32I-NEXT:    ret
+  %1 = xor i8 %a, 4
+  ret i8 %1
+}
+
+define i8 @ori(i8 %a) nounwind {
+; RV32I-LABEL: ori:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    ori a0, a0, 5
+; RV32I-NEXT:    ret
+  %1 = or i8 %a, 5
+  ret i8 %1
+}
+
+define i8 @andi(i8 %a) nounwind {
+; RV32I-LABEL: andi:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    andi a0, a0, 6
+; RV32I-NEXT:    ret
+  %1 = and i8 %a, 6
+  ret i8 %1
+}
+
+define i8 @slli(i8 %a) nounwind {
+; RV32I-LABEL: slli:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    slli a0, a0, 7
+; RV32I-NEXT:    ret
+  %1 = shl i8 %a, 7
+  ret i8 %1
+}
+
+define i8 @srli(i8 %a) nounwind {
+; RV32I-LABEL: srli:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    ret
+  %1 = lshr i8 %a, 8
+  ret i8 %1
+}
+
+define i8 @srai(i8 %a) nounwind {
+; RV32I-LABEL: srai:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    ret
+  %1 = ashr i8 %a, 9
+  ret i8 %1
+}
+
+
+define i8 @add(i8 %a, i8 %b) nounwind {
+; RV32I-LABEL: add:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    add a0, a0, a1
+; RV32I-NEXT:    ret
+  %1 = add i8 %a, %b
+  ret i8 %1
+}
+
+define i8 @sub(i8 %a, i8 %b) nounwind {
+; RV32I-LABEL: sub:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    sub a0, a0, a1
+; RV32I-NEXT:    ret
+  %1 = sub i8 %a, %b
+  ret i8 %1
+}
+
+define i8 @sll(i8 %a, i8 %b) nounwind {
+; RV32I-LABEL: sll:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    andi a1, a1, 255
+; RV32I-NEXT:    sll a0, a0, a1
+; RV32I-NEXT:    ret
+  %1 = shl i8 %a, %b
+  ret i8 %1
+}
+
+define i8 @slt(i8 %a, i8 %b) nounwind {
+; RV32I-LABEL: slt:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    slli a1, a1, 24
+; RV32I-NEXT:    srai a1, a1, 24
+; RV32I-NEXT:    slli a0, a0, 24
+; RV32I-NEXT:    srai a0, a0, 24
+; RV32I-NEXT:    slt a0, a0, a1
+; RV32I-NEXT:    ret
+  %1 = icmp slt i8 %a, %b
+  %2 = zext i1 %1 to i8
+  ret i8 %2
+}
+
+define i8 @sltu(i8 %a, i8 %b) nounwind {
+; RV32I-LABEL: sltu:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    andi a1, a1, 255
+; RV32I-NEXT:    andi a0, a0, 255
+; RV32I-NEXT:    sltu a0, a0, a1
+; RV32I-NEXT:    ret
+  %1 = icmp ult i8 %a, %b
+  %2 = zext i1 %1 to i8
+  ret i8 %2
+}
+
+define i8 @xor(i8 %a, i8 %b) nounwind {
+; RV32I-LABEL: xor:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    xor a0, a0, a1
+; RV32I-NEXT:    ret
+  %1 = xor i8 %a, %b
+  ret i8 %1
+}
+
+define i8 @srl(i8 %a, i8 %b) nounwind {
+; RV32I-LABEL: srl:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    andi a1, a1, 255
+; RV32I-NEXT:    andi a0, a0, 255
+; RV32I-NEXT:    srl a0, a0, a1
+; RV32I-NEXT:    ret
+  %1 = lshr i8 %a, %b
+  ret i8 %1
+}
+
+define i8 @sra(i8 %a, i8 %b) nounwind {
+; RV32I-LABEL: sra:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    andi a1, a1, 255
+; RV32I-NEXT:    slli a0, a0, 24
+; RV32I-NEXT:    srai a0, a0, 24
+; RV32I-NEXT:    sra a0, a0, a1
+; RV32I-NEXT:    ret
+  %1 = ashr i8 %a, %b
+  ret i8 %1
+}
+
+define i8 @or(i8 %a, i8 %b) nounwind {
+; RV32I-LABEL: or:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    or a0, a0, a1
+; RV32I-NEXT:    ret
+  %1 = or i8 %a, %b
+  ret i8 %1
+}
+
+define i8 @and(i8 %a, i8 %b) nounwind {
+; RV32I-LABEL: and:
+; RV32I:       # %bb.0:
+; RV32I-NEXT:    and a0, a0, a1
+; RV32I-NEXT:    ret
+  %1 = and i8 %a, %b
+  ret i8 %1
+}




More information about the llvm-commits mailing list