; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py ; RUN: llc -mtriple=riscv32 -verify-machineinstrs < %s \ ; RUN: | FileCheck -check-prefixes=ALL,SLOW,RV32I %s ; RUN: llc -mtriple=riscv64 -verify-machineinstrs < %s \ ; RUN: | FileCheck -check-prefixes=ALL,SLOW,RV64I %s ; RUN: llc -mtriple=riscv32 -mattr=+fast-unaligned-access -verify-machineinstrs < %s \ ; RUN: | FileCheck -check-prefixes=ALL,FAST,RV32I-FAST %s ; RUN: llc -mtriple=riscv64 -mattr=+fast-unaligned-access -verify-machineinstrs < %s \ ; RUN: | FileCheck -check-prefixes=ALL,FAST,RV64I-FAST %s ; RUN: llc -mtriple=riscv32 -mattr=+unaligned-scalar-mem -verify-machineinstrs < %s \ ; RUN: | FileCheck -check-prefixes=ALL,FAST,RV32I-FAST %s ; RUN: llc -mtriple=riscv64 -mattr=+unaligned-scalar-mem -verify-machineinstrs < %s \ ; RUN: | FileCheck -check-prefixes=ALL,FAST,RV64I-FAST %s ; A collection of cases showing codegen for unaligned loads and stores define i8 @load_i8(ptr %p) { ; ALL-LABEL: load_i8: ; ALL: # %bb.0: ; ALL-NEXT: lbu a0, 0(a0) ; ALL-NEXT: ret %res = load i8, ptr %p, align 1 ret i8 %res } define i16 @load_i16(ptr %p) { ; SLOW-LABEL: load_i16: ; SLOW: # %bb.0: ; SLOW-NEXT: lbu a1, 1(a0) ; SLOW-NEXT: lbu a0, 0(a0) ; SLOW-NEXT: slli a1, a1, 8 ; SLOW-NEXT: or a0, a1, a0 ; SLOW-NEXT: ret ; ; FAST-LABEL: load_i16: ; FAST: # %bb.0: ; FAST-NEXT: lh a0, 0(a0) ; FAST-NEXT: ret %res = load i16, ptr %p, align 1 ret i16 %res } define i24 @load_i24(ptr %p) { ; SLOW-LABEL: load_i24: ; SLOW: # %bb.0: ; SLOW-NEXT: lbu a1, 1(a0) ; SLOW-NEXT: lbu a2, 0(a0) ; SLOW-NEXT: lbu a0, 2(a0) ; SLOW-NEXT: slli a1, a1, 8 ; SLOW-NEXT: or a1, a1, a2 ; SLOW-NEXT: slli a0, a0, 16 ; SLOW-NEXT: or a0, a1, a0 ; SLOW-NEXT: ret ; ; FAST-LABEL: load_i24: ; FAST: # %bb.0: ; FAST-NEXT: lbu a1, 2(a0) ; FAST-NEXT: lhu a0, 0(a0) ; FAST-NEXT: slli a1, a1, 16 ; FAST-NEXT: or a0, a0, a1 ; FAST-NEXT: ret %res = load i24, ptr %p, align 1 ret i24 %res } define i32 @load_i32(ptr %p) { ; SLOW-LABEL: load_i32: ; SLOW: # %bb.0: ; SLOW-NEXT: lbu a1, 1(a0) ; SLOW-NEXT: lbu a2, 0(a0) ; SLOW-NEXT: lbu a3, 2(a0) ; SLOW-NEXT: lbu a0, 3(a0) ; SLOW-NEXT: slli a1, a1, 8 ; SLOW-NEXT: or a1, a1, a2 ; SLOW-NEXT: slli a3, a3, 16 ; SLOW-NEXT: slli a0, a0, 24 ; SLOW-NEXT: or a0, a0, a3 ; SLOW-NEXT: or a0, a0, a1 ; SLOW-NEXT: ret ; ; FAST-LABEL: load_i32: ; FAST: # %bb.0: ; FAST-NEXT: lw a0, 0(a0) ; FAST-NEXT: ret %res = load i32, ptr %p, align 1 ret i32 %res } define i64 @load_i64(ptr %p) { ; RV32I-LABEL: load_i64: ; RV32I: # %bb.0: ; RV32I-NEXT: lbu a1, 1(a0) ; RV32I-NEXT: lbu a2, 0(a0) ; RV32I-NEXT: lbu a3, 2(a0) ; RV32I-NEXT: lbu a4, 3(a0) ; RV32I-NEXT: slli a1, a1, 8 ; RV32I-NEXT: or a1, a1, a2 ; RV32I-NEXT: slli a3, a3, 16 ; RV32I-NEXT: slli a4, a4, 24 ; RV32I-NEXT: or a2, a4, a3 ; RV32I-NEXT: or a2, a2, a1 ; RV32I-NEXT: lbu a1, 5(a0) ; RV32I-NEXT: lbu a3, 4(a0) ; RV32I-NEXT: lbu a4, 6(a0) ; RV32I-NEXT: lbu a0, 7(a0) ; RV32I-NEXT: slli a1, a1, 8 ; RV32I-NEXT: or a1, a1, a3 ; RV32I-NEXT: slli a4, a4, 16 ; RV32I-NEXT: slli a0, a0, 24 ; RV32I-NEXT: or a0, a0, a4 ; RV32I-NEXT: or a1, a0, a1 ; RV32I-NEXT: mv a0, a2 ; RV32I-NEXT: ret ; ; RV64I-LABEL: load_i64: ; RV64I: # %bb.0: ; RV64I-NEXT: lbu a1, 1(a0) ; RV64I-NEXT: lbu a2, 0(a0) ; RV64I-NEXT: lbu a3, 2(a0) ; RV64I-NEXT: lbu a4, 3(a0) ; RV64I-NEXT: slli a1, a1, 8 ; RV64I-NEXT: or a1, a1, a2 ; RV64I-NEXT: slli a3, a3, 16 ; RV64I-NEXT: slli a4, a4, 24 ; RV64I-NEXT: or a3, a4, a3 ; RV64I-NEXT: or a1, a3, a1 ; RV64I-NEXT: lbu a2, 5(a0) ; RV64I-NEXT: lbu a3, 4(a0) ; RV64I-NEXT: lbu a4, 6(a0) ; RV64I-NEXT: lbu a0, 7(a0) ; RV64I-NEXT: slli a2, a2, 8 ; RV64I-NEXT: or a2, a2, a3 ; RV64I-NEXT: slli a4, a4, 16 ; RV64I-NEXT: slli a0, a0, 24 ; RV64I-NEXT: or a0, a0, a4 ; RV64I-NEXT: or a0, a0, a2 ; RV64I-NEXT: slli a0, a0, 32 ; RV64I-NEXT: or a0, a0, a1 ; RV64I-NEXT: ret ; ; RV32I-FAST-LABEL: load_i64: ; RV32I-FAST: # %bb.0: ; RV32I-FAST-NEXT: lw a2, 0(a0) ; RV32I-FAST-NEXT: lw a1, 4(a0) ; RV32I-FAST-NEXT: mv a0, a2 ; RV32I-FAST-NEXT: ret ; ; RV64I-FAST-LABEL: load_i64: ; RV64I-FAST: # %bb.0: ; RV64I-FAST-NEXT: ld a0, 0(a0) ; RV64I-FAST-NEXT: ret %res = load i64, ptr %p, align 1 ret i64 %res } define void @store_i8(ptr %p, i8 %v) { ; ALL-LABEL: store_i8: ; ALL: # %bb.0: ; ALL-NEXT: sb a1, 0(a0) ; ALL-NEXT: ret store i8 %v, ptr %p, align 1 ret void } define void @store_i16(ptr %p, i16 %v) { ; SLOW-LABEL: store_i16: ; SLOW: # %bb.0: ; SLOW-NEXT: sb a1, 0(a0) ; SLOW-NEXT: srli a1, a1, 8 ; SLOW-NEXT: sb a1, 1(a0) ; SLOW-NEXT: ret ; ; FAST-LABEL: store_i16: ; FAST: # %bb.0: ; FAST-NEXT: sh a1, 0(a0) ; FAST-NEXT: ret store i16 %v, ptr %p, align 1 ret void } define void @store_i24(ptr %p, i24 %v) { ; SLOW-LABEL: store_i24: ; SLOW: # %bb.0: ; SLOW-NEXT: sb a1, 0(a0) ; SLOW-NEXT: srli a2, a1, 8 ; SLOW-NEXT: sb a2, 1(a0) ; SLOW-NEXT: srli a1, a1, 16 ; SLOW-NEXT: sb a1, 2(a0) ; SLOW-NEXT: ret ; ; FAST-LABEL: store_i24: ; FAST: # %bb.0: ; FAST-NEXT: sh a1, 0(a0) ; FAST-NEXT: srli a1, a1, 16 ; FAST-NEXT: sb a1, 2(a0) ; FAST-NEXT: ret store i24 %v, ptr %p, align 1 ret void } define void @store_i32(ptr %p, i32 %v) { ; SLOW-LABEL: store_i32: ; SLOW: # %bb.0: ; SLOW-NEXT: sb a1, 0(a0) ; SLOW-NEXT: srli a2, a1, 24 ; SLOW-NEXT: sb a2, 3(a0) ; SLOW-NEXT: srli a2, a1, 16 ; SLOW-NEXT: sb a2, 2(a0) ; SLOW-NEXT: srli a1, a1, 8 ; SLOW-NEXT: sb a1, 1(a0) ; SLOW-NEXT: ret ; ; FAST-LABEL: store_i32: ; FAST: # %bb.0: ; FAST-NEXT: sw a1, 0(a0) ; FAST-NEXT: ret store i32 %v, ptr %p, align 1 ret void } define void @store_i64(ptr %p, i64 %v) { ; RV32I-LABEL: store_i64: ; RV32I: # %bb.0: ; RV32I-NEXT: sb a2, 4(a0) ; RV32I-NEXT: sb a1, 0(a0) ; RV32I-NEXT: srli a3, a2, 24 ; RV32I-NEXT: sb a3, 7(a0) ; RV32I-NEXT: srli a3, a2, 16 ; RV32I-NEXT: sb a3, 6(a0) ; RV32I-NEXT: srli a2, a2, 8 ; RV32I-NEXT: sb a2, 5(a0) ; RV32I-NEXT: srli a2, a1, 24 ; RV32I-NEXT: sb a2, 3(a0) ; RV32I-NEXT: srli a2, a1, 16 ; RV32I-NEXT: sb a2, 2(a0) ; RV32I-NEXT: srli a1, a1, 8 ; RV32I-NEXT: sb a1, 1(a0) ; RV32I-NEXT: ret ; ; RV64I-LABEL: store_i64: ; RV64I: # %bb.0: ; RV64I-NEXT: sb a1, 0(a0) ; RV64I-NEXT: srli a2, a1, 56 ; RV64I-NEXT: sb a2, 7(a0) ; RV64I-NEXT: srli a2, a1, 48 ; RV64I-NEXT: sb a2, 6(a0) ; RV64I-NEXT: srli a2, a1, 40 ; RV64I-NEXT: sb a2, 5(a0) ; RV64I-NEXT: srli a2, a1, 32 ; RV64I-NEXT: sb a2, 4(a0) ; RV64I-NEXT: srli a2, a1, 24 ; RV64I-NEXT: sb a2, 3(a0) ; RV64I-NEXT: srli a2, a1, 16 ; RV64I-NEXT: sb a2, 2(a0) ; RV64I-NEXT: srli a1, a1, 8 ; RV64I-NEXT: sb a1, 1(a0) ; RV64I-NEXT: ret ; ; RV32I-FAST-LABEL: store_i64: ; RV32I-FAST: # %bb.0: ; RV32I-FAST-NEXT: sw a2, 4(a0) ; RV32I-FAST-NEXT: sw a1, 0(a0) ; RV32I-FAST-NEXT: ret ; ; RV64I-FAST-LABEL: store_i64: ; RV64I-FAST: # %bb.0: ; RV64I-FAST-NEXT: sd a1, 0(a0) ; RV64I-FAST-NEXT: ret store i64 %v, ptr %p, align 1 ret void } define void @merge_stores_i8_i16(ptr %p) { ; SLOW-LABEL: merge_stores_i8_i16: ; SLOW: # %bb.0: ; SLOW-NEXT: sb zero, 0(a0) ; SLOW-NEXT: sb zero, 1(a0) ; SLOW-NEXT: ret ; ; FAST-LABEL: merge_stores_i8_i16: ; FAST: # %bb.0: ; FAST-NEXT: sh zero, 0(a0) ; FAST-NEXT: ret store i8 0, ptr %p %p2 = getelementptr i8, ptr %p, i32 1 store i8 0, ptr %p2 ret void } define void @merge_stores_i8_i32(ptr %p) { ; SLOW-LABEL: merge_stores_i8_i32: ; SLOW: # %bb.0: ; SLOW-NEXT: sb zero, 0(a0) ; SLOW-NEXT: sb zero, 1(a0) ; SLOW-NEXT: sb zero, 2(a0) ; SLOW-NEXT: sb zero, 3(a0) ; SLOW-NEXT: ret ; ; FAST-LABEL: merge_stores_i8_i32: ; FAST: # %bb.0: ; FAST-NEXT: sw zero, 0(a0) ; FAST-NEXT: ret store i8 0, ptr %p %p2 = getelementptr i8, ptr %p, i32 1 store i8 0, ptr %p2 %p3 = getelementptr i8, ptr %p, i32 2 store i8 0, ptr %p3 %p4 = getelementptr i8, ptr %p, i32 3 store i8 0, ptr %p4 ret void } define void @merge_stores_i8_i64(ptr %p) { ; SLOW-LABEL: merge_stores_i8_i64: ; SLOW: # %bb.0: ; SLOW-NEXT: sb zero, 0(a0) ; SLOW-NEXT: sb zero, 1(a0) ; SLOW-NEXT: sb zero, 2(a0) ; SLOW-NEXT: sb zero, 3(a0) ; SLOW-NEXT: sb zero, 4(a0) ; SLOW-NEXT: sb zero, 5(a0) ; SLOW-NEXT: sb zero, 6(a0) ; SLOW-NEXT: sb zero, 7(a0) ; SLOW-NEXT: ret ; ; RV32I-FAST-LABEL: merge_stores_i8_i64: ; RV32I-FAST: # %bb.0: ; RV32I-FAST-NEXT: sw zero, 0(a0) ; RV32I-FAST-NEXT: sw zero, 4(a0) ; RV32I-FAST-NEXT: ret ; ; RV64I-FAST-LABEL: merge_stores_i8_i64: ; RV64I-FAST: # %bb.0: ; RV64I-FAST-NEXT: sd zero, 0(a0) ; RV64I-FAST-NEXT: ret store i8 0, ptr %p %p2 = getelementptr i8, ptr %p, i32 1 store i8 0, ptr %p2 %p3 = getelementptr i8, ptr %p, i32 2 store i8 0, ptr %p3 %p4 = getelementptr i8, ptr %p, i32 3 store i8 0, ptr %p4 %p5 = getelementptr i8, ptr %p, i32 4 store i8 0, ptr %p5 %p6 = getelementptr i8, ptr %p, i32 5 store i8 0, ptr %p6 %p7 = getelementptr i8, ptr %p, i32 6 store i8 0, ptr %p7 %p8 = getelementptr i8, ptr %p, i32 7 store i8 0, ptr %p8 ret void } define void @merge_stores_i16_i32(ptr %p) { ; SLOW-LABEL: merge_stores_i16_i32: ; SLOW: # %bb.0: ; SLOW-NEXT: sh zero, 0(a0) ; SLOW-NEXT: sh zero, 2(a0) ; SLOW-NEXT: ret ; ; FAST-LABEL: merge_stores_i16_i32: ; FAST: # %bb.0: ; FAST-NEXT: sw zero, 0(a0) ; FAST-NEXT: ret store i16 0, ptr %p %p2 = getelementptr i16, ptr %p, i32 1 store i16 0, ptr %p2 ret void } define void @merge_stores_i16_i64(ptr %p) { ; SLOW-LABEL: merge_stores_i16_i64: ; SLOW: # %bb.0: ; SLOW-NEXT: sh zero, 0(a0) ; SLOW-NEXT: sh zero, 2(a0) ; SLOW-NEXT: sh zero, 4(a0) ; SLOW-NEXT: sh zero, 6(a0) ; SLOW-NEXT: ret ; ; RV32I-FAST-LABEL: merge_stores_i16_i64: ; RV32I-FAST: # %bb.0: ; RV32I-FAST-NEXT: sw zero, 0(a0) ; RV32I-FAST-NEXT: sw zero, 4(a0) ; RV32I-FAST-NEXT: ret ; ; RV64I-FAST-LABEL: merge_stores_i16_i64: ; RV64I-FAST: # %bb.0: ; RV64I-FAST-NEXT: sd zero, 0(a0) ; RV64I-FAST-NEXT: ret store i16 0, ptr %p %p2 = getelementptr i16, ptr %p, i32 1 store i16 0, ptr %p2 %p3 = getelementptr i16, ptr %p, i32 2 store i16 0, ptr %p3 %p4 = getelementptr i16, ptr %p, i32 3 store i16 0, ptr %p4 ret void } define void @merge_stores_i32_i64(ptr %p) { ; SLOW-LABEL: merge_stores_i32_i64: ; SLOW: # %bb.0: ; SLOW-NEXT: sw zero, 0(a0) ; SLOW-NEXT: sw zero, 4(a0) ; SLOW-NEXT: ret ; ; RV32I-FAST-LABEL: merge_stores_i32_i64: ; RV32I-FAST: # %bb.0: ; RV32I-FAST-NEXT: sw zero, 0(a0) ; RV32I-FAST-NEXT: sw zero, 4(a0) ; RV32I-FAST-NEXT: ret ; ; RV64I-FAST-LABEL: merge_stores_i32_i64: ; RV64I-FAST: # %bb.0: ; RV64I-FAST-NEXT: sd zero, 0(a0) ; RV64I-FAST-NEXT: ret store i32 0, ptr %p %p2 = getelementptr i32, ptr %p, i32 1 store i32 0, ptr %p2 ret void } define void @store_large_constant(ptr %x) { ; SLOW-LABEL: store_large_constant: ; SLOW: # %bb.0: ; SLOW-NEXT: li a1, 254 ; SLOW-NEXT: sb a1, 7(a0) ; SLOW-NEXT: li a1, 220 ; SLOW-NEXT: sb a1, 6(a0) ; SLOW-NEXT: li a1, 186 ; SLOW-NEXT: sb a1, 5(a0) ; SLOW-NEXT: li a1, 152 ; SLOW-NEXT: sb a1, 4(a0) ; SLOW-NEXT: li a1, 118 ; SLOW-NEXT: sb a1, 3(a0) ; SLOW-NEXT: li a1, 84 ; SLOW-NEXT: sb a1, 2(a0) ; SLOW-NEXT: li a1, 50 ; SLOW-NEXT: sb a1, 1(a0) ; SLOW-NEXT: li a1, 16 ; SLOW-NEXT: sb a1, 0(a0) ; SLOW-NEXT: ret ; ; RV32I-FAST-LABEL: store_large_constant: ; RV32I-FAST: # %bb.0: ; RV32I-FAST-NEXT: lui a1, 1043916 ; RV32I-FAST-NEXT: addi a1, a1, -1384 ; RV32I-FAST-NEXT: sw a1, 4(a0) ; RV32I-FAST-NEXT: lui a1, 484675 ; RV32I-FAST-NEXT: addi a1, a1, 528 ; RV32I-FAST-NEXT: sw a1, 0(a0) ; RV32I-FAST-NEXT: ret ; ; RV64I-FAST-LABEL: store_large_constant: ; RV64I-FAST: # %bb.0: ; RV64I-FAST-NEXT: lui a1, %hi(.LCPI16_0) ; RV64I-FAST-NEXT: ld a1, %lo(.LCPI16_0)(a1) ; RV64I-FAST-NEXT: sd a1, 0(a0) ; RV64I-FAST-NEXT: ret store i64 18364758544493064720, ptr %x, align 1 ret void }