; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 2 ; RUN: llc -O2 -verify-machineinstrs -march=mips64 -mcpu=mips64r2 \ ; RUN: -target-abi=n64 < %s -o - | FileCheck %s -check-prefix=MIPS64R2 ; RUN: llc -O2 -verify-machineinstrs -march=mips -mcpu=mips32r2 < %s -o - \ ; RUN: | FileCheck %s -check-prefix=MIPS32R2 ; RUN: llc -O2 -verify-machineinstrs -march=mips -mattr=mips16 < %s -o - \ ; RUN: | FileCheck %s -check-prefix=MIPS16 ; RUN: llc -O2 -verify-machineinstrs -march=mips64 -mcpu=mips64r2 \ ; RUN: -target-abi=n32 < %s -o - | FileCheck %s -check-prefix=MIPS64R2N32 ; #include ; #include ; struct cvmx_buf_ptr { ; struct { ; unsigned long long addr :37; ; unsigned long long addr1 :15; ; unsigned int length:14; ; uint64_t total_bytes:16; ; uint64_t segs : 6; ; } s; ; } ; ; unsigned long long foo(volatile struct cvmx_buf_ptr bufptr) { ; bufptr.s.addr = 123; ; bufptr.s.segs = 4; ; bufptr.s.length = 5; ; bufptr.s.total_bytes = bufptr.s.length; ; return bufptr.s.addr; ; } ; Testing of selection INS/DINS instruction define i64 @f123(i64 inreg %bufptr.coerce0, i64 inreg %bufptr.coerce1) local_unnamed_addr #0 { ; MIPS64R2-LABEL: f123: ; MIPS64R2: # %bb.0: # %entry ; MIPS64R2-NEXT: daddiu $sp, $sp, -16 ; MIPS64R2-NEXT: .cfi_def_cfa_offset 16 ; MIPS64R2-NEXT: sd $4, 8($sp) ; MIPS64R2-NEXT: sd $5, 0($sp) ; MIPS64R2-NEXT: daddiu $1, $zero, 123 ; MIPS64R2-NEXT: ld $2, 8($sp) ; MIPS64R2-NEXT: dinsm $2, $1, 27, 37 ; MIPS64R2-NEXT: sd $2, 8($sp) ; MIPS64R2-NEXT: daddiu $1, $zero, 4 ; MIPS64R2-NEXT: ld $2, 0($sp) ; MIPS64R2-NEXT: dinsm $2, $1, 28, 6 ; MIPS64R2-NEXT: daddiu $1, $zero, 5 ; MIPS64R2-NEXT: sd $2, 0($sp) ; MIPS64R2-NEXT: ld $2, 0($sp) ; MIPS64R2-NEXT: dinsu $2, $1, 50, 14 ; MIPS64R2-NEXT: sd $2, 0($sp) ; MIPS64R2-NEXT: ld $1, 0($sp) ; MIPS64R2-NEXT: dsrl $1, $1, 50 ; MIPS64R2-NEXT: ld $2, 0($sp) ; MIPS64R2-NEXT: dinsu $2, $1, 34, 16 ; MIPS64R2-NEXT: sd $2, 0($sp) ; MIPS64R2-NEXT: ld $1, 8($sp) ; MIPS64R2-NEXT: dsrl $2, $1, 27 ; MIPS64R2-NEXT: jr $ra ; MIPS64R2-NEXT: daddiu $sp, $sp, 16 ; ; MIPS32R2-LABEL: f123: ; MIPS32R2: # %bb.0: # %entry ; MIPS32R2-NEXT: addiu $sp, $sp, -16 ; MIPS32R2-NEXT: .cfi_def_cfa_offset 16 ; MIPS32R2-NEXT: lw $1, 8($sp) ; MIPS32R2-NEXT: lw $1, 12($sp) ; MIPS32R2-NEXT: addiu $2, $zero, 3 ; MIPS32R2-NEXT: sw $2, 8($sp) ; MIPS32R2-NEXT: ext $1, $1, 0, 27 ; MIPS32R2-NEXT: lui $2, 55296 ; MIPS32R2-NEXT: or $1, $1, $2 ; MIPS32R2-NEXT: sw $1, 12($sp) ; MIPS32R2-NEXT: addiu $1, $zero, -4 ; MIPS32R2-NEXT: lw $2, 0($sp) ; MIPS32R2-NEXT: and $1, $2, $1 ; MIPS32R2-NEXT: lw $2, 4($sp) ; MIPS32R2-NEXT: sw $1, 0($sp) ; MIPS32R2-NEXT: ext $1, $2, 0, 28 ; MIPS32R2-NEXT: lui $2, 16384 ; MIPS32R2-NEXT: or $1, $1, $2 ; MIPS32R2-NEXT: lui $2, 20 ; MIPS32R2-NEXT: sw $1, 4($sp) ; MIPS32R2-NEXT: lw $1, 0($sp) ; MIPS32R2-NEXT: lw $3, 4($sp) ; MIPS32R2-NEXT: sw $3, 4($sp) ; MIPS32R2-NEXT: ext $1, $1, 0, 18 ; MIPS32R2-NEXT: or $1, $1, $2 ; MIPS32R2-NEXT: sw $1, 0($sp) ; MIPS32R2-NEXT: lw $1, 4($sp) ; MIPS32R2-NEXT: lw $1, 0($sp) ; MIPS32R2-NEXT: lw $2, 0($sp) ; MIPS32R2-NEXT: lw $3, 4($sp) ; MIPS32R2-NEXT: sw $3, 4($sp) ; MIPS32R2-NEXT: srl $1, $1, 18 ; MIPS32R2-NEXT: ins $2, $1, 2, 16 ; MIPS32R2-NEXT: sw $2, 0($sp) ; MIPS32R2-NEXT: lw $1, 8($sp) ; MIPS32R2-NEXT: sll $2, $1, 5 ; MIPS32R2-NEXT: lw $3, 12($sp) ; MIPS32R2-NEXT: srl $3, $3, 27 ; MIPS32R2-NEXT: or $3, $3, $2 ; MIPS32R2-NEXT: srl $2, $1, 27 ; MIPS32R2-NEXT: jr $ra ; MIPS32R2-NEXT: addiu $sp, $sp, 16 ; ; MIPS16-LABEL: f123: ; MIPS16: # %bb.0: # %entry ; MIPS16-NEXT: save 16 # 16 bit inst ; MIPS16-EMPTY: ; MIPS16-NEXT: .cfi_def_cfa_offset 16 ; MIPS16-NEXT: lw $2, 8($sp) ; MIPS16-NEXT: lw $2, 12($sp) ; MIPS16-NEXT: li $3, 3 ; MIPS16-NEXT: sw $3, 8($sp) ; MIPS16-NEXT: lw $3, $CPI0_0 # 16 bit inst ; MIPS16-NEXT: and $3, $2 ; MIPS16-NEXT: lw $2, $CPI0_1 # 16 bit inst ; MIPS16-NEXT: or $2, $3 ; MIPS16-NEXT: sw $2, 12($sp) ; MIPS16-NEXT: move $2, $zero ; MIPS16-NEXT: addiu $2, -4 ; MIPS16-NEXT: lw $3, 0($sp) ; MIPS16-NEXT: and $3, $2 ; MIPS16-NEXT: lw $2, 4($sp) ; MIPS16-NEXT: sw $3, 0($sp) ; MIPS16-NEXT: lw $3, $CPI0_2 # 16 bit inst ; MIPS16-NEXT: and $3, $2 ; MIPS16-NEXT: lw $2, $CPI0_3 # 16 bit inst ; MIPS16-NEXT: or $2, $3 ; MIPS16-NEXT: sw $2, 4($sp) ; MIPS16-NEXT: lw $2, 0($sp) ; MIPS16-NEXT: lw $3, 4($sp) ; MIPS16-NEXT: sw $3, 4($sp) ; MIPS16-NEXT: lw $3, $CPI0_4 # 16 bit inst ; MIPS16-NEXT: and $3, $2 ; MIPS16-NEXT: lw $2, $CPI0_5 # 16 bit inst ; MIPS16-NEXT: or $2, $3 ; MIPS16-NEXT: sw $2, 0($sp) ; MIPS16-NEXT: lw $2, 4($sp) ; MIPS16-NEXT: lw $2, 0($sp) ; MIPS16-NEXT: lw $3, 0($sp) ; MIPS16-NEXT: lw $4, 4($sp) ; MIPS16-NEXT: sw $4, 4($sp) ; MIPS16-NEXT: srl $2, $2, 16 ; MIPS16-NEXT: li $4, 65532 ; MIPS16-NEXT: and $4, $2 ; MIPS16-NEXT: lw $2, $CPI0_6 # 16 bit inst ; MIPS16-NEXT: and $2, $3 ; MIPS16-NEXT: or $2, $4 ; MIPS16-NEXT: sw $2, 0($sp) ; MIPS16-NEXT: lw $2, 12($sp) ; MIPS16-NEXT: srl $2, $2, 27 ; MIPS16-NEXT: lw $4, 8($sp) ; MIPS16-NEXT: sll $3, $4, 5 ; MIPS16-NEXT: or $3, $2 ; MIPS16-NEXT: srl $2, $4, 27 ; MIPS16-NEXT: restore 16 # 16 bit inst ; MIPS16-EMPTY: ; MIPS16-NEXT: jrc $ra ; MIPS16-NEXT: .p2align 2 ; MIPS16-NEXT: # %bb.1: ; MIPS16-NEXT: $CPI0_0: ; MIPS16-NEXT: .4byte 134217727 # 0x7ffffff ; MIPS16-NEXT: $CPI0_1: ; MIPS16-NEXT: .4byte 3623878656 # 0xd8000000 ; MIPS16-NEXT: $CPI0_2: ; MIPS16-NEXT: .4byte 268435455 # 0xfffffff ; MIPS16-NEXT: $CPI0_3: ; MIPS16-NEXT: .4byte 1073741824 # 0x40000000 ; MIPS16-NEXT: $CPI0_4: ; MIPS16-NEXT: .4byte 262143 # 0x3ffff ; MIPS16-NEXT: $CPI0_5: ; MIPS16-NEXT: .4byte 1310720 # 0x140000 ; MIPS16-NEXT: $CPI0_6: ; MIPS16-NEXT: .4byte 4294705155 # 0xfffc0003 ; ; MIPS64R2N32-LABEL: f123: ; MIPS64R2N32: # %bb.0: # %entry ; MIPS64R2N32-NEXT: addiu $sp, $sp, -16 ; MIPS64R2N32-NEXT: .cfi_def_cfa_offset 16 ; MIPS64R2N32-NEXT: sd $4, 8($sp) ; MIPS64R2N32-NEXT: sd $5, 0($sp) ; MIPS64R2N32-NEXT: daddiu $1, $zero, 123 ; MIPS64R2N32-NEXT: ld $2, 8($sp) ; MIPS64R2N32-NEXT: dinsm $2, $1, 27, 37 ; MIPS64R2N32-NEXT: sd $2, 8($sp) ; MIPS64R2N32-NEXT: daddiu $1, $zero, 4 ; MIPS64R2N32-NEXT: ld $2, 0($sp) ; MIPS64R2N32-NEXT: dinsm $2, $1, 28, 6 ; MIPS64R2N32-NEXT: daddiu $1, $zero, 5 ; MIPS64R2N32-NEXT: sd $2, 0($sp) ; MIPS64R2N32-NEXT: ld $2, 0($sp) ; MIPS64R2N32-NEXT: dinsu $2, $1, 50, 14 ; MIPS64R2N32-NEXT: sd $2, 0($sp) ; MIPS64R2N32-NEXT: ld $1, 0($sp) ; MIPS64R2N32-NEXT: dsrl $1, $1, 50 ; MIPS64R2N32-NEXT: ld $2, 0($sp) ; MIPS64R2N32-NEXT: dinsu $2, $1, 34, 16 ; MIPS64R2N32-NEXT: sd $2, 0($sp) ; MIPS64R2N32-NEXT: ld $1, 8($sp) ; MIPS64R2N32-NEXT: dsrl $2, $1, 27 ; MIPS64R2N32-NEXT: jr $ra ; MIPS64R2N32-NEXT: addiu $sp, $sp, 16 entry: %bufptr.sroa.0 = alloca i64, align 8 %bufptr.sroa.4 = alloca i64, align 8 store i64 %bufptr.coerce0, ptr %bufptr.sroa.0, align 8 store i64 %bufptr.coerce1, ptr %bufptr.sroa.4, align 8 %bufptr.sroa.0.0.bufptr.sroa.0.0.bufptr.sroa.0.0.bf.load = load volatile i64, ptr %bufptr.sroa.0, align 8 %bf.clear = and i64 %bufptr.sroa.0.0.bufptr.sroa.0.0.bufptr.sroa.0.0.bf.load, 134217727 %bf.set = or i64 %bf.clear, 16508780544 store volatile i64 %bf.set, ptr %bufptr.sroa.0, align 8 %bufptr.sroa.4.0.bufptr.sroa.4.0.bufptr.sroa.4.8.bf.load2 = load volatile i64, ptr %bufptr.sroa.4, align 8 %bf.clear3 = and i64 %bufptr.sroa.4.0.bufptr.sroa.4.0.bufptr.sroa.4.8.bf.load2, -16911433729 %bf.set4 = or i64 %bf.clear3, 1073741824 store volatile i64 %bf.set4, ptr %bufptr.sroa.4, align 8 %bufptr.sroa.4.0.bufptr.sroa.4.0.bufptr.sroa.4.8.bf.load6 = load volatile i64, ptr %bufptr.sroa.4, align 8 %bf.clear7 = and i64 %bufptr.sroa.4.0.bufptr.sroa.4.0.bufptr.sroa.4.8.bf.load6, 1125899906842623 %bf.set8 = or i64 %bf.clear7, 5629499534213120 store volatile i64 %bf.set8, ptr %bufptr.sroa.4, align 8 %bufptr.sroa.4.0.bufptr.sroa.4.0.bufptr.sroa.4.8.bf.load11 = load volatile i64, ptr %bufptr.sroa.4, align 8 %bf.lshr = lshr i64 %bufptr.sroa.4.0.bufptr.sroa.4.0.bufptr.sroa.4.8.bf.load11, 50 %bufptr.sroa.4.0.bufptr.sroa.4.0.bufptr.sroa.4.8.bf.load13 = load volatile i64, ptr %bufptr.sroa.4, align 8 %bf.shl = shl nuw nsw i64 %bf.lshr, 34 %bf.clear14 = and i64 %bufptr.sroa.4.0.bufptr.sroa.4.0.bufptr.sroa.4.8.bf.load13, -1125882726973441 %bf.set15 = or i64 %bf.clear14, %bf.shl store volatile i64 %bf.set15, ptr %bufptr.sroa.4, align 8 %bufptr.sroa.0.0.bufptr.sroa.0.0.bufptr.sroa.0.0.bf.load17 = load volatile i64, ptr %bufptr.sroa.0, align 8 %bf.lshr18 = lshr i64 %bufptr.sroa.0.0.bufptr.sroa.0.0.bufptr.sroa.0.0.bf.load17, 27 ret i64 %bf.lshr18 } ; int foo(volatile int x) { ; int y = x; ; y = y & -4; ; x = y | 8; ; return y; ; } define i32 @foo(i32 signext %x) { ; MIPS64R2-LABEL: foo: ; MIPS64R2: # %bb.0: # %entry ; MIPS64R2-NEXT: daddiu $sp, $sp, -16 ; MIPS64R2-NEXT: .cfi_def_cfa_offset 16 ; MIPS64R2-NEXT: sw $4, 12($sp) ; MIPS64R2-NEXT: addiu $1, $zero, -4 ; MIPS64R2-NEXT: lw $2, 12($sp) ; MIPS64R2-NEXT: and $2, $2, $1 ; MIPS64R2-NEXT: ori $1, $2, 8 ; MIPS64R2-NEXT: sw $1, 12($sp) ; MIPS64R2-NEXT: jr $ra ; MIPS64R2-NEXT: daddiu $sp, $sp, 16 ; ; MIPS32R2-LABEL: foo: ; MIPS32R2: # %bb.0: # %entry ; MIPS32R2-NEXT: addiu $sp, $sp, -8 ; MIPS32R2-NEXT: .cfi_def_cfa_offset 8 ; MIPS32R2-NEXT: sw $4, 4($sp) ; MIPS32R2-NEXT: addiu $1, $zero, -4 ; MIPS32R2-NEXT: lw $2, 4($sp) ; MIPS32R2-NEXT: and $2, $2, $1 ; MIPS32R2-NEXT: ori $1, $2, 8 ; MIPS32R2-NEXT: sw $1, 4($sp) ; MIPS32R2-NEXT: jr $ra ; MIPS32R2-NEXT: addiu $sp, $sp, 8 ; ; MIPS16-LABEL: foo: ; MIPS16: # %bb.0: # %entry ; MIPS16-NEXT: save 8 # 16 bit inst ; MIPS16-EMPTY: ; MIPS16-NEXT: .cfi_def_cfa_offset 8 ; MIPS16-NEXT: sw $4, 4($sp) ; MIPS16-NEXT: move $3, $zero ; MIPS16-NEXT: addiu $3, -4 ; MIPS16-NEXT: lw $2, 4($sp) ; MIPS16-NEXT: and $2, $3 ; MIPS16-NEXT: li $3, 8 ; MIPS16-NEXT: or $3, $2 ; MIPS16-NEXT: sw $3, 4($sp) ; MIPS16-NEXT: restore 8 # 16 bit inst ; MIPS16-EMPTY: ; MIPS16-NEXT: jrc $ra ; ; MIPS64R2N32-LABEL: foo: ; MIPS64R2N32: # %bb.0: # %entry ; MIPS64R2N32-NEXT: addiu $sp, $sp, -16 ; MIPS64R2N32-NEXT: .cfi_def_cfa_offset 16 ; MIPS64R2N32-NEXT: sw $4, 12($sp) ; MIPS64R2N32-NEXT: addiu $1, $zero, -4 ; MIPS64R2N32-NEXT: lw $2, 12($sp) ; MIPS64R2N32-NEXT: and $2, $2, $1 ; MIPS64R2N32-NEXT: ori $1, $2, 8 ; MIPS64R2N32-NEXT: sw $1, 12($sp) ; MIPS64R2N32-NEXT: jr $ra ; MIPS64R2N32-NEXT: addiu $sp, $sp, 16 entry: %x.addr = alloca i32, align 4 store volatile i32 %x, ptr %x.addr, align 4 %x.addr.0.x.addr.0. = load volatile i32, ptr %x.addr, align 4 %and = and i32 %x.addr.0.x.addr.0., -4 %or = or i32 %and, 8 store volatile i32 %or, ptr %x.addr, align 4 ret i32 %and }