; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py ; RUN: llc -mtriple=riscv32 -verify-machineinstrs -mattr=+v < %s \ ; RUN: | FileCheck %s -check-prefix=RV32I ; RUN: llc -mtriple=riscv64 -verify-machineinstrs -mattr=+v < %s \ ; RUN: | FileCheck %s -check-prefix=RV64I define i32 @and_add_lsr(i32 %x, i32 %y) { ; RV32I-LABEL: and_add_lsr: ; RV32I: # %bb.0: ; RV32I-NEXT: addi a0, a0, -1 ; RV32I-NEXT: srli a1, a1, 20 ; RV32I-NEXT: and a0, a1, a0 ; RV32I-NEXT: ret ; ; RV64I-LABEL: and_add_lsr: ; RV64I: # %bb.0: ; RV64I-NEXT: addiw a0, a0, -1 ; RV64I-NEXT: srliw a1, a1, 20 ; RV64I-NEXT: and a0, a1, a0 ; RV64I-NEXT: ret %1 = add i32 %x, 4095 %2 = lshr i32 %y, 20 %r = and i32 %2, %1 ret i32 %r } ; Make sure we don't crash on fixed length vectors define <2 x i32> @and_add_lsr_vec(<2 x i32> %x, <2 x i32> %y) { ; RV32I-LABEL: and_add_lsr_vec: ; RV32I: # %bb.0: ; RV32I-NEXT: lui a0, 1 ; RV32I-NEXT: addi a0, a0, -1 ; RV32I-NEXT: vsetivli zero, 2, e32, mf2, ta, ma ; RV32I-NEXT: vadd.vx v8, v8, a0 ; RV32I-NEXT: vsrl.vi v9, v9, 20 ; RV32I-NEXT: vand.vv v8, v9, v8 ; RV32I-NEXT: ret ; ; RV64I-LABEL: and_add_lsr_vec: ; RV64I: # %bb.0: ; RV64I-NEXT: lui a0, 1 ; RV64I-NEXT: addi a0, a0, -1 ; RV64I-NEXT: vsetivli zero, 2, e32, mf2, ta, ma ; RV64I-NEXT: vadd.vx v8, v8, a0 ; RV64I-NEXT: vsrl.vi v9, v9, 20 ; RV64I-NEXT: vand.vv v8, v9, v8 ; RV64I-NEXT: ret %1 = add <2 x i32> %x, splat (i32 4095) %2 = lshr <2 x i32> %y, splat (i32 20) %r = and <2 x i32> %2, %1 ret <2 x i32> %r } ; Make sure we don't crash on scalable vectors define @and_add_lsr_vec2( %x, %y) { ; RV32I-LABEL: and_add_lsr_vec2: ; RV32I: # %bb.0: ; RV32I-NEXT: lui a0, 1 ; RV32I-NEXT: addi a0, a0, -1 ; RV32I-NEXT: vsetvli a1, zero, e32, m1, ta, ma ; RV32I-NEXT: vadd.vx v8, v8, a0 ; RV32I-NEXT: vsrl.vi v9, v9, 20 ; RV32I-NEXT: vand.vv v8, v9, v8 ; RV32I-NEXT: ret ; ; RV64I-LABEL: and_add_lsr_vec2: ; RV64I: # %bb.0: ; RV64I-NEXT: lui a0, 1 ; RV64I-NEXT: addi a0, a0, -1 ; RV64I-NEXT: vsetvli a1, zero, e32, m1, ta, ma ; RV64I-NEXT: vadd.vx v8, v8, a0 ; RV64I-NEXT: vsrl.vi v9, v9, 20 ; RV64I-NEXT: vand.vv v8, v9, v8 ; RV64I-NEXT: ret %1 = add %x, splat (i32 4095) %2 = lshr %y, splat (i32 20) %r = and %2, %1 ret %r }