; NOTE: Assertions have been autogenerated by utils/update_test_checks.py ; RUN: opt < %s -mtriple=riscv32 -mattr=+v -interleaved-access -S | FileCheck %s --check-prefix=RV32 ; RUN: opt < %s -mtriple=riscv64 -mattr=+v -interleaved-access -S | FileCheck %s --check-prefix=RV64 ; RUN: opt < %s -mtriple=riscv32 -mattr=+v -passes=interleaved-access -S | FileCheck %s --check-prefix=RV32 ; RUN: opt < %s -mtriple=riscv64 -mattr=+v -passes=interleaved-access -S | FileCheck %s --check-prefix=RV64 define void @load_factor2(ptr %ptr) { ; RV32-LABEL: @load_factor2( ; RV32-NEXT: [[TMP1:%.*]] = call { <8 x i32>, <8 x i32> } @llvm.riscv.seg2.load.mask.v8i32.p0.i32(ptr [[PTR:%.*]], <8 x i1> splat (i1 true), i32 8) ; RV32-NEXT: [[TMP2:%.*]] = extractvalue { <8 x i32>, <8 x i32> } [[TMP1]], 1 ; RV32-NEXT: [[TMP3:%.*]] = extractvalue { <8 x i32>, <8 x i32> } [[TMP1]], 0 ; RV32-NEXT: ret void ; ; RV64-LABEL: @load_factor2( ; RV64-NEXT: [[TMP1:%.*]] = call { <8 x i32>, <8 x i32> } @llvm.riscv.seg2.load.mask.v8i32.p0.i64(ptr [[PTR:%.*]], <8 x i1> splat (i1 true), i64 8) ; RV64-NEXT: [[TMP2:%.*]] = extractvalue { <8 x i32>, <8 x i32> } [[TMP1]], 1 ; RV64-NEXT: [[TMP3:%.*]] = extractvalue { <8 x i32>, <8 x i32> } [[TMP1]], 0 ; RV64-NEXT: ret void ; %interleaved.vec = load <16 x i32>, ptr %ptr %v0 = shufflevector <16 x i32> %interleaved.vec, <16 x i32> poison, <8 x i32> %v1 = shufflevector <16 x i32> %interleaved.vec, <16 x i32> poison, <8 x i32> ret void } define void @load_factor2_vscale(ptr %ptr) { ; RV32-LABEL: @load_factor2_vscale( ; RV32-NEXT: [[TMP1:%.*]] = call target("riscv.vector.tuple", , 2) @llvm.riscv.vlseg2.mask.triscv.vector.tuple_nxv32i8_2t.p0.nxv8i1.i32(target("riscv.vector.tuple", , 2) poison, ptr [[PTR:%.*]], splat (i1 true), i32 -1, i32 3, i32 5) ; RV32-NEXT: [[TMP2:%.*]] = call @llvm.riscv.tuple.extract.nxv8i32.triscv.vector.tuple_nxv32i8_2t(target("riscv.vector.tuple", , 2) [[TMP1]], i32 0) ; RV32-NEXT: [[TMP3:%.*]] = insertvalue { , } poison, [[TMP2]], 0 ; RV32-NEXT: [[TMP4:%.*]] = call @llvm.riscv.tuple.extract.nxv8i32.triscv.vector.tuple_nxv32i8_2t(target("riscv.vector.tuple", , 2) [[TMP1]], i32 1) ; RV32-NEXT: [[TMP5:%.*]] = insertvalue { , } [[TMP3]], [[TMP4]], 1 ; RV32-NEXT: [[TMP6:%.*]] = extractvalue { , } [[TMP5]], 0 ; RV32-NEXT: [[TMP7:%.*]] = extractvalue { , } [[TMP5]], 1 ; RV32-NEXT: ret void ; ; RV64-LABEL: @load_factor2_vscale( ; RV64-NEXT: [[TMP1:%.*]] = call target("riscv.vector.tuple", , 2) @llvm.riscv.vlseg2.mask.triscv.vector.tuple_nxv32i8_2t.p0.nxv8i1.i64(target("riscv.vector.tuple", , 2) poison, ptr [[PTR:%.*]], splat (i1 true), i64 -1, i64 3, i64 5) ; RV64-NEXT: [[TMP2:%.*]] = call @llvm.riscv.tuple.extract.nxv8i32.triscv.vector.tuple_nxv32i8_2t(target("riscv.vector.tuple", , 2) [[TMP1]], i32 0) ; RV64-NEXT: [[TMP3:%.*]] = insertvalue { , } poison, [[TMP2]], 0 ; RV64-NEXT: [[TMP4:%.*]] = call @llvm.riscv.tuple.extract.nxv8i32.triscv.vector.tuple_nxv32i8_2t(target("riscv.vector.tuple", , 2) [[TMP1]], i32 1) ; RV64-NEXT: [[TMP5:%.*]] = insertvalue { , } [[TMP3]], [[TMP4]], 1 ; RV64-NEXT: [[TMP6:%.*]] = extractvalue { , } [[TMP5]], 0 ; RV64-NEXT: [[TMP7:%.*]] = extractvalue { , } [[TMP5]], 1 ; RV64-NEXT: ret void ; %interleaved.vec = load , ptr %ptr %v = call { , } @llvm.vector.deinterleave2.nxv16i32( %interleaved.vec) %t0 = extractvalue { , } %v, 0 %t1 = extractvalue { , } %v, 1 ret void } define void @load_factor3(ptr %ptr) { ; RV32-LABEL: @load_factor3( ; RV32-NEXT: [[TMP1:%.*]] = call { <4 x i32>, <4 x i32>, <4 x i32> } @llvm.riscv.seg3.load.mask.v4i32.p0.i32(ptr [[PTR:%.*]], <4 x i1> splat (i1 true), i32 4) ; RV32-NEXT: [[TMP2:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 2 ; RV32-NEXT: [[TMP3:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 1 ; RV32-NEXT: [[TMP4:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 0 ; RV32-NEXT: ret void ; ; RV64-LABEL: @load_factor3( ; RV64-NEXT: [[TMP1:%.*]] = call { <4 x i32>, <4 x i32>, <4 x i32> } @llvm.riscv.seg3.load.mask.v4i32.p0.i64(ptr [[PTR:%.*]], <4 x i1> splat (i1 true), i64 4) ; RV64-NEXT: [[TMP2:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 2 ; RV64-NEXT: [[TMP3:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 1 ; RV64-NEXT: [[TMP4:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 0 ; RV64-NEXT: ret void ; %interleaved.vec = load <12 x i32>, ptr %ptr %v0 = shufflevector <12 x i32> %interleaved.vec, <12 x i32> poison, <4 x i32> %v1 = shufflevector <12 x i32> %interleaved.vec, <12 x i32> poison, <4 x i32> %v2 = shufflevector <12 x i32> %interleaved.vec, <12 x i32> poison, <4 x i32> ret void } define void @load_factor3_vscale(ptr %ptr) { ; RV32-LABEL: @load_factor3_vscale( ; RV32-NEXT: [[TMP1:%.*]] = call target("riscv.vector.tuple", , 3) @llvm.riscv.vlseg3.mask.triscv.vector.tuple_nxv8i8_3t.p0.nxv2i1.i32(target("riscv.vector.tuple", , 3) poison, ptr [[PTR:%.*]], splat (i1 true), i32 -1, i32 3, i32 5) ; RV32-NEXT: [[TMP2:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_3t(target("riscv.vector.tuple", , 3) [[TMP1]], i32 0) ; RV32-NEXT: [[TMP3:%.*]] = insertvalue { , , } poison, [[TMP2]], 0 ; RV32-NEXT: [[TMP4:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_3t(target("riscv.vector.tuple", , 3) [[TMP1]], i32 1) ; RV32-NEXT: [[TMP5:%.*]] = insertvalue { , , } [[TMP3]], [[TMP4]], 1 ; RV32-NEXT: [[TMP6:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_3t(target("riscv.vector.tuple", , 3) [[TMP1]], i32 2) ; RV32-NEXT: [[V:%.*]] = insertvalue { , , } [[TMP5]], [[TMP6]], 2 ; RV32-NEXT: [[T0:%.*]] = extractvalue { , , } [[V]], 0 ; RV32-NEXT: [[T1:%.*]] = extractvalue { , , } [[V]], 1 ; RV32-NEXT: [[T2:%.*]] = extractvalue { , , } [[V]], 2 ; RV32-NEXT: ret void ; ; RV64-LABEL: @load_factor3_vscale( ; RV64-NEXT: [[TMP1:%.*]] = call target("riscv.vector.tuple", , 3) @llvm.riscv.vlseg3.mask.triscv.vector.tuple_nxv8i8_3t.p0.nxv2i1.i64(target("riscv.vector.tuple", , 3) poison, ptr [[PTR:%.*]], splat (i1 true), i64 -1, i64 3, i64 5) ; RV64-NEXT: [[TMP2:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_3t(target("riscv.vector.tuple", , 3) [[TMP1]], i32 0) ; RV64-NEXT: [[TMP3:%.*]] = insertvalue { , , } poison, [[TMP2]], 0 ; RV64-NEXT: [[TMP4:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_3t(target("riscv.vector.tuple", , 3) [[TMP1]], i32 1) ; RV64-NEXT: [[TMP5:%.*]] = insertvalue { , , } [[TMP3]], [[TMP4]], 1 ; RV64-NEXT: [[TMP6:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_3t(target("riscv.vector.tuple", , 3) [[TMP1]], i32 2) ; RV64-NEXT: [[V:%.*]] = insertvalue { , , } [[TMP5]], [[TMP6]], 2 ; RV64-NEXT: [[T0:%.*]] = extractvalue { , , } [[V]], 0 ; RV64-NEXT: [[T1:%.*]] = extractvalue { , , } [[V]], 1 ; RV64-NEXT: [[T2:%.*]] = extractvalue { , , } [[V]], 2 ; RV64-NEXT: ret void ; %interleaved.vec = load , ptr %ptr %v = call { , , } @llvm.vector.deinterleave3.nxv6i32( %interleaved.vec) %t0 = extractvalue { , , } %v, 0 %t1 = extractvalue { , , } %v, 1 %t2 = extractvalue { , , } %v, 2 ret void } define void @load_factor4(ptr %ptr) { ; RV32-LABEL: @load_factor4( ; RV32-NEXT: [[TMP1:%.*]] = call { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } @llvm.riscv.seg4.load.mask.v4i32.p0.i32(ptr [[PTR:%.*]], <4 x i1> splat (i1 true), i32 4) ; RV32-NEXT: [[TMP2:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 3 ; RV32-NEXT: [[TMP3:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 2 ; RV32-NEXT: [[TMP4:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 1 ; RV32-NEXT: [[TMP5:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 0 ; RV32-NEXT: ret void ; ; RV64-LABEL: @load_factor4( ; RV64-NEXT: [[TMP1:%.*]] = call { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } @llvm.riscv.seg4.load.mask.v4i32.p0.i64(ptr [[PTR:%.*]], <4 x i1> splat (i1 true), i64 4) ; RV64-NEXT: [[TMP2:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 3 ; RV64-NEXT: [[TMP3:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 2 ; RV64-NEXT: [[TMP4:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 1 ; RV64-NEXT: [[TMP5:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 0 ; RV64-NEXT: ret void ; %interleaved.vec = load <16 x i32>, ptr %ptr %v0 = shufflevector <16 x i32> %interleaved.vec, <16 x i32> poison, <4 x i32> %v1 = shufflevector <16 x i32> %interleaved.vec, <16 x i32> poison, <4 x i32> %v2 = shufflevector <16 x i32> %interleaved.vec, <16 x i32> poison, <4 x i32> %v3 = shufflevector <16 x i32> %interleaved.vec, <16 x i32> poison, <4 x i32> ret void } define void @load_factor4_vscale(ptr %ptr) { ; RV32-LABEL: @load_factor4_vscale( ; RV32-NEXT: [[TMP1:%.*]] = call target("riscv.vector.tuple", , 4) @llvm.riscv.vlseg4.mask.triscv.vector.tuple_nxv16i8_4t.p0.nxv4i1.i32(target("riscv.vector.tuple", , 4) poison, ptr [[PTR:%.*]], splat (i1 true), i32 -1, i32 3, i32 5) ; RV32-NEXT: [[TMP2:%.*]] = call @llvm.riscv.tuple.extract.nxv4i32.triscv.vector.tuple_nxv16i8_4t(target("riscv.vector.tuple", , 4) [[TMP1]], i32 0) ; RV32-NEXT: [[TMP3:%.*]] = insertvalue { , , , } poison, [[TMP2]], 0 ; RV32-NEXT: [[TMP4:%.*]] = call @llvm.riscv.tuple.extract.nxv4i32.triscv.vector.tuple_nxv16i8_4t(target("riscv.vector.tuple", , 4) [[TMP1]], i32 1) ; RV32-NEXT: [[TMP5:%.*]] = insertvalue { , , , } [[TMP3]], [[TMP4]], 1 ; RV32-NEXT: [[TMP6:%.*]] = call @llvm.riscv.tuple.extract.nxv4i32.triscv.vector.tuple_nxv16i8_4t(target("riscv.vector.tuple", , 4) [[TMP1]], i32 2) ; RV32-NEXT: [[TMP7:%.*]] = insertvalue { , , , } [[TMP5]], [[TMP6]], 2 ; RV32-NEXT: [[TMP8:%.*]] = call @llvm.riscv.tuple.extract.nxv4i32.triscv.vector.tuple_nxv16i8_4t(target("riscv.vector.tuple", , 4) [[TMP1]], i32 3) ; RV32-NEXT: [[TMP9:%.*]] = insertvalue { , , , } [[TMP7]], [[TMP8]], 3 ; RV32-NEXT: [[TMP10:%.*]] = extractvalue { , , , } [[TMP9]], 0 ; RV32-NEXT: [[TMP11:%.*]] = extractvalue { , , , } [[TMP9]], 1 ; RV32-NEXT: [[TMP12:%.*]] = extractvalue { , , , } [[TMP9]], 2 ; RV32-NEXT: [[TMP13:%.*]] = extractvalue { , , , } [[TMP9]], 3 ; RV32-NEXT: ret void ; ; RV64-LABEL: @load_factor4_vscale( ; RV64-NEXT: [[TMP1:%.*]] = call target("riscv.vector.tuple", , 4) @llvm.riscv.vlseg4.mask.triscv.vector.tuple_nxv16i8_4t.p0.nxv4i1.i64(target("riscv.vector.tuple", , 4) poison, ptr [[PTR:%.*]], splat (i1 true), i64 -1, i64 3, i64 5) ; RV64-NEXT: [[TMP2:%.*]] = call @llvm.riscv.tuple.extract.nxv4i32.triscv.vector.tuple_nxv16i8_4t(target("riscv.vector.tuple", , 4) [[TMP1]], i32 0) ; RV64-NEXT: [[TMP3:%.*]] = insertvalue { , , , } poison, [[TMP2]], 0 ; RV64-NEXT: [[TMP4:%.*]] = call @llvm.riscv.tuple.extract.nxv4i32.triscv.vector.tuple_nxv16i8_4t(target("riscv.vector.tuple", , 4) [[TMP1]], i32 1) ; RV64-NEXT: [[TMP5:%.*]] = insertvalue { , , , } [[TMP3]], [[TMP4]], 1 ; RV64-NEXT: [[TMP6:%.*]] = call @llvm.riscv.tuple.extract.nxv4i32.triscv.vector.tuple_nxv16i8_4t(target("riscv.vector.tuple", , 4) [[TMP1]], i32 2) ; RV64-NEXT: [[TMP7:%.*]] = insertvalue { , , , } [[TMP5]], [[TMP6]], 2 ; RV64-NEXT: [[TMP8:%.*]] = call @llvm.riscv.tuple.extract.nxv4i32.triscv.vector.tuple_nxv16i8_4t(target("riscv.vector.tuple", , 4) [[TMP1]], i32 3) ; RV64-NEXT: [[TMP9:%.*]] = insertvalue { , , , } [[TMP7]], [[TMP8]], 3 ; RV64-NEXT: [[TMP10:%.*]] = extractvalue { , , , } [[TMP9]], 0 ; RV64-NEXT: [[TMP11:%.*]] = extractvalue { , , , } [[TMP9]], 1 ; RV64-NEXT: [[TMP12:%.*]] = extractvalue { , , , } [[TMP9]], 2 ; RV64-NEXT: [[TMP13:%.*]] = extractvalue { , , , } [[TMP9]], 3 ; RV64-NEXT: ret void ; %interleaved.vec = load , ptr %ptr %v = call { , , , } @llvm.vector.deinterleave4.nxv16i32( %interleaved.vec) %t0 = extractvalue { , , , } %v, 0 %t1 = extractvalue { , , , } %v, 1 %t2 = extractvalue { , , , } %v, 2 %t3 = extractvalue { , , , } %v, 3 ret void } define void @load_factor5(ptr %ptr) { ; RV32-LABEL: @load_factor5( ; RV32-NEXT: [[TMP1:%.*]] = call { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } @llvm.riscv.seg5.load.mask.v4i32.p0.i32(ptr [[PTR:%.*]], <4 x i1> splat (i1 true), i32 4) ; RV32-NEXT: [[TMP2:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 4 ; RV32-NEXT: [[TMP3:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 3 ; RV32-NEXT: [[TMP4:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 2 ; RV32-NEXT: [[TMP5:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 1 ; RV32-NEXT: [[TMP6:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 0 ; RV32-NEXT: ret void ; ; RV64-LABEL: @load_factor5( ; RV64-NEXT: [[TMP1:%.*]] = call { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } @llvm.riscv.seg5.load.mask.v4i32.p0.i64(ptr [[PTR:%.*]], <4 x i1> splat (i1 true), i64 4) ; RV64-NEXT: [[TMP2:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 4 ; RV64-NEXT: [[TMP3:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 3 ; RV64-NEXT: [[TMP4:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 2 ; RV64-NEXT: [[TMP5:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 1 ; RV64-NEXT: [[TMP6:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 0 ; RV64-NEXT: ret void ; %interleaved.vec = load <20 x i32>, ptr %ptr %v0 = shufflevector <20 x i32> %interleaved.vec, <20 x i32> poison, <4 x i32> %v1 = shufflevector <20 x i32> %interleaved.vec, <20 x i32> poison, <4 x i32> %v2 = shufflevector <20 x i32> %interleaved.vec, <20 x i32> poison, <4 x i32> %v3 = shufflevector <20 x i32> %interleaved.vec, <20 x i32> poison, <4 x i32> %v4 = shufflevector <20 x i32> %interleaved.vec, <20 x i32> poison, <4 x i32> ret void } define void @load_factor5_vscale(ptr %ptr) { ; RV32-LABEL: @load_factor5_vscale( ; RV32-NEXT: [[TMP1:%.*]] = call target("riscv.vector.tuple", , 5) @llvm.riscv.vlseg5.mask.triscv.vector.tuple_nxv8i8_5t.p0.nxv2i1.i32(target("riscv.vector.tuple", , 5) poison, ptr [[PTR:%.*]], splat (i1 true), i32 -1, i32 3, i32 5) ; RV32-NEXT: [[TMP2:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_5t(target("riscv.vector.tuple", , 5) [[TMP1]], i32 0) ; RV32-NEXT: [[TMP3:%.*]] = insertvalue { , , , , } poison, [[TMP2]], 0 ; RV32-NEXT: [[TMP4:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_5t(target("riscv.vector.tuple", , 5) [[TMP1]], i32 1) ; RV32-NEXT: [[TMP5:%.*]] = insertvalue { , , , , } [[TMP3]], [[TMP4]], 1 ; RV32-NEXT: [[TMP6:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_5t(target("riscv.vector.tuple", , 5) [[TMP1]], i32 2) ; RV32-NEXT: [[TMP7:%.*]] = insertvalue { , , , , } [[TMP5]], [[TMP6]], 2 ; RV32-NEXT: [[TMP8:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_5t(target("riscv.vector.tuple", , 5) [[TMP1]], i32 3) ; RV32-NEXT: [[TMP9:%.*]] = insertvalue { , , , , } [[TMP7]], [[TMP8]], 3 ; RV32-NEXT: [[TMP10:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_5t(target("riscv.vector.tuple", , 5) [[TMP1]], i32 4) ; RV32-NEXT: [[V:%.*]] = insertvalue { , , , , } [[TMP9]], [[TMP10]], 4 ; RV32-NEXT: [[T0:%.*]] = extractvalue { , , , , } [[V]], 0 ; RV32-NEXT: [[T1:%.*]] = extractvalue { , , , , } [[V]], 1 ; RV32-NEXT: [[T2:%.*]] = extractvalue { , , , , } [[V]], 2 ; RV32-NEXT: [[T3:%.*]] = extractvalue { , , , , } [[V]], 3 ; RV32-NEXT: [[T4:%.*]] = extractvalue { , , , , } [[V]], 4 ; RV32-NEXT: ret void ; ; RV64-LABEL: @load_factor5_vscale( ; RV64-NEXT: [[TMP1:%.*]] = call target("riscv.vector.tuple", , 5) @llvm.riscv.vlseg5.mask.triscv.vector.tuple_nxv8i8_5t.p0.nxv2i1.i64(target("riscv.vector.tuple", , 5) poison, ptr [[PTR:%.*]], splat (i1 true), i64 -1, i64 3, i64 5) ; RV64-NEXT: [[TMP2:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_5t(target("riscv.vector.tuple", , 5) [[TMP1]], i32 0) ; RV64-NEXT: [[TMP3:%.*]] = insertvalue { , , , , } poison, [[TMP2]], 0 ; RV64-NEXT: [[TMP4:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_5t(target("riscv.vector.tuple", , 5) [[TMP1]], i32 1) ; RV64-NEXT: [[TMP5:%.*]] = insertvalue { , , , , } [[TMP3]], [[TMP4]], 1 ; RV64-NEXT: [[TMP6:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_5t(target("riscv.vector.tuple", , 5) [[TMP1]], i32 2) ; RV64-NEXT: [[TMP7:%.*]] = insertvalue { , , , , } [[TMP5]], [[TMP6]], 2 ; RV64-NEXT: [[TMP8:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_5t(target("riscv.vector.tuple", , 5) [[TMP1]], i32 3) ; RV64-NEXT: [[TMP9:%.*]] = insertvalue { , , , , } [[TMP7]], [[TMP8]], 3 ; RV64-NEXT: [[TMP10:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_5t(target("riscv.vector.tuple", , 5) [[TMP1]], i32 4) ; RV64-NEXT: [[V:%.*]] = insertvalue { , , , , } [[TMP9]], [[TMP10]], 4 ; RV64-NEXT: [[T0:%.*]] = extractvalue { , , , , } [[V]], 0 ; RV64-NEXT: [[T1:%.*]] = extractvalue { , , , , } [[V]], 1 ; RV64-NEXT: [[T2:%.*]] = extractvalue { , , , , } [[V]], 2 ; RV64-NEXT: [[T3:%.*]] = extractvalue { , , , , } [[V]], 3 ; RV64-NEXT: [[T4:%.*]] = extractvalue { , , , , } [[V]], 4 ; RV64-NEXT: ret void ; %interleaved.vec = load , ptr %ptr %v = call { , , , , } @llvm.vector.deinterleave5.nxv10i32( %interleaved.vec) %t0 = extractvalue { , , , , } %v, 0 %t1 = extractvalue { , , , , } %v, 1 %t2 = extractvalue { , , , , } %v, 2 %t3 = extractvalue { , , , , } %v, 3 %t4 = extractvalue { , , , , } %v, 4 ret void } define void @load_factor6(ptr %ptr) { ; RV32-LABEL: @load_factor6( ; RV32-NEXT: [[TMP1:%.*]] = call { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } @llvm.riscv.seg6.load.mask.v4i32.p0.i32(ptr [[PTR:%.*]], <4 x i1> splat (i1 true), i32 4) ; RV32-NEXT: [[TMP2:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 5 ; RV32-NEXT: [[TMP3:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 4 ; RV32-NEXT: [[TMP4:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 3 ; RV32-NEXT: [[TMP5:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 2 ; RV32-NEXT: [[TMP6:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 1 ; RV32-NEXT: [[TMP7:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 0 ; RV32-NEXT: ret void ; ; RV64-LABEL: @load_factor6( ; RV64-NEXT: [[TMP1:%.*]] = call { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } @llvm.riscv.seg6.load.mask.v4i32.p0.i64(ptr [[PTR:%.*]], <4 x i1> splat (i1 true), i64 4) ; RV64-NEXT: [[TMP2:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 5 ; RV64-NEXT: [[TMP3:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 4 ; RV64-NEXT: [[TMP4:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 3 ; RV64-NEXT: [[TMP5:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 2 ; RV64-NEXT: [[TMP6:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 1 ; RV64-NEXT: [[TMP7:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 0 ; RV64-NEXT: ret void ; %interleaved.vec = load <24 x i32>, ptr %ptr %v0 = shufflevector <24 x i32> %interleaved.vec, <24 x i32> poison, <4 x i32> %v1 = shufflevector <24 x i32> %interleaved.vec, <24 x i32> poison, <4 x i32> %v2 = shufflevector <24 x i32> %interleaved.vec, <24 x i32> poison, <4 x i32> %v3 = shufflevector <24 x i32> %interleaved.vec, <24 x i32> poison, <4 x i32> %v4 = shufflevector <24 x i32> %interleaved.vec, <24 x i32> poison, <4 x i32> %v5 = shufflevector <24 x i32> %interleaved.vec, <24 x i32> poison, <4 x i32> ret void } define void @load_factor6_vscale(ptr %ptr) { ; RV32-LABEL: @load_factor6_vscale( ; RV32-NEXT: [[TMP1:%.*]] = call target("riscv.vector.tuple", , 6) @llvm.riscv.vlseg6.mask.triscv.vector.tuple_nxv8i8_6t.p0.nxv2i1.i32(target("riscv.vector.tuple", , 6) poison, ptr [[PTR:%.*]], splat (i1 true), i32 -1, i32 3, i32 5) ; RV32-NEXT: [[TMP2:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_6t(target("riscv.vector.tuple", , 6) [[TMP1]], i32 0) ; RV32-NEXT: [[TMP3:%.*]] = insertvalue { , , , , , } poison, [[TMP2]], 0 ; RV32-NEXT: [[TMP4:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_6t(target("riscv.vector.tuple", , 6) [[TMP1]], i32 1) ; RV32-NEXT: [[TMP5:%.*]] = insertvalue { , , , , , } [[TMP3]], [[TMP4]], 1 ; RV32-NEXT: [[TMP6:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_6t(target("riscv.vector.tuple", , 6) [[TMP1]], i32 2) ; RV32-NEXT: [[TMP7:%.*]] = insertvalue { , , , , , } [[TMP5]], [[TMP6]], 2 ; RV32-NEXT: [[TMP8:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_6t(target("riscv.vector.tuple", , 6) [[TMP1]], i32 3) ; RV32-NEXT: [[TMP9:%.*]] = insertvalue { , , , , , } [[TMP7]], [[TMP8]], 3 ; RV32-NEXT: [[TMP10:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_6t(target("riscv.vector.tuple", , 6) [[TMP1]], i32 4) ; RV32-NEXT: [[TMP11:%.*]] = insertvalue { , , , , , } [[TMP9]], [[TMP10]], 4 ; RV32-NEXT: [[TMP12:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_6t(target("riscv.vector.tuple", , 6) [[TMP1]], i32 5) ; RV32-NEXT: [[TMP13:%.*]] = insertvalue { , , , , , } [[TMP11]], [[TMP12]], 5 ; RV32-NEXT: [[TMP14:%.*]] = extractvalue { , , , , , } [[TMP13]], 0 ; RV32-NEXT: [[TMP15:%.*]] = extractvalue { , , , , , } [[TMP13]], 1 ; RV32-NEXT: [[TMP16:%.*]] = extractvalue { , , , , , } [[TMP13]], 2 ; RV32-NEXT: [[TMP17:%.*]] = extractvalue { , , , , , } [[TMP13]], 3 ; RV32-NEXT: [[TMP18:%.*]] = extractvalue { , , , , , } [[TMP13]], 4 ; RV32-NEXT: [[TMP19:%.*]] = extractvalue { , , , , , } [[TMP13]], 5 ; RV32-NEXT: ret void ; ; RV64-LABEL: @load_factor6_vscale( ; RV64-NEXT: [[TMP1:%.*]] = call target("riscv.vector.tuple", , 6) @llvm.riscv.vlseg6.mask.triscv.vector.tuple_nxv8i8_6t.p0.nxv2i1.i64(target("riscv.vector.tuple", , 6) poison, ptr [[PTR:%.*]], splat (i1 true), i64 -1, i64 3, i64 5) ; RV64-NEXT: [[TMP2:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_6t(target("riscv.vector.tuple", , 6) [[TMP1]], i32 0) ; RV64-NEXT: [[TMP3:%.*]] = insertvalue { , , , , , } poison, [[TMP2]], 0 ; RV64-NEXT: [[TMP4:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_6t(target("riscv.vector.tuple", , 6) [[TMP1]], i32 1) ; RV64-NEXT: [[TMP5:%.*]] = insertvalue { , , , , , } [[TMP3]], [[TMP4]], 1 ; RV64-NEXT: [[TMP6:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_6t(target("riscv.vector.tuple", , 6) [[TMP1]], i32 2) ; RV64-NEXT: [[TMP7:%.*]] = insertvalue { , , , , , } [[TMP5]], [[TMP6]], 2 ; RV64-NEXT: [[TMP8:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_6t(target("riscv.vector.tuple", , 6) [[TMP1]], i32 3) ; RV64-NEXT: [[TMP9:%.*]] = insertvalue { , , , , , } [[TMP7]], [[TMP8]], 3 ; RV64-NEXT: [[TMP10:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_6t(target("riscv.vector.tuple", , 6) [[TMP1]], i32 4) ; RV64-NEXT: [[TMP11:%.*]] = insertvalue { , , , , , } [[TMP9]], [[TMP10]], 4 ; RV64-NEXT: [[TMP12:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_6t(target("riscv.vector.tuple", , 6) [[TMP1]], i32 5) ; RV64-NEXT: [[TMP13:%.*]] = insertvalue { , , , , , } [[TMP11]], [[TMP12]], 5 ; RV64-NEXT: [[TMP14:%.*]] = extractvalue { , , , , , } [[TMP13]], 0 ; RV64-NEXT: [[TMP15:%.*]] = extractvalue { , , , , , } [[TMP13]], 1 ; RV64-NEXT: [[TMP16:%.*]] = extractvalue { , , , , , } [[TMP13]], 2 ; RV64-NEXT: [[TMP17:%.*]] = extractvalue { , , , , , } [[TMP13]], 3 ; RV64-NEXT: [[TMP18:%.*]] = extractvalue { , , , , , } [[TMP13]], 4 ; RV64-NEXT: [[TMP19:%.*]] = extractvalue { , , , , , } [[TMP13]], 5 ; RV64-NEXT: ret void ; %interleaved.vec = load , ptr %ptr %v = call { , , , , , } @llvm.vector.deinterleave6.nxv12i32( %interleaved.vec) %t0 = extractvalue { , , , , , } %v, 0 %t1 = extractvalue { , , , , , } %v, 1 %t2 = extractvalue { , , , , , } %v, 2 %t3 = extractvalue { , , , , , } %v, 3 %t4 = extractvalue { , , , , , } %v, 4 %t5 = extractvalue { , , , , , } %v, 5 ret void } define void @load_factor7(ptr %ptr) { ; RV32-LABEL: @load_factor7( ; RV32-NEXT: [[TMP1:%.*]] = call { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } @llvm.riscv.seg7.load.mask.v4i32.p0.i32(ptr [[PTR:%.*]], <4 x i1> splat (i1 true), i32 4) ; RV32-NEXT: [[TMP2:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 6 ; RV32-NEXT: [[TMP3:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 5 ; RV32-NEXT: [[TMP4:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 4 ; RV32-NEXT: [[TMP5:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 3 ; RV32-NEXT: [[TMP6:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 2 ; RV32-NEXT: [[TMP7:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 1 ; RV32-NEXT: [[TMP8:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 0 ; RV32-NEXT: ret void ; ; RV64-LABEL: @load_factor7( ; RV64-NEXT: [[TMP1:%.*]] = call { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } @llvm.riscv.seg7.load.mask.v4i32.p0.i64(ptr [[PTR:%.*]], <4 x i1> splat (i1 true), i64 4) ; RV64-NEXT: [[TMP2:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 6 ; RV64-NEXT: [[TMP3:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 5 ; RV64-NEXT: [[TMP4:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 4 ; RV64-NEXT: [[TMP5:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 3 ; RV64-NEXT: [[TMP6:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 2 ; RV64-NEXT: [[TMP7:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 1 ; RV64-NEXT: [[TMP8:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 0 ; RV64-NEXT: ret void ; %interleaved.vec = load <28 x i32>, ptr %ptr %v0 = shufflevector <28 x i32> %interleaved.vec, <28 x i32> poison, <4 x i32> %v1 = shufflevector <28 x i32> %interleaved.vec, <28 x i32> poison, <4 x i32> %v2 = shufflevector <28 x i32> %interleaved.vec, <28 x i32> poison, <4 x i32> %v3 = shufflevector <28 x i32> %interleaved.vec, <28 x i32> poison, <4 x i32> %v4 = shufflevector <28 x i32> %interleaved.vec, <28 x i32> poison, <4 x i32> %v5 = shufflevector <28 x i32> %interleaved.vec, <28 x i32> poison, <4 x i32> %v6 = shufflevector <28 x i32> %interleaved.vec, <28 x i32> poison, <4 x i32> ret void } define void @load_factor7_vscale(ptr %ptr) { ; RV32-LABEL: @load_factor7_vscale( ; RV32-NEXT: [[TMP1:%.*]] = call target("riscv.vector.tuple", , 7) @llvm.riscv.vlseg7.mask.triscv.vector.tuple_nxv8i8_7t.p0.nxv2i1.i32(target("riscv.vector.tuple", , 7) poison, ptr [[PTR:%.*]], splat (i1 true), i32 -1, i32 3, i32 5) ; RV32-NEXT: [[TMP2:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_7t(target("riscv.vector.tuple", , 7) [[TMP1]], i32 0) ; RV32-NEXT: [[TMP3:%.*]] = insertvalue { , , , , , , } poison, [[TMP2]], 0 ; RV32-NEXT: [[TMP4:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_7t(target("riscv.vector.tuple", , 7) [[TMP1]], i32 1) ; RV32-NEXT: [[TMP5:%.*]] = insertvalue { , , , , , , } [[TMP3]], [[TMP4]], 1 ; RV32-NEXT: [[TMP6:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_7t(target("riscv.vector.tuple", , 7) [[TMP1]], i32 2) ; RV32-NEXT: [[TMP7:%.*]] = insertvalue { , , , , , , } [[TMP5]], [[TMP6]], 2 ; RV32-NEXT: [[TMP8:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_7t(target("riscv.vector.tuple", , 7) [[TMP1]], i32 3) ; RV32-NEXT: [[TMP9:%.*]] = insertvalue { , , , , , , } [[TMP7]], [[TMP8]], 3 ; RV32-NEXT: [[TMP10:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_7t(target("riscv.vector.tuple", , 7) [[TMP1]], i32 4) ; RV32-NEXT: [[TMP11:%.*]] = insertvalue { , , , , , , } [[TMP9]], [[TMP10]], 4 ; RV32-NEXT: [[TMP12:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_7t(target("riscv.vector.tuple", , 7) [[TMP1]], i32 5) ; RV32-NEXT: [[TMP13:%.*]] = insertvalue { , , , , , , } [[TMP11]], [[TMP12]], 5 ; RV32-NEXT: [[TMP14:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_7t(target("riscv.vector.tuple", , 7) [[TMP1]], i32 6) ; RV32-NEXT: [[V:%.*]] = insertvalue { , , , , , , } [[TMP13]], [[TMP14]], 6 ; RV32-NEXT: [[T0:%.*]] = extractvalue { , , , , , , } [[V]], 0 ; RV32-NEXT: [[T1:%.*]] = extractvalue { , , , , , , } [[V]], 1 ; RV32-NEXT: [[T2:%.*]] = extractvalue { , , , , , , } [[V]], 2 ; RV32-NEXT: [[T3:%.*]] = extractvalue { , , , , , , } [[V]], 3 ; RV32-NEXT: [[T4:%.*]] = extractvalue { , , , , , , } [[V]], 4 ; RV32-NEXT: [[T5:%.*]] = extractvalue { , , , , , , } [[V]], 5 ; RV32-NEXT: [[T6:%.*]] = extractvalue { , , , , , , } [[V]], 6 ; RV32-NEXT: ret void ; ; RV64-LABEL: @load_factor7_vscale( ; RV64-NEXT: [[TMP1:%.*]] = call target("riscv.vector.tuple", , 7) @llvm.riscv.vlseg7.mask.triscv.vector.tuple_nxv8i8_7t.p0.nxv2i1.i64(target("riscv.vector.tuple", , 7) poison, ptr [[PTR:%.*]], splat (i1 true), i64 -1, i64 3, i64 5) ; RV64-NEXT: [[TMP2:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_7t(target("riscv.vector.tuple", , 7) [[TMP1]], i32 0) ; RV64-NEXT: [[TMP3:%.*]] = insertvalue { , , , , , , } poison, [[TMP2]], 0 ; RV64-NEXT: [[TMP4:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_7t(target("riscv.vector.tuple", , 7) [[TMP1]], i32 1) ; RV64-NEXT: [[TMP5:%.*]] = insertvalue { , , , , , , } [[TMP3]], [[TMP4]], 1 ; RV64-NEXT: [[TMP6:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_7t(target("riscv.vector.tuple", , 7) [[TMP1]], i32 2) ; RV64-NEXT: [[TMP7:%.*]] = insertvalue { , , , , , , } [[TMP5]], [[TMP6]], 2 ; RV64-NEXT: [[TMP8:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_7t(target("riscv.vector.tuple", , 7) [[TMP1]], i32 3) ; RV64-NEXT: [[TMP9:%.*]] = insertvalue { , , , , , , } [[TMP7]], [[TMP8]], 3 ; RV64-NEXT: [[TMP10:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_7t(target("riscv.vector.tuple", , 7) [[TMP1]], i32 4) ; RV64-NEXT: [[TMP11:%.*]] = insertvalue { , , , , , , } [[TMP9]], [[TMP10]], 4 ; RV64-NEXT: [[TMP12:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_7t(target("riscv.vector.tuple", , 7) [[TMP1]], i32 5) ; RV64-NEXT: [[TMP13:%.*]] = insertvalue { , , , , , , } [[TMP11]], [[TMP12]], 5 ; RV64-NEXT: [[TMP14:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_7t(target("riscv.vector.tuple", , 7) [[TMP1]], i32 6) ; RV64-NEXT: [[V:%.*]] = insertvalue { , , , , , , } [[TMP13]], [[TMP14]], 6 ; RV64-NEXT: [[T0:%.*]] = extractvalue { , , , , , , } [[V]], 0 ; RV64-NEXT: [[T1:%.*]] = extractvalue { , , , , , , } [[V]], 1 ; RV64-NEXT: [[T2:%.*]] = extractvalue { , , , , , , } [[V]], 2 ; RV64-NEXT: [[T3:%.*]] = extractvalue { , , , , , , } [[V]], 3 ; RV64-NEXT: [[T4:%.*]] = extractvalue { , , , , , , } [[V]], 4 ; RV64-NEXT: [[T5:%.*]] = extractvalue { , , , , , , } [[V]], 5 ; RV64-NEXT: [[T6:%.*]] = extractvalue { , , , , , , } [[V]], 6 ; RV64-NEXT: ret void ; %interleaved.vec = load , ptr %ptr %v = call { , , , , , , } @llvm.vector.deinterleave7.nxv14i32( %interleaved.vec) %t0 = extractvalue { , , , , , , } %v, 0 %t1 = extractvalue { , , , , , , } %v, 1 %t2 = extractvalue { , , , , , , } %v, 2 %t3 = extractvalue { , , , , , , } %v, 3 %t4 = extractvalue { , , , , , , } %v, 4 %t5 = extractvalue { , , , , , , } %v, 5 %t6 = extractvalue { , , , , , , } %v, 6 ret void } define void @load_factor8(ptr %ptr) { ; RV32-LABEL: @load_factor8( ; RV32-NEXT: [[TMP1:%.*]] = call { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } @llvm.riscv.seg8.load.mask.v4i32.p0.i32(ptr [[PTR:%.*]], <4 x i1> splat (i1 true), i32 4) ; RV32-NEXT: [[TMP2:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 7 ; RV32-NEXT: [[TMP3:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 6 ; RV32-NEXT: [[TMP4:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 5 ; RV32-NEXT: [[TMP5:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 4 ; RV32-NEXT: [[TMP6:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 3 ; RV32-NEXT: [[TMP7:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 2 ; RV32-NEXT: [[TMP8:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 1 ; RV32-NEXT: [[TMP9:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 0 ; RV32-NEXT: ret void ; ; RV64-LABEL: @load_factor8( ; RV64-NEXT: [[TMP1:%.*]] = call { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } @llvm.riscv.seg8.load.mask.v4i32.p0.i64(ptr [[PTR:%.*]], <4 x i1> splat (i1 true), i64 4) ; RV64-NEXT: [[TMP2:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 7 ; RV64-NEXT: [[TMP3:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 6 ; RV64-NEXT: [[TMP4:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 5 ; RV64-NEXT: [[TMP5:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 4 ; RV64-NEXT: [[TMP6:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 3 ; RV64-NEXT: [[TMP7:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 2 ; RV64-NEXT: [[TMP8:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 1 ; RV64-NEXT: [[TMP9:%.*]] = extractvalue { <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32>, <4 x i32> } [[TMP1]], 0 ; RV64-NEXT: ret void ; %interleaved.vec = load <32 x i32>, ptr %ptr %v0 = shufflevector <32 x i32> %interleaved.vec, <32 x i32> poison, <4 x i32> %v1 = shufflevector <32 x i32> %interleaved.vec, <32 x i32> poison, <4 x i32> %v2 = shufflevector <32 x i32> %interleaved.vec, <32 x i32> poison, <4 x i32> %v3 = shufflevector <32 x i32> %interleaved.vec, <32 x i32> poison, <4 x i32> %v4 = shufflevector <32 x i32> %interleaved.vec, <32 x i32> poison, <4 x i32> %v5 = shufflevector <32 x i32> %interleaved.vec, <32 x i32> poison, <4 x i32> %v6 = shufflevector <32 x i32> %interleaved.vec, <32 x i32> poison, <4 x i32> %v7 = shufflevector <32 x i32> %interleaved.vec, <32 x i32> poison, <4 x i32> ret void } define void @load_factor8_vscale(ptr %ptr) { ; RV32-LABEL: @load_factor8_vscale( ; RV32-NEXT: [[TMP1:%.*]] = call target("riscv.vector.tuple", , 8) @llvm.riscv.vlseg8.mask.triscv.vector.tuple_nxv8i8_8t.p0.nxv2i1.i32(target("riscv.vector.tuple", , 8) poison, ptr [[PTR:%.*]], splat (i1 true), i32 -1, i32 3, i32 5) ; RV32-NEXT: [[TMP2:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_8t(target("riscv.vector.tuple", , 8) [[TMP1]], i32 0) ; RV32-NEXT: [[TMP3:%.*]] = insertvalue { , , , , , , , } poison, [[TMP2]], 0 ; RV32-NEXT: [[TMP4:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_8t(target("riscv.vector.tuple", , 8) [[TMP1]], i32 1) ; RV32-NEXT: [[TMP5:%.*]] = insertvalue { , , , , , , , } [[TMP3]], [[TMP4]], 1 ; RV32-NEXT: [[TMP6:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_8t(target("riscv.vector.tuple", , 8) [[TMP1]], i32 2) ; RV32-NEXT: [[TMP7:%.*]] = insertvalue { , , , , , , , } [[TMP5]], [[TMP6]], 2 ; RV32-NEXT: [[TMP8:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_8t(target("riscv.vector.tuple", , 8) [[TMP1]], i32 3) ; RV32-NEXT: [[TMP9:%.*]] = insertvalue { , , , , , , , } [[TMP7]], [[TMP8]], 3 ; RV32-NEXT: [[TMP10:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_8t(target("riscv.vector.tuple", , 8) [[TMP1]], i32 4) ; RV32-NEXT: [[TMP11:%.*]] = insertvalue { , , , , , , , } [[TMP9]], [[TMP10]], 4 ; RV32-NEXT: [[TMP12:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_8t(target("riscv.vector.tuple", , 8) [[TMP1]], i32 5) ; RV32-NEXT: [[TMP13:%.*]] = insertvalue { , , , , , , , } [[TMP11]], [[TMP12]], 5 ; RV32-NEXT: [[TMP14:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_8t(target("riscv.vector.tuple", , 8) [[TMP1]], i32 6) ; RV32-NEXT: [[TMP15:%.*]] = insertvalue { , , , , , , , } [[TMP13]], [[TMP14]], 6 ; RV32-NEXT: [[TMP16:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_8t(target("riscv.vector.tuple", , 8) [[TMP1]], i32 7) ; RV32-NEXT: [[TMP17:%.*]] = insertvalue { , , , , , , , } [[TMP15]], [[TMP16]], 7 ; RV32-NEXT: [[TMP18:%.*]] = extractvalue { , , , , , , , } [[TMP17]], 0 ; RV32-NEXT: [[TMP19:%.*]] = extractvalue { , , , , , , , } [[TMP17]], 1 ; RV32-NEXT: [[TMP20:%.*]] = extractvalue { , , , , , , , } [[TMP17]], 2 ; RV32-NEXT: [[TMP21:%.*]] = extractvalue { , , , , , , , } [[TMP17]], 3 ; RV32-NEXT: [[TMP22:%.*]] = extractvalue { , , , , , , , } [[TMP17]], 4 ; RV32-NEXT: [[TMP23:%.*]] = extractvalue { , , , , , , , } [[TMP17]], 5 ; RV32-NEXT: [[TMP24:%.*]] = extractvalue { , , , , , , , } [[TMP17]], 6 ; RV32-NEXT: [[TMP25:%.*]] = extractvalue { , , , , , , , } [[TMP17]], 7 ; RV32-NEXT: ret void ; ; RV64-LABEL: @load_factor8_vscale( ; RV64-NEXT: [[TMP1:%.*]] = call target("riscv.vector.tuple", , 8) @llvm.riscv.vlseg8.mask.triscv.vector.tuple_nxv8i8_8t.p0.nxv2i1.i64(target("riscv.vector.tuple", , 8) poison, ptr [[PTR:%.*]], splat (i1 true), i64 -1, i64 3, i64 5) ; RV64-NEXT: [[TMP2:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_8t(target("riscv.vector.tuple", , 8) [[TMP1]], i32 0) ; RV64-NEXT: [[TMP3:%.*]] = insertvalue { , , , , , , , } poison, [[TMP2]], 0 ; RV64-NEXT: [[TMP4:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_8t(target("riscv.vector.tuple", , 8) [[TMP1]], i32 1) ; RV64-NEXT: [[TMP5:%.*]] = insertvalue { , , , , , , , } [[TMP3]], [[TMP4]], 1 ; RV64-NEXT: [[TMP6:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_8t(target("riscv.vector.tuple", , 8) [[TMP1]], i32 2) ; RV64-NEXT: [[TMP7:%.*]] = insertvalue { , , , , , , , } [[TMP5]], [[TMP6]], 2 ; RV64-NEXT: [[TMP8:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_8t(target("riscv.vector.tuple", , 8) [[TMP1]], i32 3) ; RV64-NEXT: [[TMP9:%.*]] = insertvalue { , , , , , , , } [[TMP7]], [[TMP8]], 3 ; RV64-NEXT: [[TMP10:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_8t(target("riscv.vector.tuple", , 8) [[TMP1]], i32 4) ; RV64-NEXT: [[TMP11:%.*]] = insertvalue { , , , , , , , } [[TMP9]], [[TMP10]], 4 ; RV64-NEXT: [[TMP12:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_8t(target("riscv.vector.tuple", , 8) [[TMP1]], i32 5) ; RV64-NEXT: [[TMP13:%.*]] = insertvalue { , , , , , , , } [[TMP11]], [[TMP12]], 5 ; RV64-NEXT: [[TMP14:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_8t(target("riscv.vector.tuple", , 8) [[TMP1]], i32 6) ; RV64-NEXT: [[TMP15:%.*]] = insertvalue { , , , , , , , } [[TMP13]], [[TMP14]], 6 ; RV64-NEXT: [[TMP16:%.*]] = call @llvm.riscv.tuple.extract.nxv2i32.triscv.vector.tuple_nxv8i8_8t(target("riscv.vector.tuple", , 8) [[TMP1]], i32 7) ; RV64-NEXT: [[TMP17:%.*]] = insertvalue { , , , , , , , } [[TMP15]], [[TMP16]], 7 ; RV64-NEXT: [[TMP18:%.*]] = extractvalue { , , , , , , , } [[TMP17]], 0 ; RV64-NEXT: [[TMP19:%.*]] = extractvalue { , , , , , , , } [[TMP17]], 1 ; RV64-NEXT: [[TMP20:%.*]] = extractvalue { , , , , , , , } [[TMP17]], 2 ; RV64-NEXT: [[TMP21:%.*]] = extractvalue { , , , , , , , } [[TMP17]], 3 ; RV64-NEXT: [[TMP22:%.*]] = extractvalue { , , , , , , , } [[TMP17]], 4 ; RV64-NEXT: [[TMP23:%.*]] = extractvalue { , , , , , , , } [[TMP17]], 5 ; RV64-NEXT: [[TMP24:%.*]] = extractvalue { , , , , , , , } [[TMP17]], 6 ; RV64-NEXT: [[TMP25:%.*]] = extractvalue { , , , , , , , } [[TMP17]], 7 ; RV64-NEXT: ret void ; %interleaved.vec = load , ptr %ptr %v = call { , , , , , , , } @llvm.vector.deinterleave8.nxv16i32( %interleaved.vec) %t0 = extractvalue { , , , , , , , } %v, 0 %t1 = extractvalue { , , , , , , , } %v, 1 %t2 = extractvalue { , , , , , , , } %v, 2 %t3 = extractvalue { , , , , , , , } %v, 3 %t4 = extractvalue { , , , , , , , } %v, 4 %t5 = extractvalue { , , , , , , , } %v, 5 %t6 = extractvalue { , , , , , , , } %v, 6 %t7 = extractvalue { , , , , , , , } %v, 7 ret void } define void @store_factor2(ptr %ptr, <8 x i8> %v0, <8 x i8> %v1) { ; RV32-LABEL: @store_factor2( ; RV32-NEXT: [[TMP1:%.*]] = shufflevector <8 x i8> [[V0:%.*]], <8 x i8> [[V1:%.*]], <8 x i32> ; RV32-NEXT: [[TMP2:%.*]] = shufflevector <8 x i8> [[V0]], <8 x i8> [[V1]], <8 x i32> ; RV32-NEXT: call void @llvm.riscv.seg2.store.mask.v8i8.p0.i32(<8 x i8> [[TMP1]], <8 x i8> [[TMP2]], ptr [[PTR:%.*]], <8 x i1> splat (i1 true), i32 8) ; RV32-NEXT: ret void ; ; RV64-LABEL: @store_factor2( ; RV64-NEXT: [[TMP1:%.*]] = shufflevector <8 x i8> [[V0:%.*]], <8 x i8> [[V1:%.*]], <8 x i32> ; RV64-NEXT: [[TMP2:%.*]] = shufflevector <8 x i8> [[V0]], <8 x i8> [[V1]], <8 x i32> ; RV64-NEXT: call void @llvm.riscv.seg2.store.mask.v8i8.p0.i64(<8 x i8> [[TMP1]], <8 x i8> [[TMP2]], ptr [[PTR:%.*]], <8 x i1> splat (i1 true), i64 8) ; RV64-NEXT: ret void ; %interleaved.vec = shufflevector <8 x i8> %v0, <8 x i8> %v1, <16 x i32> store <16 x i8> %interleaved.vec, ptr %ptr, align 4 ret void } define void @store_factor2_vscale(ptr %ptr, %v0, %v1) { ; RV32-LABEL: @store_factor2_vscale( ; RV32-NEXT: [[TMP1:%.*]] = call target("riscv.vector.tuple", , 2) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_2t.nxv8i8(target("riscv.vector.tuple", , 2) poison, [[V0:%.*]], i32 0) ; RV32-NEXT: [[TMP2:%.*]] = call target("riscv.vector.tuple", , 2) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_2t.nxv8i8(target("riscv.vector.tuple", , 2) [[TMP1]], [[V1:%.*]], i32 1) ; RV32-NEXT: call void @llvm.riscv.vsseg2.mask.triscv.vector.tuple_nxv8i8_2t.p0.nxv8i1.i32(target("riscv.vector.tuple", , 2) [[TMP2]], ptr [[PTR:%.*]], splat (i1 true), i32 -1, i32 3) ; RV32-NEXT: ret void ; ; RV64-LABEL: @store_factor2_vscale( ; RV64-NEXT: [[TMP1:%.*]] = call target("riscv.vector.tuple", , 2) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_2t.nxv8i8(target("riscv.vector.tuple", , 2) poison, [[V0:%.*]], i32 0) ; RV64-NEXT: [[TMP2:%.*]] = call target("riscv.vector.tuple", , 2) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_2t.nxv8i8(target("riscv.vector.tuple", , 2) [[TMP1]], [[V1:%.*]], i32 1) ; RV64-NEXT: call void @llvm.riscv.vsseg2.mask.triscv.vector.tuple_nxv8i8_2t.p0.nxv8i1.i64(target("riscv.vector.tuple", , 2) [[TMP2]], ptr [[PTR:%.*]], splat (i1 true), i64 -1, i64 3) ; RV64-NEXT: ret void ; %interleaved.vec = call @llvm.vector.interleave2.nxv8i8( %v0, %v1) store %interleaved.vec, ptr %ptr, align 4 ret void } define void @store_factor3(ptr %ptr, <4 x i32> %v0, <4 x i32> %v1, <4 x i32> %v2) { ; RV32-LABEL: @store_factor3( ; RV32-NEXT: [[S0:%.*]] = shufflevector <4 x i32> [[V0:%.*]], <4 x i32> [[V1:%.*]], <8 x i32> ; RV32-NEXT: [[S1:%.*]] = shufflevector <4 x i32> [[V2:%.*]], <4 x i32> poison, <8 x i32> ; RV32-NEXT: [[TMP1:%.*]] = shufflevector <8 x i32> [[S0]], <8 x i32> [[S1]], <4 x i32> ; RV32-NEXT: [[TMP2:%.*]] = shufflevector <8 x i32> [[S0]], <8 x i32> [[S1]], <4 x i32> ; RV32-NEXT: [[TMP3:%.*]] = shufflevector <8 x i32> [[S0]], <8 x i32> [[S1]], <4 x i32> ; RV32-NEXT: call void @llvm.riscv.seg3.store.mask.v4i32.p0.i32(<4 x i32> [[TMP1]], <4 x i32> [[TMP2]], <4 x i32> [[TMP3]], ptr [[PTR:%.*]], <4 x i1> splat (i1 true), i32 4) ; RV32-NEXT: ret void ; ; RV64-LABEL: @store_factor3( ; RV64-NEXT: [[S0:%.*]] = shufflevector <4 x i32> [[V0:%.*]], <4 x i32> [[V1:%.*]], <8 x i32> ; RV64-NEXT: [[S1:%.*]] = shufflevector <4 x i32> [[V2:%.*]], <4 x i32> poison, <8 x i32> ; RV64-NEXT: [[TMP1:%.*]] = shufflevector <8 x i32> [[S0]], <8 x i32> [[S1]], <4 x i32> ; RV64-NEXT: [[TMP2:%.*]] = shufflevector <8 x i32> [[S0]], <8 x i32> [[S1]], <4 x i32> ; RV64-NEXT: [[TMP3:%.*]] = shufflevector <8 x i32> [[S0]], <8 x i32> [[S1]], <4 x i32> ; RV64-NEXT: call void @llvm.riscv.seg3.store.mask.v4i32.p0.i64(<4 x i32> [[TMP1]], <4 x i32> [[TMP2]], <4 x i32> [[TMP3]], ptr [[PTR:%.*]], <4 x i1> splat (i1 true), i64 4) ; RV64-NEXT: ret void ; %s0 = shufflevector <4 x i32> %v0, <4 x i32> %v1, <8 x i32> %s1 = shufflevector <4 x i32> %v2, <4 x i32> poison, <8 x i32> %interleaved.vec = shufflevector <8 x i32> %s0, <8 x i32> %s1, <12 x i32> store <12 x i32> %interleaved.vec, ptr %ptr, align 4 ret void } define void @store_factor3_vscale(ptr %ptr, %v0, %v1, %v2) { ; RV32-LABEL: @store_factor3_vscale( ; RV32-NEXT: [[TMP1:%.*]] = call target("riscv.vector.tuple", , 3) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_3t.nxv8i8(target("riscv.vector.tuple", , 3) poison, [[V0:%.*]], i32 0) ; RV32-NEXT: [[TMP2:%.*]] = call target("riscv.vector.tuple", , 3) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_3t.nxv8i8(target("riscv.vector.tuple", , 3) [[TMP1]], [[V1:%.*]], i32 1) ; RV32-NEXT: [[TMP3:%.*]] = call target("riscv.vector.tuple", , 3) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_3t.nxv8i8(target("riscv.vector.tuple", , 3) [[TMP2]], [[V2:%.*]], i32 2) ; RV32-NEXT: call void @llvm.riscv.vsseg3.mask.triscv.vector.tuple_nxv8i8_3t.p0.nxv8i1.i32(target("riscv.vector.tuple", , 3) [[TMP3]], ptr [[PTR:%.*]], splat (i1 true), i32 -1, i32 3) ; RV32-NEXT: ret void ; ; RV64-LABEL: @store_factor3_vscale( ; RV64-NEXT: [[TMP1:%.*]] = call target("riscv.vector.tuple", , 3) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_3t.nxv8i8(target("riscv.vector.tuple", , 3) poison, [[V0:%.*]], i32 0) ; RV64-NEXT: [[TMP2:%.*]] = call target("riscv.vector.tuple", , 3) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_3t.nxv8i8(target("riscv.vector.tuple", , 3) [[TMP1]], [[V1:%.*]], i32 1) ; RV64-NEXT: [[TMP3:%.*]] = call target("riscv.vector.tuple", , 3) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_3t.nxv8i8(target("riscv.vector.tuple", , 3) [[TMP2]], [[V2:%.*]], i32 2) ; RV64-NEXT: call void @llvm.riscv.vsseg3.mask.triscv.vector.tuple_nxv8i8_3t.p0.nxv8i1.i64(target("riscv.vector.tuple", , 3) [[TMP3]], ptr [[PTR:%.*]], splat (i1 true), i64 -1, i64 3) ; RV64-NEXT: ret void ; %interleaved.vec = call @llvm.vector.interleave3.nxv8i8( %v0, %v1, %v2) store %interleaved.vec, ptr %ptr ret void } define void @store_factor4(ptr %ptr, <4 x i32> %v0, <4 x i32> %v1, <4 x i32> %v2, <4 x i32> %v3) { ; RV32-LABEL: @store_factor4( ; RV32-NEXT: [[S0:%.*]] = shufflevector <4 x i32> [[V0:%.*]], <4 x i32> [[V1:%.*]], <8 x i32> ; RV32-NEXT: [[S1:%.*]] = shufflevector <4 x i32> [[V2:%.*]], <4 x i32> [[V3:%.*]], <8 x i32> ; RV32-NEXT: [[TMP1:%.*]] = shufflevector <8 x i32> [[S0]], <8 x i32> [[S1]], <4 x i32> ; RV32-NEXT: [[TMP2:%.*]] = shufflevector <8 x i32> [[S0]], <8 x i32> [[S1]], <4 x i32> ; RV32-NEXT: [[TMP3:%.*]] = shufflevector <8 x i32> [[S0]], <8 x i32> [[S1]], <4 x i32> ; RV32-NEXT: [[TMP4:%.*]] = shufflevector <8 x i32> [[S0]], <8 x i32> [[S1]], <4 x i32> ; RV32-NEXT: call void @llvm.riscv.seg4.store.mask.v4i32.p0.i32(<4 x i32> [[TMP1]], <4 x i32> [[TMP2]], <4 x i32> [[TMP3]], <4 x i32> [[TMP4]], ptr [[PTR:%.*]], <4 x i1> splat (i1 true), i32 4) ; RV32-NEXT: ret void ; ; RV64-LABEL: @store_factor4( ; RV64-NEXT: [[S0:%.*]] = shufflevector <4 x i32> [[V0:%.*]], <4 x i32> [[V1:%.*]], <8 x i32> ; RV64-NEXT: [[S1:%.*]] = shufflevector <4 x i32> [[V2:%.*]], <4 x i32> [[V3:%.*]], <8 x i32> ; RV64-NEXT: [[TMP1:%.*]] = shufflevector <8 x i32> [[S0]], <8 x i32> [[S1]], <4 x i32> ; RV64-NEXT: [[TMP2:%.*]] = shufflevector <8 x i32> [[S0]], <8 x i32> [[S1]], <4 x i32> ; RV64-NEXT: [[TMP3:%.*]] = shufflevector <8 x i32> [[S0]], <8 x i32> [[S1]], <4 x i32> ; RV64-NEXT: [[TMP4:%.*]] = shufflevector <8 x i32> [[S0]], <8 x i32> [[S1]], <4 x i32> ; RV64-NEXT: call void @llvm.riscv.seg4.store.mask.v4i32.p0.i64(<4 x i32> [[TMP1]], <4 x i32> [[TMP2]], <4 x i32> [[TMP3]], <4 x i32> [[TMP4]], ptr [[PTR:%.*]], <4 x i1> splat (i1 true), i64 4) ; RV64-NEXT: ret void ; %s0 = shufflevector <4 x i32> %v0, <4 x i32> %v1, <8 x i32> %s1 = shufflevector <4 x i32> %v2, <4 x i32> %v3, <8 x i32> %interleaved.vec = shufflevector <8 x i32> %s0, <8 x i32> %s1, <16 x i32> store <16 x i32> %interleaved.vec, ptr %ptr, align 4 ret void } define void @store_factor4_vscale(ptr %ptr, %v0, %v1, %v2, %v3) { ; RV32-LABEL: @store_factor4_vscale( ; RV32-NEXT: [[TMP1:%.*]] = call target("riscv.vector.tuple", , 4) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_4t.nxv8i8(target("riscv.vector.tuple", , 4) poison, [[V0:%.*]], i32 0) ; RV32-NEXT: [[TMP2:%.*]] = call target("riscv.vector.tuple", , 4) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_4t.nxv8i8(target("riscv.vector.tuple", , 4) [[TMP1]], [[V1:%.*]], i32 1) ; RV32-NEXT: [[TMP3:%.*]] = call target("riscv.vector.tuple", , 4) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_4t.nxv8i8(target("riscv.vector.tuple", , 4) [[TMP2]], [[V2:%.*]], i32 2) ; RV32-NEXT: [[TMP4:%.*]] = call target("riscv.vector.tuple", , 4) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_4t.nxv8i8(target("riscv.vector.tuple", , 4) [[TMP3]], [[V3:%.*]], i32 3) ; RV32-NEXT: call void @llvm.riscv.vsseg4.mask.triscv.vector.tuple_nxv8i8_4t.p0.nxv8i1.i32(target("riscv.vector.tuple", , 4) [[TMP4]], ptr [[PTR:%.*]], splat (i1 true), i32 -1, i32 3) ; RV32-NEXT: ret void ; ; RV64-LABEL: @store_factor4_vscale( ; RV64-NEXT: [[TMP1:%.*]] = call target("riscv.vector.tuple", , 4) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_4t.nxv8i8(target("riscv.vector.tuple", , 4) poison, [[V0:%.*]], i32 0) ; RV64-NEXT: [[TMP2:%.*]] = call target("riscv.vector.tuple", , 4) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_4t.nxv8i8(target("riscv.vector.tuple", , 4) [[TMP1]], [[V1:%.*]], i32 1) ; RV64-NEXT: [[TMP3:%.*]] = call target("riscv.vector.tuple", , 4) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_4t.nxv8i8(target("riscv.vector.tuple", , 4) [[TMP2]], [[V2:%.*]], i32 2) ; RV64-NEXT: [[TMP4:%.*]] = call target("riscv.vector.tuple", , 4) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_4t.nxv8i8(target("riscv.vector.tuple", , 4) [[TMP3]], [[V3:%.*]], i32 3) ; RV64-NEXT: call void @llvm.riscv.vsseg4.mask.triscv.vector.tuple_nxv8i8_4t.p0.nxv8i1.i64(target("riscv.vector.tuple", , 4) [[TMP4]], ptr [[PTR:%.*]], splat (i1 true), i64 -1, i64 3) ; RV64-NEXT: ret void ; %interleaved.vec = call @llvm.vector.interleave4.nxv8i8( %v0, %v1, %v2, %v3) store %interleaved.vec, ptr %ptr ret void } define void @store_factor5_vscale(ptr %ptr, %v0, %v1, %v2, %v3, %v4) { ; RV32-LABEL: @store_factor5_vscale( ; RV32-NEXT: [[TMP1:%.*]] = call target("riscv.vector.tuple", , 5) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_5t.nxv8i8(target("riscv.vector.tuple", , 5) poison, [[V0:%.*]], i32 0) ; RV32-NEXT: [[TMP2:%.*]] = call target("riscv.vector.tuple", , 5) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_5t.nxv8i8(target("riscv.vector.tuple", , 5) [[TMP1]], [[V1:%.*]], i32 1) ; RV32-NEXT: [[TMP3:%.*]] = call target("riscv.vector.tuple", , 5) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_5t.nxv8i8(target("riscv.vector.tuple", , 5) [[TMP2]], [[V2:%.*]], i32 2) ; RV32-NEXT: [[TMP4:%.*]] = call target("riscv.vector.tuple", , 5) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_5t.nxv8i8(target("riscv.vector.tuple", , 5) [[TMP3]], [[V3:%.*]], i32 3) ; RV32-NEXT: [[TMP5:%.*]] = call target("riscv.vector.tuple", , 5) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_5t.nxv8i8(target("riscv.vector.tuple", , 5) [[TMP4]], [[V4:%.*]], i32 4) ; RV32-NEXT: call void @llvm.riscv.vsseg5.mask.triscv.vector.tuple_nxv8i8_5t.p0.nxv8i1.i32(target("riscv.vector.tuple", , 5) [[TMP5]], ptr [[PTR:%.*]], splat (i1 true), i32 -1, i32 3) ; RV32-NEXT: ret void ; ; RV64-LABEL: @store_factor5_vscale( ; RV64-NEXT: [[TMP1:%.*]] = call target("riscv.vector.tuple", , 5) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_5t.nxv8i8(target("riscv.vector.tuple", , 5) poison, [[V0:%.*]], i32 0) ; RV64-NEXT: [[TMP2:%.*]] = call target("riscv.vector.tuple", , 5) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_5t.nxv8i8(target("riscv.vector.tuple", , 5) [[TMP1]], [[V1:%.*]], i32 1) ; RV64-NEXT: [[TMP3:%.*]] = call target("riscv.vector.tuple", , 5) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_5t.nxv8i8(target("riscv.vector.tuple", , 5) [[TMP2]], [[V2:%.*]], i32 2) ; RV64-NEXT: [[TMP4:%.*]] = call target("riscv.vector.tuple", , 5) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_5t.nxv8i8(target("riscv.vector.tuple", , 5) [[TMP3]], [[V3:%.*]], i32 3) ; RV64-NEXT: [[TMP5:%.*]] = call target("riscv.vector.tuple", , 5) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_5t.nxv8i8(target("riscv.vector.tuple", , 5) [[TMP4]], [[V4:%.*]], i32 4) ; RV64-NEXT: call void @llvm.riscv.vsseg5.mask.triscv.vector.tuple_nxv8i8_5t.p0.nxv8i1.i64(target("riscv.vector.tuple", , 5) [[TMP5]], ptr [[PTR:%.*]], splat (i1 true), i64 -1, i64 3) ; RV64-NEXT: ret void ; %interleaved.vec = call @llvm.vector.interleave5.nxv8i8( %v0, %v1, %v2, %v3, %v4) store %interleaved.vec, ptr %ptr ret void } define void @store_factor2_wide(ptr %ptr, <8 x i32> %v0, <8 x i32> %v1) { ; RV32-LABEL: @store_factor2_wide( ; RV32-NEXT: [[TMP1:%.*]] = shufflevector <8 x i32> [[V0:%.*]], <8 x i32> [[V1:%.*]], <8 x i32> ; RV32-NEXT: [[TMP2:%.*]] = shufflevector <8 x i32> [[V0]], <8 x i32> [[V1]], <8 x i32> ; RV32-NEXT: call void @llvm.riscv.seg2.store.mask.v8i32.p0.i32(<8 x i32> [[TMP1]], <8 x i32> [[TMP2]], ptr [[PTR:%.*]], <8 x i1> splat (i1 true), i32 8) ; RV32-NEXT: ret void ; ; RV64-LABEL: @store_factor2_wide( ; RV64-NEXT: [[TMP1:%.*]] = shufflevector <8 x i32> [[V0:%.*]], <8 x i32> [[V1:%.*]], <8 x i32> ; RV64-NEXT: [[TMP2:%.*]] = shufflevector <8 x i32> [[V0]], <8 x i32> [[V1]], <8 x i32> ; RV64-NEXT: call void @llvm.riscv.seg2.store.mask.v8i32.p0.i64(<8 x i32> [[TMP1]], <8 x i32> [[TMP2]], ptr [[PTR:%.*]], <8 x i1> splat (i1 true), i64 8) ; RV64-NEXT: ret void ; %interleaved.vec = shufflevector <8 x i32> %v0, <8 x i32> %v1, <16 x i32> store <16 x i32> %interleaved.vec, ptr %ptr, align 4 ret void } define void @store_factor3_wide(ptr %ptr, <8 x i32> %v0, <8 x i32> %v1, <8 x i32> %v2) { ; RV32-LABEL: @store_factor3_wide( ; RV32-NEXT: [[S0:%.*]] = shufflevector <8 x i32> [[V0:%.*]], <8 x i32> [[V1:%.*]], <16 x i32> ; RV32-NEXT: [[S1:%.*]] = shufflevector <8 x i32> [[V2:%.*]], <8 x i32> poison, <16 x i32> ; RV32-NEXT: [[TMP1:%.*]] = shufflevector <16 x i32> [[S0]], <16 x i32> [[S1]], <8 x i32> ; RV32-NEXT: [[TMP2:%.*]] = shufflevector <16 x i32> [[S0]], <16 x i32> [[S1]], <8 x i32> ; RV32-NEXT: [[TMP3:%.*]] = shufflevector <16 x i32> [[S0]], <16 x i32> [[S1]], <8 x i32> ; RV32-NEXT: call void @llvm.riscv.seg3.store.mask.v8i32.p0.i32(<8 x i32> [[TMP1]], <8 x i32> [[TMP2]], <8 x i32> [[TMP3]], ptr [[PTR:%.*]], <8 x i1> splat (i1 true), i32 8) ; RV32-NEXT: ret void ; ; RV64-LABEL: @store_factor3_wide( ; RV64-NEXT: [[S0:%.*]] = shufflevector <8 x i32> [[V0:%.*]], <8 x i32> [[V1:%.*]], <16 x i32> ; RV64-NEXT: [[S1:%.*]] = shufflevector <8 x i32> [[V2:%.*]], <8 x i32> poison, <16 x i32> ; RV64-NEXT: [[TMP1:%.*]] = shufflevector <16 x i32> [[S0]], <16 x i32> [[S1]], <8 x i32> ; RV64-NEXT: [[TMP2:%.*]] = shufflevector <16 x i32> [[S0]], <16 x i32> [[S1]], <8 x i32> ; RV64-NEXT: [[TMP3:%.*]] = shufflevector <16 x i32> [[S0]], <16 x i32> [[S1]], <8 x i32> ; RV64-NEXT: call void @llvm.riscv.seg3.store.mask.v8i32.p0.i64(<8 x i32> [[TMP1]], <8 x i32> [[TMP2]], <8 x i32> [[TMP3]], ptr [[PTR:%.*]], <8 x i1> splat (i1 true), i64 8) ; RV64-NEXT: ret void ; %s0 = shufflevector <8 x i32> %v0, <8 x i32> %v1, <16 x i32> %s1 = shufflevector <8 x i32> %v2, <8 x i32> poison, <16 x i32> %interleaved.vec = shufflevector <16 x i32> %s0, <16 x i32> %s1, <24 x i32> store <24 x i32> %interleaved.vec, ptr %ptr, align 4 ret void } define void @store_factor4_wide(ptr %ptr, <8 x i32> %v0, <8 x i32> %v1, <8 x i32> %v2, <8 x i32> %v3) { ; RV32-LABEL: @store_factor4_wide( ; RV32-NEXT: [[S0:%.*]] = shufflevector <8 x i32> [[V0:%.*]], <8 x i32> [[V1:%.*]], <16 x i32> ; RV32-NEXT: [[S1:%.*]] = shufflevector <8 x i32> [[V2:%.*]], <8 x i32> [[V3:%.*]], <16 x i32> ; RV32-NEXT: [[TMP1:%.*]] = shufflevector <16 x i32> [[S0]], <16 x i32> [[S1]], <8 x i32> ; RV32-NEXT: [[TMP2:%.*]] = shufflevector <16 x i32> [[S0]], <16 x i32> [[S1]], <8 x i32> ; RV32-NEXT: [[TMP3:%.*]] = shufflevector <16 x i32> [[S0]], <16 x i32> [[S1]], <8 x i32> ; RV32-NEXT: [[TMP4:%.*]] = shufflevector <16 x i32> [[S0]], <16 x i32> [[S1]], <8 x i32> ; RV32-NEXT: call void @llvm.riscv.seg4.store.mask.v8i32.p0.i32(<8 x i32> [[TMP1]], <8 x i32> [[TMP2]], <8 x i32> [[TMP3]], <8 x i32> [[TMP4]], ptr [[PTR:%.*]], <8 x i1> splat (i1 true), i32 8) ; RV32-NEXT: ret void ; ; RV64-LABEL: @store_factor4_wide( ; RV64-NEXT: [[S0:%.*]] = shufflevector <8 x i32> [[V0:%.*]], <8 x i32> [[V1:%.*]], <16 x i32> ; RV64-NEXT: [[S1:%.*]] = shufflevector <8 x i32> [[V2:%.*]], <8 x i32> [[V3:%.*]], <16 x i32> ; RV64-NEXT: [[TMP1:%.*]] = shufflevector <16 x i32> [[S0]], <16 x i32> [[S1]], <8 x i32> ; RV64-NEXT: [[TMP2:%.*]] = shufflevector <16 x i32> [[S0]], <16 x i32> [[S1]], <8 x i32> ; RV64-NEXT: [[TMP3:%.*]] = shufflevector <16 x i32> [[S0]], <16 x i32> [[S1]], <8 x i32> ; RV64-NEXT: [[TMP4:%.*]] = shufflevector <16 x i32> [[S0]], <16 x i32> [[S1]], <8 x i32> ; RV64-NEXT: call void @llvm.riscv.seg4.store.mask.v8i32.p0.i64(<8 x i32> [[TMP1]], <8 x i32> [[TMP2]], <8 x i32> [[TMP3]], <8 x i32> [[TMP4]], ptr [[PTR:%.*]], <8 x i1> splat (i1 true), i64 8) ; RV64-NEXT: ret void ; %s0 = shufflevector <8 x i32> %v0, <8 x i32> %v1, <16 x i32> %s1 = shufflevector <8 x i32> %v2, <8 x i32> %v3, <16 x i32> %interleaved.vec = shufflevector <16 x i32> %s0, <16 x i32> %s1, <32 x i32> store <32 x i32> %interleaved.vec, ptr %ptr, align 4 ret void } define void @store_factor6_vscale(ptr %ptr, %v0, %v1, %v2, %v3, %v4, %v5) { ; RV32-LABEL: @store_factor6_vscale( ; RV32-NEXT: [[TMP1:%.*]] = call target("riscv.vector.tuple", , 6) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_6t.nxv8i8(target("riscv.vector.tuple", , 6) poison, [[V0:%.*]], i32 0) ; RV32-NEXT: [[TMP2:%.*]] = call target("riscv.vector.tuple", , 6) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_6t.nxv8i8(target("riscv.vector.tuple", , 6) [[TMP1]], [[V1:%.*]], i32 1) ; RV32-NEXT: [[TMP3:%.*]] = call target("riscv.vector.tuple", , 6) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_6t.nxv8i8(target("riscv.vector.tuple", , 6) [[TMP2]], [[V2:%.*]], i32 2) ; RV32-NEXT: [[TMP4:%.*]] = call target("riscv.vector.tuple", , 6) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_6t.nxv8i8(target("riscv.vector.tuple", , 6) [[TMP3]], [[V3:%.*]], i32 3) ; RV32-NEXT: [[TMP5:%.*]] = call target("riscv.vector.tuple", , 6) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_6t.nxv8i8(target("riscv.vector.tuple", , 6) [[TMP4]], [[V4:%.*]], i32 4) ; RV32-NEXT: [[TMP6:%.*]] = call target("riscv.vector.tuple", , 6) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_6t.nxv8i8(target("riscv.vector.tuple", , 6) [[TMP5]], [[V5:%.*]], i32 5) ; RV32-NEXT: call void @llvm.riscv.vsseg6.mask.triscv.vector.tuple_nxv8i8_6t.p0.nxv8i1.i32(target("riscv.vector.tuple", , 6) [[TMP6]], ptr [[PTR:%.*]], splat (i1 true), i32 -1, i32 3) ; RV32-NEXT: ret void ; ; RV64-LABEL: @store_factor6_vscale( ; RV64-NEXT: [[TMP1:%.*]] = call target("riscv.vector.tuple", , 6) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_6t.nxv8i8(target("riscv.vector.tuple", , 6) poison, [[V0:%.*]], i32 0) ; RV64-NEXT: [[TMP2:%.*]] = call target("riscv.vector.tuple", , 6) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_6t.nxv8i8(target("riscv.vector.tuple", , 6) [[TMP1]], [[V1:%.*]], i32 1) ; RV64-NEXT: [[TMP3:%.*]] = call target("riscv.vector.tuple", , 6) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_6t.nxv8i8(target("riscv.vector.tuple", , 6) [[TMP2]], [[V2:%.*]], i32 2) ; RV64-NEXT: [[TMP4:%.*]] = call target("riscv.vector.tuple", , 6) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_6t.nxv8i8(target("riscv.vector.tuple", , 6) [[TMP3]], [[V3:%.*]], i32 3) ; RV64-NEXT: [[TMP5:%.*]] = call target("riscv.vector.tuple", , 6) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_6t.nxv8i8(target("riscv.vector.tuple", , 6) [[TMP4]], [[V4:%.*]], i32 4) ; RV64-NEXT: [[TMP6:%.*]] = call target("riscv.vector.tuple", , 6) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_6t.nxv8i8(target("riscv.vector.tuple", , 6) [[TMP5]], [[V5:%.*]], i32 5) ; RV64-NEXT: call void @llvm.riscv.vsseg6.mask.triscv.vector.tuple_nxv8i8_6t.p0.nxv8i1.i64(target("riscv.vector.tuple", , 6) [[TMP6]], ptr [[PTR:%.*]], splat (i1 true), i64 -1, i64 3) ; RV64-NEXT: ret void ; %interleaved.vec = call @llvm.vector.interleave6.nxv8i8( %v0, %v1, %v2, %v3, %v4, %v5) store %interleaved.vec, ptr %ptr ret void } define void @store_factor7_vscale(ptr %ptr, %v0, %v1, %v2, %v3, %v4, %v5, %v6) { ; RV32-LABEL: @store_factor7_vscale( ; RV32-NEXT: [[TMP1:%.*]] = call target("riscv.vector.tuple", , 7) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_7t.nxv8i8(target("riscv.vector.tuple", , 7) poison, [[V0:%.*]], i32 0) ; RV32-NEXT: [[TMP2:%.*]] = call target("riscv.vector.tuple", , 7) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_7t.nxv8i8(target("riscv.vector.tuple", , 7) [[TMP1]], [[V1:%.*]], i32 1) ; RV32-NEXT: [[TMP3:%.*]] = call target("riscv.vector.tuple", , 7) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_7t.nxv8i8(target("riscv.vector.tuple", , 7) [[TMP2]], [[V2:%.*]], i32 2) ; RV32-NEXT: [[TMP4:%.*]] = call target("riscv.vector.tuple", , 7) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_7t.nxv8i8(target("riscv.vector.tuple", , 7) [[TMP3]], [[V3:%.*]], i32 3) ; RV32-NEXT: [[TMP5:%.*]] = call target("riscv.vector.tuple", , 7) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_7t.nxv8i8(target("riscv.vector.tuple", , 7) [[TMP4]], [[V4:%.*]], i32 4) ; RV32-NEXT: [[TMP6:%.*]] = call target("riscv.vector.tuple", , 7) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_7t.nxv8i8(target("riscv.vector.tuple", , 7) [[TMP5]], [[V5:%.*]], i32 5) ; RV32-NEXT: [[TMP7:%.*]] = call target("riscv.vector.tuple", , 7) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_7t.nxv8i8(target("riscv.vector.tuple", , 7) [[TMP6]], [[V6:%.*]], i32 6) ; RV32-NEXT: call void @llvm.riscv.vsseg7.mask.triscv.vector.tuple_nxv8i8_7t.p0.nxv8i1.i32(target("riscv.vector.tuple", , 7) [[TMP7]], ptr [[PTR:%.*]], splat (i1 true), i32 -1, i32 3) ; RV32-NEXT: ret void ; ; RV64-LABEL: @store_factor7_vscale( ; RV64-NEXT: [[TMP1:%.*]] = call target("riscv.vector.tuple", , 7) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_7t.nxv8i8(target("riscv.vector.tuple", , 7) poison, [[V0:%.*]], i32 0) ; RV64-NEXT: [[TMP2:%.*]] = call target("riscv.vector.tuple", , 7) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_7t.nxv8i8(target("riscv.vector.tuple", , 7) [[TMP1]], [[V1:%.*]], i32 1) ; RV64-NEXT: [[TMP3:%.*]] = call target("riscv.vector.tuple", , 7) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_7t.nxv8i8(target("riscv.vector.tuple", , 7) [[TMP2]], [[V2:%.*]], i32 2) ; RV64-NEXT: [[TMP4:%.*]] = call target("riscv.vector.tuple", , 7) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_7t.nxv8i8(target("riscv.vector.tuple", , 7) [[TMP3]], [[V3:%.*]], i32 3) ; RV64-NEXT: [[TMP5:%.*]] = call target("riscv.vector.tuple", , 7) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_7t.nxv8i8(target("riscv.vector.tuple", , 7) [[TMP4]], [[V4:%.*]], i32 4) ; RV64-NEXT: [[TMP6:%.*]] = call target("riscv.vector.tuple", , 7) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_7t.nxv8i8(target("riscv.vector.tuple", , 7) [[TMP5]], [[V5:%.*]], i32 5) ; RV64-NEXT: [[TMP7:%.*]] = call target("riscv.vector.tuple", , 7) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_7t.nxv8i8(target("riscv.vector.tuple", , 7) [[TMP6]], [[V6:%.*]], i32 6) ; RV64-NEXT: call void @llvm.riscv.vsseg7.mask.triscv.vector.tuple_nxv8i8_7t.p0.nxv8i1.i64(target("riscv.vector.tuple", , 7) [[TMP7]], ptr [[PTR:%.*]], splat (i1 true), i64 -1, i64 3) ; RV64-NEXT: ret void ; %interleaved.vec = call @llvm.vector.interleave7.nxv8i8( %v0, %v1, %v2, %v3, %v4, %v5, %v6) store %interleaved.vec, ptr %ptr ret void } define void @store_factor8_vscale(ptr %ptr, %v0, %v1, %v2, %v3, %v4, %v5, %v6, %v7) { ; RV32-LABEL: @store_factor8_vscale( ; RV32-NEXT: [[TMP1:%.*]] = call target("riscv.vector.tuple", , 8) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_8t.nxv8i8(target("riscv.vector.tuple", , 8) poison, [[V0:%.*]], i32 0) ; RV32-NEXT: [[TMP2:%.*]] = call target("riscv.vector.tuple", , 8) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_8t.nxv8i8(target("riscv.vector.tuple", , 8) [[TMP1]], [[V1:%.*]], i32 1) ; RV32-NEXT: [[TMP3:%.*]] = call target("riscv.vector.tuple", , 8) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_8t.nxv8i8(target("riscv.vector.tuple", , 8) [[TMP2]], [[V2:%.*]], i32 2) ; RV32-NEXT: [[TMP4:%.*]] = call target("riscv.vector.tuple", , 8) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_8t.nxv8i8(target("riscv.vector.tuple", , 8) [[TMP3]], [[V3:%.*]], i32 3) ; RV32-NEXT: [[TMP5:%.*]] = call target("riscv.vector.tuple", , 8) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_8t.nxv8i8(target("riscv.vector.tuple", , 8) [[TMP4]], [[V4:%.*]], i32 4) ; RV32-NEXT: [[TMP6:%.*]] = call target("riscv.vector.tuple", , 8) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_8t.nxv8i8(target("riscv.vector.tuple", , 8) [[TMP5]], [[V5:%.*]], i32 5) ; RV32-NEXT: [[TMP7:%.*]] = call target("riscv.vector.tuple", , 8) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_8t.nxv8i8(target("riscv.vector.tuple", , 8) [[TMP6]], [[V6:%.*]], i32 6) ; RV32-NEXT: [[TMP8:%.*]] = call target("riscv.vector.tuple", , 8) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_8t.nxv8i8(target("riscv.vector.tuple", , 8) [[TMP7]], [[V7:%.*]], i32 7) ; RV32-NEXT: call void @llvm.riscv.vsseg8.mask.triscv.vector.tuple_nxv8i8_8t.p0.nxv8i1.i32(target("riscv.vector.tuple", , 8) [[TMP8]], ptr [[PTR:%.*]], splat (i1 true), i32 -1, i32 3) ; RV32-NEXT: ret void ; ; RV64-LABEL: @store_factor8_vscale( ; RV64-NEXT: [[TMP1:%.*]] = call target("riscv.vector.tuple", , 8) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_8t.nxv8i8(target("riscv.vector.tuple", , 8) poison, [[V0:%.*]], i32 0) ; RV64-NEXT: [[TMP2:%.*]] = call target("riscv.vector.tuple", , 8) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_8t.nxv8i8(target("riscv.vector.tuple", , 8) [[TMP1]], [[V1:%.*]], i32 1) ; RV64-NEXT: [[TMP3:%.*]] = call target("riscv.vector.tuple", , 8) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_8t.nxv8i8(target("riscv.vector.tuple", , 8) [[TMP2]], [[V2:%.*]], i32 2) ; RV64-NEXT: [[TMP4:%.*]] = call target("riscv.vector.tuple", , 8) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_8t.nxv8i8(target("riscv.vector.tuple", , 8) [[TMP3]], [[V3:%.*]], i32 3) ; RV64-NEXT: [[TMP5:%.*]] = call target("riscv.vector.tuple", , 8) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_8t.nxv8i8(target("riscv.vector.tuple", , 8) [[TMP4]], [[V4:%.*]], i32 4) ; RV64-NEXT: [[TMP6:%.*]] = call target("riscv.vector.tuple", , 8) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_8t.nxv8i8(target("riscv.vector.tuple", , 8) [[TMP5]], [[V5:%.*]], i32 5) ; RV64-NEXT: [[TMP7:%.*]] = call target("riscv.vector.tuple", , 8) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_8t.nxv8i8(target("riscv.vector.tuple", , 8) [[TMP6]], [[V6:%.*]], i32 6) ; RV64-NEXT: [[TMP8:%.*]] = call target("riscv.vector.tuple", , 8) @llvm.riscv.tuple.insert.triscv.vector.tuple_nxv8i8_8t.nxv8i8(target("riscv.vector.tuple", , 8) [[TMP7]], [[V7:%.*]], i32 7) ; RV64-NEXT: call void @llvm.riscv.vsseg8.mask.triscv.vector.tuple_nxv8i8_8t.p0.nxv8i1.i64(target("riscv.vector.tuple", , 8) [[TMP8]], ptr [[PTR:%.*]], splat (i1 true), i64 -1, i64 3) ; RV64-NEXT: ret void ; %interleaved.vec = call @llvm.vector.interleave8.nxv8i8( %v0, %v1, %v2, %v3, %v4, %v5, %v6, %v7) store %interleaved.vec, ptr %ptr ret void } define void @load_factor2_fp128(ptr %ptr) { ; RV32-LABEL: @load_factor2_fp128( ; RV32-NEXT: [[INTERLEAVED_VEC:%.*]] = load <4 x fp128>, ptr [[PTR:%.*]], align 16 ; RV32-NEXT: [[V0:%.*]] = shufflevector <4 x fp128> [[INTERLEAVED_VEC]], <4 x fp128> poison, <2 x i32> ; RV32-NEXT: [[V1:%.*]] = shufflevector <4 x fp128> [[INTERLEAVED_VEC]], <4 x fp128> poison, <2 x i32> ; RV32-NEXT: ret void ; ; RV64-LABEL: @load_factor2_fp128( ; RV64-NEXT: [[INTERLEAVED_VEC:%.*]] = load <4 x fp128>, ptr [[PTR:%.*]], align 16 ; RV64-NEXT: [[V0:%.*]] = shufflevector <4 x fp128> [[INTERLEAVED_VEC]], <4 x fp128> poison, <2 x i32> ; RV64-NEXT: [[V1:%.*]] = shufflevector <4 x fp128> [[INTERLEAVED_VEC]], <4 x fp128> poison, <2 x i32> ; RV64-NEXT: ret void ; %interleaved.vec = load <4 x fp128>, ptr %ptr, align 16 %v0 = shufflevector <4 x fp128> %interleaved.vec, <4 x fp128> poison, <2 x i32> %v1 = shufflevector <4 x fp128> %interleaved.vec, <4 x fp128> poison, <2 x i32> ret void } define void @load_factor2_f32(ptr %ptr) { ; RV32-LABEL: @load_factor2_f32( ; RV32-NEXT: [[TMP1:%.*]] = call { <8 x float>, <8 x float> } @llvm.riscv.seg2.load.mask.v8f32.p0.i32(ptr [[PTR:%.*]], <8 x i1> splat (i1 true), i32 8) ; RV32-NEXT: [[TMP2:%.*]] = extractvalue { <8 x float>, <8 x float> } [[TMP1]], 1 ; RV32-NEXT: [[TMP3:%.*]] = extractvalue { <8 x float>, <8 x float> } [[TMP1]], 0 ; RV32-NEXT: ret void ; ; RV64-LABEL: @load_factor2_f32( ; RV64-NEXT: [[TMP1:%.*]] = call { <8 x float>, <8 x float> } @llvm.riscv.seg2.load.mask.v8f32.p0.i64(ptr [[PTR:%.*]], <8 x i1> splat (i1 true), i64 8) ; RV64-NEXT: [[TMP2:%.*]] = extractvalue { <8 x float>, <8 x float> } [[TMP1]], 1 ; RV64-NEXT: [[TMP3:%.*]] = extractvalue { <8 x float>, <8 x float> } [[TMP1]], 0 ; RV64-NEXT: ret void ; %interleaved.vec = load <16 x float>, ptr %ptr %v0 = shufflevector <16 x float> %interleaved.vec, <16 x float> poison, <8 x i32> %v1 = shufflevector <16 x float> %interleaved.vec, <16 x float> poison, <8 x i32> ret void } define void @load_factor2_f64(ptr %ptr) { ; RV32-LABEL: @load_factor2_f64( ; RV32-NEXT: [[TMP1:%.*]] = call { <8 x double>, <8 x double> } @llvm.riscv.seg2.load.mask.v8f64.p0.i32(ptr [[PTR:%.*]], <8 x i1> splat (i1 true), i32 8) ; RV32-NEXT: [[TMP2:%.*]] = extractvalue { <8 x double>, <8 x double> } [[TMP1]], 1 ; RV32-NEXT: [[TMP3:%.*]] = extractvalue { <8 x double>, <8 x double> } [[TMP1]], 0 ; RV32-NEXT: ret void ; ; RV64-LABEL: @load_factor2_f64( ; RV64-NEXT: [[TMP1:%.*]] = call { <8 x double>, <8 x double> } @llvm.riscv.seg2.load.mask.v8f64.p0.i64(ptr [[PTR:%.*]], <8 x i1> splat (i1 true), i64 8) ; RV64-NEXT: [[TMP2:%.*]] = extractvalue { <8 x double>, <8 x double> } [[TMP1]], 1 ; RV64-NEXT: [[TMP3:%.*]] = extractvalue { <8 x double>, <8 x double> } [[TMP1]], 0 ; RV64-NEXT: ret void ; %interleaved.vec = load <16 x double>, ptr %ptr %v0 = shufflevector <16 x double> %interleaved.vec, <16 x double> poison, <8 x i32> %v1 = shufflevector <16 x double> %interleaved.vec, <16 x double> poison, <8 x i32> ret void } define void @load_factor2_bf16(ptr %ptr) { ; RV32-LABEL: @load_factor2_bf16( ; RV32-NEXT: [[INTERLEAVED_VEC:%.*]] = load <16 x bfloat>, ptr [[PTR:%.*]], align 32 ; RV32-NEXT: [[V0:%.*]] = shufflevector <16 x bfloat> [[INTERLEAVED_VEC]], <16 x bfloat> poison, <8 x i32> ; RV32-NEXT: [[V1:%.*]] = shufflevector <16 x bfloat> [[INTERLEAVED_VEC]], <16 x bfloat> poison, <8 x i32> ; RV32-NEXT: ret void ; ; RV64-LABEL: @load_factor2_bf16( ; RV64-NEXT: [[INTERLEAVED_VEC:%.*]] = load <16 x bfloat>, ptr [[PTR:%.*]], align 32 ; RV64-NEXT: [[V0:%.*]] = shufflevector <16 x bfloat> [[INTERLEAVED_VEC]], <16 x bfloat> poison, <8 x i32> ; RV64-NEXT: [[V1:%.*]] = shufflevector <16 x bfloat> [[INTERLEAVED_VEC]], <16 x bfloat> poison, <8 x i32> ; RV64-NEXT: ret void ; %interleaved.vec = load <16 x bfloat>, ptr %ptr %v0 = shufflevector <16 x bfloat> %interleaved.vec, <16 x bfloat> poison, <8 x i32> %v1 = shufflevector <16 x bfloat> %interleaved.vec, <16 x bfloat> poison, <8 x i32> ret void } define void @load_factor2_f16(ptr %ptr) { ; RV32-LABEL: @load_factor2_f16( ; RV32-NEXT: [[INTERLEAVED_VEC:%.*]] = load <16 x half>, ptr [[PTR:%.*]], align 32 ; RV32-NEXT: [[V0:%.*]] = shufflevector <16 x half> [[INTERLEAVED_VEC]], <16 x half> poison, <8 x i32> ; RV32-NEXT: [[V1:%.*]] = shufflevector <16 x half> [[INTERLEAVED_VEC]], <16 x half> poison, <8 x i32> ; RV32-NEXT: ret void ; ; RV64-LABEL: @load_factor2_f16( ; RV64-NEXT: [[INTERLEAVED_VEC:%.*]] = load <16 x half>, ptr [[PTR:%.*]], align 32 ; RV64-NEXT: [[V0:%.*]] = shufflevector <16 x half> [[INTERLEAVED_VEC]], <16 x half> poison, <8 x i32> ; RV64-NEXT: [[V1:%.*]] = shufflevector <16 x half> [[INTERLEAVED_VEC]], <16 x half> poison, <8 x i32> ; RV64-NEXT: ret void ; %interleaved.vec = load <16 x half>, ptr %ptr %v0 = shufflevector <16 x half> %interleaved.vec, <16 x half> poison, <8 x i32> %v1 = shufflevector <16 x half> %interleaved.vec, <16 x half> poison, <8 x i32> ret void }