blob: 6d8515061694abaac70498d64ea21a0f6048c447 (
plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
|
#include "tree-vect.h"
#if VECTOR_BITS > 256
#define NINTS (VECTOR_BITS / 32)
#else
#define NINTS 8
#endif
#define N (NINTS * 2)
#define RESULT (NINTS * (NINTS - 1) / 2 * N + NINTS)
extern void abort (void);
typedef struct giga
{
unsigned int g[N];
} giga;
unsigned long __attribute__((noinline,noclone))
addfst(giga const *gptr, int num)
{
unsigned int retval = 0;
int i;
for (i = 0; i < num; i++)
retval += gptr[i].g[0];
return retval;
}
int main ()
{
struct giga g[NINTS];
unsigned int n = 1;
int i, j;
check_vect ();
for (i = 0; i < NINTS; ++i)
for (j = 0; j < N; ++j)
{
g[i].g[j] = n++;
__asm__ volatile ("");
}
if (addfst (g, NINTS) != RESULT)
abort ();
return 0;
}
/* We don't want to vectorize the single-element interleaving in the way
we currently do that (without ignoring not needed vectors in the
gap between gptr[0].g[0] and gptr[1].g[0]), because that's very
sub-optimal and causes memory explosion (even though the cost model
should reject that in the end). */
/* { dg-final { scan-tree-dump-times "vectorized 0 loops in function" 2 "vect" { target {! riscv*-*-* } } } } */
/* We should end up using gathers for the strided load on RISC-V. */
/* { dg-final { scan-tree-dump-times "vectorized 1 loops in function" 1 "vect" { target { riscv*-*-* } } } } */
/* { dg-final { scan-tree-dump "using gather/scatter for strided/grouped access" "vect" { target { riscv*-*-* } } } } */
|