/* PR target/98737 */ /* { dg-do compile { target i?86-*-* x86_64-*-* powerpc*-*-* aarch64*-*-* } } */ /* { dg-options "-O2 -fdump-tree-optimized -fcompare-debug" } */ /* { dg-additional-options "-march=i686" { target ia32 } } */ /* Note the choice between __atomic_fetch_and_* and __atomic_and_*_fetch if the result is not used does not matter. so check if make sure we don't have __atomic_fetch_ with an assignment */ /* { dg-final { scan-tree-dump-not "= __atomic_fetch_" "optimized" } } */ /* { dg-final { scan-tree-dump-not "= __sync_fetch_and_" "optimized" } } */ typedef signed char schar; typedef unsigned long ulong; typedef unsigned int uint; typedef unsigned short ushort; typedef unsigned char uchar; long vlong; int vint; short vshort; schar vschar; ulong vulong; uint vuint; ushort vushort; uchar vuchar; #define A(n, t, ut, f, o, ...) \ t fn##n (t x) \ { \ ut z = f (&v##t, x, ##__VA_ARGS__); \ t w = (t) z; \ return w o x; \ } #define B(n, f, o, ...) \ A(n##0, long, ulong, f, o, ##__VA_ARGS__) \ A(n##1, int, uint, f, o, ##__VA_ARGS__) \ A(n##2, short, ushort, f, o, ##__VA_ARGS__) \ A(n##3, schar, uchar, f, o, ##__VA_ARGS__) \ A(n##4, ulong, ulong, f, o, ##__VA_ARGS__) \ A(n##5, uint, uint, f, o, ##__VA_ARGS__) \ A(n##6, ushort, ushort, f, o, ##__VA_ARGS__) \ A(n##7, uchar, uchar, f, o, ##__VA_ARGS__) B(00, __atomic_fetch_add, +, __ATOMIC_RELAXED) B(01, __atomic_fetch_sub, -, __ATOMIC_RELAXED) B(02, __atomic_fetch_and, &, __ATOMIC_RELAXED) B(03, __atomic_fetch_xor, ^, __ATOMIC_RELAXED) B(04, __atomic_fetch_or, |, __ATOMIC_RELAXED) B(05, __sync_fetch_and_add, +) B(06, __sync_fetch_and_sub, -) B(07, __sync_fetch_and_and, &) B(08, __sync_fetch_and_xor, ^) B(09, __sync_fetch_and_or, |) #undef A #define A(n, t, ut, f, o, ...) \ t fn##n (void) \ { \ ut z = f (&v##t, 42, ##__VA_ARGS__); \ t w = (t) z; \ return w o 42; \ } B(10, __atomic_fetch_add, +, __ATOMIC_RELAXED) B(11, __atomic_fetch_sub, -, __ATOMIC_RELAXED) B(12, __atomic_fetch_and, &, __ATOMIC_RELAXED) B(13, __atomic_fetch_xor, ^, __ATOMIC_RELAXED) B(14, __atomic_fetch_or, |, __ATOMIC_RELAXED) B(15, __sync_fetch_and_add, +) B(16, __sync_fetch_and_sub, -) B(17, __sync_fetch_and_and, &) B(18, __sync_fetch_and_xor, ^) B(19, __sync_fetch_and_or, |) #undef A #define A(n, t, ut, f, o, ...) \ t fn##n (t x) \ { \ ut z = f (&v##t, x, ##__VA_ARGS__); \ t w = (t) z; \ t v = w o x; \ return v == 0; \ } B(20, __atomic_fetch_add, +, __ATOMIC_RELAXED) B(21, __atomic_fetch_sub, -, __ATOMIC_RELAXED) B(22, __atomic_fetch_and, &, __ATOMIC_RELAXED) B(23, __atomic_fetch_xor, ^, __ATOMIC_RELAXED) B(24, __atomic_fetch_or, |, __ATOMIC_RELAXED) B(25, __sync_fetch_and_add, +) B(26, __sync_fetch_and_sub, -) B(27, __sync_fetch_and_and, &) B(28, __sync_fetch_and_xor, ^) B(29, __sync_fetch_and_or, |) #undef A #define A(n, t, ut, f, o, ...) \ t fn##n (void) \ { \ ut z = f (&v##t, 42, ##__VA_ARGS__); \ t w = (t) z; \ t v = w o 42; \ return v != 0; \ } B(30, __atomic_fetch_add, +, __ATOMIC_RELAXED) B(31, __atomic_fetch_sub, -, __ATOMIC_RELAXED) B(32, __atomic_fetch_and, &, __ATOMIC_RELAXED) B(33, __atomic_fetch_xor, ^, __ATOMIC_RELAXED) B(34, __atomic_fetch_or, |, __ATOMIC_RELAXED) B(35, __sync_fetch_and_add, +) B(36, __sync_fetch_and_sub, -) B(37, __sync_fetch_and_and, &) B(38, __sync_fetch_and_xor, ^) B(39, __sync_fetch_and_or, |) #undef A #define A(n, t, ut, f, o, ...) \ t fn##n (t x) \ { \ return (t) (((t) f (&v##t, x, ##__VA_ARGS__)) \ o x) != 0; \ } B(40, __atomic_fetch_add, +, __ATOMIC_RELAXED) B(41, __atomic_fetch_sub, -, __ATOMIC_RELAXED) B(42, __atomic_fetch_and, &, __ATOMIC_RELAXED) B(43, __atomic_fetch_xor, ^, __ATOMIC_RELAXED) B(44, __atomic_fetch_or, |, __ATOMIC_RELAXED) B(45, __sync_fetch_and_add, +) B(46, __sync_fetch_and_sub, -) B(47, __sync_fetch_and_and, &) B(48, __sync_fetch_and_xor, ^) B(49, __sync_fetch_and_or, |) #undef A #define A(n, t, ut, f, o, ...) \ t fn##n (void) \ { \ return (t) (((t) f (&v##t, 42, ##__VA_ARGS__))\ o 42) == 0; \ } B(50, __atomic_fetch_add, +, __ATOMIC_RELAXED) B(51, __atomic_fetch_sub, -, __ATOMIC_RELAXED) B(52, __atomic_fetch_and, &, __ATOMIC_RELAXED) B(53, __atomic_fetch_xor, ^, __ATOMIC_RELAXED) /* (whatever | 42) == 0 is 0, so we can't test this. */ /* B(54, __atomic_fetch_or, |, __ATOMIC_RELAXED) */ B(55, __sync_fetch_and_add, +) B(56, __sync_fetch_and_sub, -) B(57, __sync_fetch_and_and, &) B(58, __sync_fetch_and_xor, ^) /* B(59, __sync_fetch_and_or, |) */