aboutsummaryrefslogtreecommitdiff
path: root/gcc
diff options
context:
space:
mode:
authorJose E. Marchesi <jose.marchesi@oracle.com>2024-08-05 16:23:47 +0200
committerJose E. Marchesi <jose.marchesi@oracle.com>2024-08-05 18:09:07 +0200
commitc26534d21159dd4c5d1472f0050b65e148161691 (patch)
tree8cab28ce7c4aef153c318bda5b4bdc70d5cefad6 /gcc
parent70ffc57fd2fdb3c8fa67f11d2e8e6b6275dcc7c0 (diff)
downloadgcc-c26534d21159dd4c5d1472f0050b65e148161691.zip
gcc-c26534d21159dd4c5d1472f0050b65e148161691.tar.gz
gcc-c26534d21159dd4c5d1472f0050b65e148161691.tar.bz2
bpf: do not emit BPF non-fetching atomic instructions
When GCC finds a call to one of the __atomic_OP_fetch built-ins in which the return value is not used it optimizes it into the corresponding non-fetching atomic operation. Up to now we had definitions in gcc/config/bpf/atomic.md to implement both atomic_OP and atomic_fetch_OP sets of insns: atomic_add -> aadd (aka xadd) atomic_and -> aand atomic_or -> aor atomic_xor -> axor atomic_fetch_add -> afadd atomic_fetch_and -> afand atomic_fetch_or -> afor atomic_fetch_xor -> afxor This was not correct, because as it happens the non-fetching BPF atomic instructions imply different memory ordering semantics than the fetching BPF atomic instructions, and they cannot be used interchangeably, as it would be expected. This patch modifies config/bpf/atomic.md in order to not define the atomic_{add,and,or,xor} insns. This makes GCC to implement them in terms of the corresponding fetching operations; this is less efficient, but correct. It also updates the expected results in the corresponding tests, which are also updated to cover cases where the value resulting from the __atomic_fetch_* operations is actually used. Tested in bpf-unknown-none target in x86_64-linux-gnu host. gcc/ChangeLog * config/bpf/atomic.md ("atomic_add<AMO:mode>"): Remove insn. ("atomic_and<AMO:mode>"): Likewise ("atomic_or<AMO:mode>"): Likewise. ("atomic_xor<AMO:mode>"): Likewise. gcc/testsuite/ChangeLog * gcc.target/bpf/atomic-op-1.c (test_used_atomic_add): New function. (test_used_atomic_sub): Likewise. (test_used_atomic_and): Likewise. (test_used_atomic_nand): Likewise. (test_used_atomic_or): Likewise. (test_used_atomic_xor): Likewise. * gcc.target/bpf/atomic-op-2.c (test_used_atomic_add): Likewise. (test_used_atomic_sub): Likewise. (test_used_atomic_and): Likewise. (test_used_atomic_nand): Likewise. (test_used_atomic_or): Likewise. (test_used_atomic_xor): Likewise. * gcc.target/bpf/sync-fetch-and-add.c: Expected results updated.
Diffstat (limited to 'gcc')
-rw-r--r--gcc/config/bpf/atomic.md59
-rw-r--r--gcc/testsuite/gcc.target/bpf/atomic-op-1.c53
-rw-r--r--gcc/testsuite/gcc.target/bpf/atomic-op-2.c53
-rw-r--r--gcc/testsuite/gcc.target/bpf/sync-fetch-and-add.c4
4 files changed, 111 insertions, 58 deletions
diff --git a/gcc/config/bpf/atomic.md b/gcc/config/bpf/atomic.md
index be4511b..4e94c03 100644
--- a/gcc/config/bpf/atomic.md
+++ b/gcc/config/bpf/atomic.md
@@ -22,50 +22,21 @@
;;; Plain atomic modify operations.
-;; Non-fetching atomic add predates all other BPF atomic insns.
-;; Use xadd{w,dw} for compatibility with older GAS without support
-;; for v3 atomics. Newer GAS supports "aadd[32]" in line with the
-;; other atomic operations.
-(define_insn "atomic_add<AMO:mode>"
- [(set (match_operand:AMO 0 "memory_operand" "+m")
- (unspec_volatile:AMO
- [(plus:AMO (match_dup 0)
- (match_operand:AMO 1 "register_operand" "r"))
- (match_operand:SI 2 "const_int_operand")] ;; Memory model.
- UNSPEC_AADD))]
- ""
- "{xadd<mop>\t%0,%1|lock *(<smop> *)%w0 += %w1}"
- [(set_attr "type" "atomic")])
-
-(define_insn "atomic_and<AMO:mode>"
- [(set (match_operand:AMO 0 "memory_operand" "+m")
- (unspec_volatile:AMO
- [(and:AMO (match_dup 0)
- (match_operand:AMO 1 "register_operand" "r"))
- (match_operand:SI 2 "const_int_operand")] ;; Memory model.
- UNSPEC_AAND))]
- "bpf_has_v3_atomics"
- "{aand<msuffix>\t%0,%1|lock *(<smop> *)%w0 &= %w1}")
-
-(define_insn "atomic_or<AMO:mode>"
- [(set (match_operand:AMO 0 "memory_operand" "+m")
- (unspec_volatile:AMO
- [(ior:AMO (match_dup 0)
- (match_operand:AMO 1 "register_operand" "r"))
- (match_operand:SI 2 "const_int_operand")] ;; Memory model.
- UNSPEC_AOR))]
- "bpf_has_v3_atomics"
- "{aor<msuffix>\t%0,%1|lock *(<smop> *)%w0 %|= %w1}")
-
-(define_insn "atomic_xor<AMO:mode>"
- [(set (match_operand:AMO 0 "memory_operand" "+m")
- (unspec_volatile:AMO
- [(xor:AMO (match_dup 0)
- (match_operand:AMO 1 "register_operand" "r"))
- (match_operand:SI 2 "const_int_operand")] ;; Memory model.
- UNSPEC_AXOR))]
- "bpf_has_v3_atomics"
- "{axor<msuffix>\t%0,%1|lock *(<smop> *)%w0 ^= %w1}")
+;; The BPF instruction set provides non-fetching atomic instructions
+;; that could be used to implement the corresponding named insns:
+;;
+;; atomic_add -> aadd (aka xadd)
+;; atomic_and -> aand
+;; atomic_or -> aor
+;; atomic_xor -> axor
+;;
+;; However, we are not including insns for these here because the
+;; non-fetching BPF atomic instruction imply different memory ordering
+;; semantics than the fetching BPF atomic instruction used to
+;; implement the atomic_fetch_* insns below (afadd, afand, afor,
+;; afxor) and they cannot be used interchangeably, as it is expected
+;; by GCC when it uses a non-fetching variant as an optimization of a
+;; fetching operation where the returned value is not used.
;;; Feching (read-modify-store) versions of atomic operations.
diff --git a/gcc/testsuite/gcc.target/bpf/atomic-op-1.c b/gcc/testsuite/gcc.target/bpf/atomic-op-1.c
index 5c87dcb..b7d3d83f 100644
--- a/gcc/testsuite/gcc.target/bpf/atomic-op-1.c
+++ b/gcc/testsuite/gcc.target/bpf/atomic-op-1.c
@@ -1,7 +1,12 @@
-/* Test 64-bit non-fetch atomic operations. */
+/* Test 64-bit fetch and non-fetch atomic operations. */
/* { dg-do compile } */
/* { dg-options "-mv3-atomics -O2 -masm=normal" } */
+/* Note that GCC optimizes __atomic_add_fetch calls whose return value is not
+ used into non-fetching operations that, in BPF, generate fetching
+ instructions anyway. See note in gcc/config/bpf/atomic.md on this
+ regard. */
+
long val;
void
@@ -10,40 +15,76 @@ test_atomic_add (long x)
__atomic_add_fetch (&val, x, __ATOMIC_ACQUIRE);
}
+long
+test_used_atomic_add (long x)
+{
+ return __atomic_add_fetch (&val, x, __ATOMIC_ACQUIRE);
+}
+
void
test_atomic_sub (long x)
{
__atomic_add_fetch (&val, x, __ATOMIC_ACQUIRE);
}
+long
+test_used_atomic_sub (long x)
+{
+ return __atomic_add_fetch (&val, x, __ATOMIC_ACQUIRE);
+}
+
void
test_atomic_and (long x)
{
__atomic_and_fetch (&val, x, __ATOMIC_ACQUIRE);
}
+long
+test_used_atomic_and (long x)
+{
+ return __atomic_and_fetch (&val, x, __ATOMIC_ACQUIRE);
+}
+
void
test_atomic_nand (long x)
{
__atomic_nand_fetch (&val, x, __ATOMIC_ACQUIRE);
}
+long
+test_used_atomic_nand (long x)
+{
+ return __atomic_nand_fetch (&val, x, __ATOMIC_ACQUIRE);
+}
+
void
test_atomic_or (long x)
{
__atomic_or_fetch (&val, x, __ATOMIC_ACQUIRE);
}
+long
+test_used_atomic_or (long x)
+{
+ return __atomic_or_fetch (&val, x, __ATOMIC_ACQUIRE);
+}
+
void
test_atomic_xor (long x)
{
__atomic_xor_fetch (&val, x, __ATOMIC_ACQUIRE);
}
-/* sub implemented in terms of add, and we output xadd to support older GAS. */
-/* { dg-final { scan-assembler-times "xadddw\t" 2 } } */
-/* { dg-final { scan-assembler-times "aand\t" 1 } } */
+long
+test_used_atomic_xor (long x)
+{
+ return __atomic_xor_fetch (&val, x, __ATOMIC_ACQUIRE);
+}
+
+/* sub implemented in terms of add. */
+/* { dg-final { scan-assembler-times "afadd\t" 4 } } */
+/* { dg-final { scan-assembler-times "afand\t" 2 } } */
/* nand must use an exchange loop */
/* { dg-final { scan-assembler "acmp\t" } } */
-/* { dg-final { scan-assembler-times "aor\t" 1 } } */
-/* { dg-final { scan-assembler-times "axor\t" 1 } } */
+/* { dg-final { scan-assembler-times "afor\t" 2 } } */
+/* { dg-final { scan-assembler-times "afxor\t" 2 } } */
diff --git a/gcc/testsuite/gcc.target/bpf/atomic-op-2.c b/gcc/testsuite/gcc.target/bpf/atomic-op-2.c
index 8331d33..902142d 100644
--- a/gcc/testsuite/gcc.target/bpf/atomic-op-2.c
+++ b/gcc/testsuite/gcc.target/bpf/atomic-op-2.c
@@ -1,7 +1,12 @@
-/* Test 32-bit non-fetch atomic operations. */
+/* Test 32-bit fetch and non-fetch atomic operations. */
/* { dg-do compile } */
/* { dg-options "-mv3-atomics -O2 -masm=normal" } */
+/* Note that GCC optimizes __atomic_add_fetch calls whose return value is not
+ used into non-fetching operations that, in BPF, generate fetching
+ instructions anyway. See note in gcc/config/bpf/atomic.md on this
+ regard. */
+
int val;
void
@@ -10,40 +15,76 @@ test_atomic_add (int x)
__atomic_add_fetch (&val, x, __ATOMIC_ACQUIRE);
}
+long
+test_used_atomic_add (int x)
+{
+ return __atomic_add_fetch (&val, x, __ATOMIC_ACQUIRE);
+}
+
void
test_atomic_sub (int x)
{
__atomic_add_fetch (&val, x, __ATOMIC_ACQUIRE);
}
+long
+test_used_atomic_sub (int x)
+{
+ return __atomic_add_fetch (&val, x, __ATOMIC_ACQUIRE);
+}
+
void
test_atomic_and (int x)
{
__atomic_and_fetch (&val, x, __ATOMIC_ACQUIRE);
}
+long
+test_used_atomic_and (int x)
+{
+ return __atomic_and_fetch (&val, x, __ATOMIC_ACQUIRE);
+}
+
void
test_atomic_nand (int x)
{
__atomic_nand_fetch (&val, x, __ATOMIC_ACQUIRE);
}
+long
+test_used_atomic_nand (int x)
+{
+ return __atomic_nand_fetch (&val, x, __ATOMIC_ACQUIRE);
+}
+
void
test_atomic_or (int x)
{
__atomic_or_fetch (&val, x, __ATOMIC_ACQUIRE);
}
+long
+test_used_atomic_or (int x)
+{
+ return __atomic_or_fetch (&val, x, __ATOMIC_ACQUIRE);
+}
+
void
test_atomic_xor (int x)
{
__atomic_xor_fetch (&val, x, __ATOMIC_ACQUIRE);
}
-/* sub implemented in terms of add, and we output xadd to support older GAS. */
-/* { dg-final { scan-assembler-times "xaddw\t" 2 } } */
-/* { dg-final { scan-assembler-times "aand32\t" 1 } } */
+long
+test_used_atomic_xor (int x)
+{
+ return __atomic_xor_fetch (&val, x, __ATOMIC_ACQUIRE);
+}
+
+/* sub implemented in terms of add. */
+/* { dg-final { scan-assembler-times "afadd32" 4 } } */
+/* { dg-final { scan-assembler-times "afand32\t" 2 } } */
/* nand must use an exchange loop */
/* { dg-final { scan-assembler "acmp32\t" } } */
-/* { dg-final { scan-assembler-times "aor32\t" 1 } } */
-/* { dg-final { scan-assembler-times "axor32\t" 1 } } */
+/* { dg-final { scan-assembler-times "afor32\t" 2 } } */
+/* { dg-final { scan-assembler-times "afxor32\t" 2 } } */
diff --git a/gcc/testsuite/gcc.target/bpf/sync-fetch-and-add.c b/gcc/testsuite/gcc.target/bpf/sync-fetch-and-add.c
index 6902aab..fe4f81a 100644
--- a/gcc/testsuite/gcc.target/bpf/sync-fetch-and-add.c
+++ b/gcc/testsuite/gcc.target/bpf/sync-fetch-and-add.c
@@ -11,5 +11,5 @@ foo ()
__sync_fetch_and_add((int *)val, (int)delta);
}
-/* { dg-final { scan-assembler "xadddw\t.*" } } */
-/* { dg-final { scan-assembler "xaddw\t.*" } } */
+/* { dg-final { scan-assembler "afadd\t.*" } } */
+/* { dg-final { scan-assembler "afadd32\t.*" } } */