;; Machine description for LoongArch atomic operations. ;; Copyright (C) 2021-2024 Free Software Foundation, Inc. ;; Contributed by Loongson Ltd. ;; Based on MIPS and RISC-V target for GNU compiler. ;; This file is part of GCC. ;; GCC is free software; you can redistribute it and/or modify ;; it under the terms of the GNU General Public License as published by ;; the Free Software Foundation; either version 3, or (at your option) ;; any later version. ;; GCC is distributed in the hope that it will be useful, ;; but WITHOUT ANY WARRANTY; without even the implied warranty of ;; MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ;; GNU General Public License for more details. ;; You should have received a copy of the GNU General Public License ;; along with GCC; see the file COPYING3. If not see ;; . (define_c_enum "unspec" [ UNSPEC_COMPARE_AND_SWAP UNSPEC_COMPARE_AND_SWAP_ADD UNSPEC_COMPARE_AND_SWAP_SUB UNSPEC_COMPARE_AND_SWAP_AND UNSPEC_COMPARE_AND_SWAP_XOR UNSPEC_COMPARE_AND_SWAP_OR UNSPEC_COMPARE_AND_SWAP_NAND UNSPEC_SYNC_OLD_OP UNSPEC_SYNC_EXCHANGE UNSPEC_ATOMIC_STORE UNSPEC_ATOMIC_LOAD UNSPEC_MEMORY_BARRIER ]) (define_code_iterator any_atomic [plus ior xor and]) (define_code_attr atomic_optab [(plus "add") (ior "or") (xor "xor") (and "and")]) ;; This attribute gives the format suffix for atomic memory operations. (define_mode_attr amo [(QI "b") (HI "h") (SI "w") (DI "d")]) ;; expands to the name of the atomic operand that implements a ;; particular code. (define_code_attr amop [(ior "or") (xor "xor") (and "and") (plus "add")]) ;; Memory barriers. (define_expand "mem_thread_fence" [(match_operand:SI 0 "const_int_operand" "")] ;; model "" { rtx mem = gen_rtx_MEM (BLKmode, gen_rtx_SCRATCH (Pmode)); MEM_VOLATILE_P (mem) = 1; emit_insn (gen_mem_thread_fence_1 (mem, operands[0])); DONE; }) ;; DBAR hint encoding for LA664 and later micro-architectures, paraphrased from ;; the Linux patch revealing it [1]: ;; ;; - Bit 4: kind of constraint (0: completion, 1: ordering) ;; - Bit 3: barrier for previous read (0: true, 1: false) ;; - Bit 2: barrier for previous write (0: true, 1: false) ;; - Bit 1: barrier for succeeding read (0: true, 1: false) ;; - Bit 0: barrier for succeeding write (0: true, 1: false) ;; ;; [1]: https://git.kernel.org/torvalds/c/e031a5f3f1ed ;; ;; Implementations without support for the finer-granularity hints simply treat ;; all as the full barrier (DBAR 0), so we can unconditionally start emiting the ;; more precise hints right away. (define_insn "mem_thread_fence_1" [(set (match_operand:BLK 0 "" "") (unspec:BLK [(match_dup 0)] UNSPEC_MEMORY_BARRIER)) (match_operand:SI 1 "const_int_operand" "")] ;; model "" { enum memmodel model = memmodel_base (INTVAL (operands[1])); switch (model) { case MEMMODEL_ACQUIRE: return "dbar\t0b10100"; case MEMMODEL_RELEASE: return "dbar\t0b10010"; case MEMMODEL_ACQ_REL: case MEMMODEL_SEQ_CST: return "dbar\t0b10000"; default: /* GCC internal: "For the '__ATOMIC_RELAXED' model no instructions need to be issued and this expansion is not invoked." __atomic builtins doc: "Consume is implemented using the stronger acquire memory order because of a deficiency in C++11's semantics." See PR 59448 and get_memmodel in builtins.cc. Other values should not be returned by memmodel_base. */ gcc_unreachable (); } }) ;; Atomic memory operations. (define_insn "atomic_load" [(set (match_operand:QHWD 0 "register_operand" "=r") (unspec_volatile:QHWD [(match_operand:QHWD 1 "memory_operand" "+m") (match_operand:SI 2 "const_int_operand")] ;; model UNSPEC_ATOMIC_LOAD))] "" { enum memmodel model = memmodel_base (INTVAL (operands[2])); switch (model) { case MEMMODEL_SEQ_CST: return "dbar\t0x11\\n\\t" "ld.\t%0,%1\\n\\t" "dbar\t0x14"; case MEMMODEL_ACQUIRE: return "ld.\t%0,%1\\n\\t" "dbar\t0x14"; case MEMMODEL_RELAXED: return ISA_HAS_LD_SEQ_SA ? "ld.\t%0,%1" : "ld.\t%0,%1\\n\\t" "dbar\t0x700"; default: /* The valid memory order variants are __ATOMIC_RELAXED, __ATOMIC_SEQ_CST, __ATOMIC_CONSUME and __ATOMIC_ACQUIRE. The expand_builtin_atomic_store function converts all invalid memmodels to MEMMODEL_SEQ_CST. __atomic builtins doc: "Consume is implemented using the stronger acquire memory order because of a deficiency in C++11's semantics." See PR 59448 and get_memmodel in builtins.cc. */ gcc_unreachable (); } } [(set (attr "length") (const_int 12))]) ;; Implement atomic stores with amoswap. Fall back to fences for atomic loads. (define_insn "atomic_store" [(set (match_operand:QHWD 0 "memory_operand" "+m") (unspec_volatile:QHWD [(match_operand:QHWD 1 "reg_or_0_operand" "rJ") (match_operand:SI 2 "const_int_operand")] ;; model UNSPEC_ATOMIC_STORE))] "" { enum memmodel model = memmodel_base (INTVAL (operands[2])); switch (model) { case MEMMODEL_SEQ_CST: return "dbar\t0x12\\n\\t" "st.\t%z1,%0\\n\\t" "dbar\t0x18\\n\\t"; case MEMMODEL_RELEASE: return "dbar\t0x12\\n\\t" "st.\t%z1,%0\\n\\t"; case MEMMODEL_RELAXED: return "st.\t%z1,%0"; default: /* The valid memory order variants are __ATOMIC_RELAXED, __ATOMIC_SEQ_CST, and __ATOMIC_RELEASE. The expand_builtin_atomic_store function converts all invalid memmodels to MEMMODEL_SEQ_CST. */ gcc_unreachable (); } } [(set (attr "length") (const_int 12))]) (define_insn "atomic_" [(set (match_operand:GPR 0 "memory_operand" "+ZB") (unspec_volatile:GPR [(any_atomic:GPR (match_dup 0) (match_operand:GPR 1 "reg_or_0_operand" "rJ")) (match_operand:SI 2 "const_int_operand")] ;; model UNSPEC_SYNC_OLD_OP))] "" "am%A2.\t$zero,%z1,%0" [(set (attr "length") (const_int 4))]) (define_insn "atomic_add" [(set (match_operand:SHORT 0 "memory_operand" "+ZB") (unspec_volatile:SHORT [(plus:SHORT (match_dup 0) (match_operand:SHORT 1 "reg_or_0_operand" "rJ")) (match_operand:SI 2 "const_int_operand")] ;; model UNSPEC_SYNC_OLD_OP))] "ISA_HAS_LAM_BH" "amadd%A2.\t$zero,%z1,%0" [(set (attr "length") (const_int 4))]) (define_insn "atomic_fetch_" [(set (match_operand:GPR 0 "register_operand" "=&r") (match_operand:GPR 1 "memory_operand" "+ZB")) (set (match_dup 1) (unspec_volatile:GPR [(any_atomic:GPR (match_dup 1) (match_operand:GPR 2 "reg_or_0_operand" "rJ")) (match_operand:SI 3 "const_int_operand")] ;; model UNSPEC_SYNC_OLD_OP))] "" "am%A3.\t%0,%z2,%1" [(set (attr "length") (const_int 4))]) (define_insn "atomic_exchange" [(set (match_operand:GPR 0 "register_operand" "=&r") (unspec_volatile:GPR [(match_operand:GPR 1 "memory_operand" "+ZB") (match_operand:SI 3 "const_int_operand")] ;; model UNSPEC_SYNC_EXCHANGE)) (set (match_dup 1) (match_operand:GPR 2 "register_operand" "r"))] "" "amswap%A3.\t%0,%z2,%1" [(set (attr "length") (const_int 4))]) (define_insn "atomic_exchange_short" [(set (match_operand:SHORT 0 "register_operand" "=&r") (unspec_volatile:SHORT [(match_operand:SHORT 1 "memory_operand" "+ZB") (match_operand:SI 3 "const_int_operand")] ;; model UNSPEC_SYNC_EXCHANGE)) (set (match_dup 1) (match_operand:SHORT 2 "register_operand" "r"))] "ISA_HAS_LAM_BH" "amswap%A3.\t%0,%z2,%1" [(set (attr "length") (const_int 4))]) (define_insn "atomic_cas_value_strong" [(set (match_operand:GPR 0 "register_operand" "=&r") (match_operand:GPR 1 "memory_operand" "+ZC")) (set (match_dup 1) (unspec_volatile:GPR [(match_operand:GPR 2 "reg_or_0_operand" "rJ") (match_operand:GPR 3 "reg_or_0_operand" "rJ") (match_operand:SI 4 "const_int_operand")] ;; mod_s UNSPEC_COMPARE_AND_SWAP)) (clobber (match_scratch:GPR 5 "=&r"))] "" { output_asm_insn ("1:", operands); output_asm_insn ("ll.\t%0,%1", operands); /* Like the test case atomic-cas-int.C, in loongarch64, O1 and higher, the return value of the val_without_const_folding will not be truncated and will be passed directly to the function compare_exchange_strong. However, the instruction 'bne' does not distinguish between 32-bit and 64-bit operations. so if the upper 32 bits of the register are not extended by the 32nd bit symbol, then the comparison may not be valid here. This will affect the result of the operation. */ if (TARGET_64BIT && REG_P (operands[2]) && GET_MODE (operands[2]) == SImode) { output_asm_insn ("addi.w\t%5,%2,0", operands); output_asm_insn ("bne\t%0,%5,2f", operands); } else output_asm_insn ("bne\t%0,%z2,2f", operands); output_asm_insn ("or%i3\t%5,$zero,%3", operands); output_asm_insn ("sc.\t%5,%1", operands); output_asm_insn ("beqz\t%5,1b", operands); output_asm_insn ("b\t3f", operands); output_asm_insn ("2:", operands); output_asm_insn ("%G4", operands); output_asm_insn ("3:", operands); return ""; } [(set (attr "length") (if_then_else (and (match_test "GET_MODE (operands[2]) == SImode") (match_test "REG_P (operands[2])")) (const_int 32) (const_int 28)))]) (define_insn "atomic_cas_value_strong_amcas" [(set (match_operand:QHWD 0 "register_operand" "=&r") (match_operand:QHWD 1 "memory_operand" "+ZB")) (set (match_dup 1) (unspec_volatile:QHWD [(match_operand:QHWD 2 "reg_or_0_operand" "rJ") (match_operand:QHWD 3 "reg_or_0_operand" "rJ") (match_operand:SI 4 "const_int_operand")] ;; mod_s UNSPEC_COMPARE_AND_SWAP))] "ISA_HAS_LAMCAS" "ori\t%0,%z2,0\n\tamcas%A4.\t%0,%z3,%1" [(set (attr "length") (const_int 8))]) (define_expand "atomic_compare_and_swap" [(match_operand:SI 0 "register_operand" "") ;; bool output (match_operand:GPR 1 "register_operand" "") ;; val output (match_operand:GPR 2 "memory_operand" "") ;; memory (match_operand:GPR 3 "reg_or_0_operand" "") ;; expected value (match_operand:GPR 4 "reg_or_0_operand" "") ;; desired value (match_operand:SI 5 "const_int_operand" "") ;; is_weak (match_operand:SI 6 "const_int_operand" "") ;; mod_s (match_operand:SI 7 "const_int_operand" "")] ;; mod_f "" { rtx mod_s, mod_f; mod_s = operands[6]; mod_f = operands[7]; /* Normally the succ memory model must be stronger than fail, but in the unlikely event of fail being ACQUIRE and succ being RELEASE we need to promote succ to ACQ_REL so that we don't lose the acquire semantics. */ if (is_mm_acquire (memmodel_base (INTVAL (mod_f))) && is_mm_release (memmodel_base (INTVAL (mod_s)))) mod_s = GEN_INT (MEMMODEL_ACQ_REL); operands[6] = mod_s; if (ISA_HAS_LAMCAS) emit_insn (gen_atomic_cas_value_strong_amcas (operands[1], operands[2], operands[3], operands[4], operands[6])); else emit_insn (gen_atomic_cas_value_strong (operands[1], operands[2], operands[3], operands[4], operands[6])); rtx compare = operands[1]; if (operands[3] != const0_rtx) { rtx difference = gen_rtx_MINUS (mode, operands[1], operands[3]); compare = gen_reg_rtx (mode); emit_insn (gen_rtx_SET (compare, difference)); } if (word_mode != mode) { rtx reg = gen_reg_rtx (word_mode); emit_insn (gen_rtx_SET (reg, gen_rtx_SIGN_EXTEND (word_mode, compare))); compare = reg; } emit_insn (gen_rtx_SET (operands[0], gen_rtx_EQ (SImode, compare, const0_rtx))); DONE; }) (define_expand "atomic_test_and_set" [(match_operand:QI 0 "register_operand" "") ;; bool output (match_operand:QI 1 "memory_operand" "+ZB") ;; memory (match_operand:SI 2 "const_int_operand" "")] ;; model "" { /* We have no QImode atomics, so use the address LSBs to form a mask, then use an aligned SImode atomic. */ rtx result = operands[0]; rtx mem = operands[1]; rtx model = operands[2]; rtx addr = force_reg (Pmode, XEXP (mem, 0)); rtx tmp_reg = gen_reg_rtx (Pmode); rtx zero_reg = gen_rtx_REG (Pmode, 0); rtx aligned_addr = gen_reg_rtx (Pmode); emit_move_insn (tmp_reg, gen_rtx_PLUS (Pmode, zero_reg, GEN_INT (-4))); emit_move_insn (aligned_addr, gen_rtx_AND (Pmode, addr, tmp_reg)); rtx aligned_mem = change_address (mem, SImode, aligned_addr); set_mem_alias_set (aligned_mem, 0); rtx offset = gen_reg_rtx (SImode); emit_move_insn (offset, gen_rtx_AND (SImode, gen_lowpart (SImode, addr), GEN_INT (3))); rtx tmp = gen_reg_rtx (SImode); emit_move_insn (tmp, GEN_INT (1)); rtx shmt = gen_reg_rtx (SImode); emit_move_insn (shmt, gen_rtx_ASHIFT (SImode, offset, GEN_INT (3))); rtx word = gen_reg_rtx (SImode); emit_move_insn (word, gen_rtx_ASHIFT (SImode, tmp, shmt)); tmp = gen_reg_rtx (SImode); emit_insn (gen_atomic_fetch_orsi (tmp, aligned_mem, word, model)); emit_move_insn (gen_lowpart (SImode, result), gen_rtx_LSHIFTRT (SImode, tmp, shmt)); DONE; }) (define_insn "atomic_cas_value_cmp_and_7_" [(set (match_operand:GPR 0 "register_operand" "=&r") (match_operand:GPR 1 "memory_operand" "+ZC")) (set (match_dup 1) (unspec_volatile:GPR [(match_operand:GPR 2 "reg_or_0_operand" "rJ") (match_operand:GPR 3 "reg_or_0_operand" "rJ") (match_operand:GPR 4 "reg_or_0_operand" "rJ") (match_operand:GPR 5 "reg_or_0_operand" "rJ") (match_operand:SI 6 "const_int_operand")] ;; model UNSPEC_COMPARE_AND_SWAP)) (clobber (match_scratch:GPR 7 "=&r"))] "" { return "1:\\n\\t" "ll.\\t%0,%1\\n\\t" "and\\t%7,%0,%2\\n\\t" "bne\\t%7,%z4,2f\\n\\t" "and\\t%7,%0,%z3\\n\\t" "or%i5\\t%7,%7,%5\\n\\t" "sc.\\t%7,%1\\n\\t" "beq\\t$zero,%7,1b\\n\\t" "b\\t3f\\n\\t" "2:\\n\\t" "%G6\\n\\t" "3:\\n\\t"; } [(set (attr "length") (const_int 36))]) (define_expand "atomic_compare_and_swap" [(match_operand:SI 0 "register_operand" "") ;; bool output (match_operand:SHORT 1 "register_operand" "") ;; val output (match_operand:SHORT 2 "memory_operand" "") ;; memory (match_operand:SHORT 3 "reg_or_0_operand" "") ;; expected value (match_operand:SHORT 4 "reg_or_0_operand" "") ;; desired value (match_operand:SI 5 "const_int_operand" "") ;; is_weak (match_operand:SI 6 "const_int_operand" "") ;; mod_s (match_operand:SI 7 "const_int_operand" "")] ;; mod_f "" { rtx mod_s, mod_f; mod_s = operands[6]; mod_f = operands[7]; /* Normally the succ memory model must be stronger than fail, but in the unlikely event of fail being ACQUIRE and succ being RELEASE we need to promote succ to ACQ_REL so that we don't lose the acquire semantics. */ if (is_mm_acquire (memmodel_base (INTVAL (mod_f))) && is_mm_release (memmodel_base (INTVAL (mod_s)))) mod_s = GEN_INT (MEMMODEL_ACQ_REL); operands[6] = mod_s; if (ISA_HAS_LAMCAS) emit_insn (gen_atomic_cas_value_strong_amcas (operands[1], operands[2], operands[3], operands[4], operands[6])); else { union loongarch_gen_fn_ptrs generator; generator.fn_7 = gen_atomic_cas_value_cmp_and_7_si; loongarch_expand_atomic_qihi (generator, operands[1], operands[2], operands[3], operands[4], operands[6]); } rtx compare = operands[1]; if (operands[3] != const0_rtx) { machine_mode mode = GET_MODE (operands[3]); rtx op1 = convert_modes (SImode, mode, operands[1], true); rtx op3 = convert_modes (SImode, mode, operands[3], true); rtx difference = gen_rtx_MINUS (SImode, op1, op3); compare = gen_reg_rtx (SImode); emit_insn (gen_rtx_SET (compare, difference)); } if (word_mode != mode) { rtx reg = gen_reg_rtx (word_mode); emit_insn (gen_rtx_SET (reg, gen_rtx_SIGN_EXTEND (word_mode, compare))); compare = reg; } emit_insn (gen_rtx_SET (operands[0], gen_rtx_EQ (SImode, compare, const0_rtx))); DONE; }) (define_insn "atomic_cas_value_add_7_" [(set (match_operand:GPR 0 "register_operand" "=&r") ;; res (match_operand:GPR 1 "memory_operand" "+ZC")) (set (match_dup 1) (unspec_volatile:GPR [(match_operand:GPR 2 "reg_or_0_operand" "rJ") ;; mask (match_operand:GPR 3 "reg_or_0_operand" "rJ") ;; inverted_mask (match_operand:GPR 4 "reg_or_0_operand" "rJ") ;; old val (match_operand:GPR 5 "reg_or_0_operand" "rJ") ;; new val (match_operand:SI 6 "const_int_operand")] ;; model UNSPEC_COMPARE_AND_SWAP_ADD)) (clobber (match_scratch:GPR 7 "=&r")) (clobber (match_scratch:GPR 8 "=&r"))] "" { return "1:\\n\\t" "ll.\\t%0,%1\\n\\t" "and\\t%7,%0,%3\\n\\t" "add.w\\t%8,%0,%z5\\n\\t" "and\\t%8,%8,%z2\\n\\t" "or%i8\\t%7,%7,%8\\n\\t" "sc.\\t%7,%1\\n\\t" "beq\\t$zero,%7,1b"; } [(set (attr "length") (const_int 28))]) (define_insn "atomic_cas_value_sub_7_" [(set (match_operand:GPR 0 "register_operand" "=&r") ;; res (match_operand:GPR 1 "memory_operand" "+ZC")) (set (match_dup 1) (unspec_volatile:GPR [(match_operand:GPR 2 "reg_or_0_operand" "rJ") ;; mask (match_operand:GPR 3 "reg_or_0_operand" "rJ") ;; inverted_mask (match_operand:GPR 4 "reg_or_0_operand" "rJ") ;; old val (match_operand:GPR 5 "reg_or_0_operand" "rJ") ;; new val (match_operand:SI 6 "const_int_operand")] ;; model UNSPEC_COMPARE_AND_SWAP_SUB)) (clobber (match_scratch:GPR 7 "=&r")) (clobber (match_scratch:GPR 8 "=&r"))] "" { return "1:\\n\\t" "ll.\\t%0,%1\\n\\t" "and\\t%7,%0,%3\\n\\t" "sub.w\\t%8,%0,%z5\\n\\t" "and\\t%8,%8,%z2\\n\\t" "or%i8\\t%7,%7,%8\\n\\t" "sc.\\t%7,%1\\n\\t" "beq\\t$zero,%7,1b"; } [(set (attr "length") (const_int 28))]) (define_insn "atomic_cas_value_and_7_" [(set (match_operand:GPR 0 "register_operand" "=&r") ;; res (match_operand:GPR 1 "memory_operand" "+ZC")) (set (match_dup 1) (unspec_volatile:GPR [(match_operand:GPR 2 "reg_or_0_operand" "rJ") ;; mask (match_operand:GPR 3 "reg_or_0_operand" "rJ") ;; inverted_mask (match_operand:GPR 4 "reg_or_0_operand" "rJ") ;; old val (match_operand:GPR 5 "reg_or_0_operand" "rJ") ;; new val (match_operand:SI 6 "const_int_operand")] ;; model UNSPEC_COMPARE_AND_SWAP_AND)) (clobber (match_scratch:GPR 7 "=&r")) (clobber (match_scratch:GPR 8 "=&r"))] "" { return "1:\\n\\t" "ll.\\t%0,%1\\n\\t" "and\\t%7,%0,%3\\n\\t" "and\\t%8,%0,%z5\\n\\t" "and\\t%8,%8,%z2\\n\\t" "or%i8\\t%7,%7,%8\\n\\t" "sc.\\t%7,%1\\n\\t" "beq\\t$zero,%7,1b"; } [(set (attr "length") (const_int 28))]) (define_insn "atomic_cas_value_xor_7_" [(set (match_operand:GPR 0 "register_operand" "=&r") ;; res (match_operand:GPR 1 "memory_operand" "+ZC")) (set (match_dup 1) (unspec_volatile:GPR [(match_operand:GPR 2 "reg_or_0_operand" "rJ") ;; mask (match_operand:GPR 3 "reg_or_0_operand" "rJ") ;; inverted_mask (match_operand:GPR 4 "reg_or_0_operand" "rJ") ;; old val (match_operand:GPR 5 "reg_or_0_operand" "rJ") ;; new val (match_operand:SI 6 "const_int_operand")] ;; model UNSPEC_COMPARE_AND_SWAP_XOR)) (clobber (match_scratch:GPR 7 "=&r")) (clobber (match_scratch:GPR 8 "=&r"))] "" { return "1:\\n\\t" "ll.\\t%0,%1\\n\\t" "and\\t%7,%0,%3\\n\\t" "xor\\t%8,%0,%z5\\n\\t" "and\\t%8,%8,%z2\\n\\t" "or%i8\\t%7,%7,%8\\n\\t" "sc.\\t%7,%1\\n\\t" "beq\\t$zero,%7,1b"; } [(set (attr "length") (const_int 28))]) (define_insn "atomic_cas_value_or_7_" [(set (match_operand:GPR 0 "register_operand" "=&r") ;; res (match_operand:GPR 1 "memory_operand" "+ZC")) (set (match_dup 1) (unspec_volatile:GPR [(match_operand:GPR 2 "reg_or_0_operand" "rJ") ;; mask (match_operand:GPR 3 "reg_or_0_operand" "rJ") ;; inverted_mask (match_operand:GPR 4 "reg_or_0_operand" "rJ") ;; old val (match_operand:GPR 5 "reg_or_0_operand" "rJ") ;; new val (match_operand:SI 6 "const_int_operand")] ;; model UNSPEC_COMPARE_AND_SWAP_OR)) (clobber (match_scratch:GPR 7 "=&r")) (clobber (match_scratch:GPR 8 "=&r"))] "" { return "1:\\n\\t" "ll.\\t%0,%1\\n\\t" "and\\t%7,%0,%3\\n\\t" "or\\t%8,%0,%z5\\n\\t" "and\\t%8,%8,%z2\\n\\t" "or%i8\\t%7,%7,%8\\n\\t" "sc.\\t%7,%1\\n\\t" "beq\\t$zero,%7,1b"; } [(set (attr "length") (const_int 28))]) (define_insn "atomic_cas_value_nand_7_" [(set (match_operand:GPR 0 "register_operand" "=&r") ;; res (match_operand:GPR 1 "memory_operand" "+ZC")) (set (match_dup 1) (unspec_volatile:GPR [(match_operand:GPR 2 "reg_or_0_operand" "rJ") ;; mask (match_operand:GPR 3 "reg_or_0_operand" "rJ") ;; inverted_mask (match_operand:GPR 4 "reg_or_0_operand" "rJ") ;; old val (match_operand:GPR 5 "reg_or_0_operand" "rJ") ;; new val (match_operand:SI 6 "const_int_operand")] ;; model UNSPEC_COMPARE_AND_SWAP_NAND)) (clobber (match_scratch:GPR 7 "=&r")) (clobber (match_scratch:GPR 8 "=&r"))] "" { return "1:\\n\\t" "ll.\\t%0,%1\\n\\t" "and\\t%7,%0,%3\\n\\t" "and\\t%8,%0,%z5\\n\\t" "xor\\t%8,%8,%z2\\n\\t" "or%i8\\t%7,%7,%8\\n\\t" "sc.\\t%7,%1\\n\\t" "beq\\t$zero,%7,1b"; } [(set (attr "length") (const_int 28))]) (define_insn "atomic_cas_value_exchange_7_" [(set (match_operand:GPR 0 "register_operand" "=&r") (match_operand:GPR 1 "memory_operand" "+ZC")) (set (match_dup 1) (unspec_volatile:GPR [(match_operand:GPR 2 "reg_or_0_operand" "rJ") (match_operand:GPR 3 "reg_or_0_operand" "rJ") (match_operand:GPR 4 "reg_or_0_operand" "rJ") (match_operand:GPR 5 "reg_or_0_operand" "rJ") (match_operand:SI 6 "const_int_operand")] ;; model UNSPEC_SYNC_EXCHANGE)) (clobber (match_scratch:GPR 7 "=&r"))] "" { return "1:\\n\\t" "ll.\\t%0,%1\\n\\t" "and\\t%7,%0,%z3\\n\\t" "or%i5\\t%7,%7,%5\\n\\t" "sc.\\t%7,%1\\n\\t" "beqz\\t%7,1b\\n\\t"; } [(set (attr "length") (const_int 20))]) (define_expand "atomic_exchange" [(set (match_operand:SHORT 0 "register_operand") (unspec_volatile:SHORT [(match_operand:SHORT 1 "memory_operand") (match_operand:SI 3 "const_int_operand")] ;; model UNSPEC_SYNC_EXCHANGE)) (set (match_dup 1) (match_operand:SHORT 2 "register_operand"))] "" { if (ISA_HAS_LAM_BH) emit_insn (gen_atomic_exchange_short (operands[0], operands[1], operands[2], operands[3])); else { union loongarch_gen_fn_ptrs generator; generator.fn_7 = gen_atomic_cas_value_exchange_7_si; loongarch_expand_atomic_qihi (generator, operands[0], operands[1], const0_rtx, operands[2], operands[3]); } DONE; }) (define_insn "atomic_fetch_add_short" [(set (match_operand:SHORT 0 "register_operand" "=&r") (match_operand:SHORT 1 "memory_operand" "+ZB")) (set (match_dup 1) (unspec_volatile:SHORT [(plus:SHORT (match_dup 1) (match_operand:SHORT 2 "reg_or_0_operand" "rJ")) (match_operand:SI 3 "const_int_operand")] ;; model UNSPEC_SYNC_OLD_OP))] "ISA_HAS_LAM_BH" "amadd%A3.\t%0,%z2,%1" [(set (attr "length") (const_int 4))]) (define_expand "atomic_fetch_add" [(set (match_operand:SHORT 0 "register_operand" "=&r") (match_operand:SHORT 1 "memory_operand" "+ZB")) (set (match_dup 1) (unspec_volatile:SHORT [(plus:SHORT (match_dup 1) (match_operand:SHORT 2 "reg_or_0_operand" "rJ")) (match_operand:SI 3 "const_int_operand")] ;; model UNSPEC_SYNC_OLD_OP))] "" { if (ISA_HAS_LAM_BH) emit_insn (gen_atomic_fetch_add_short (operands[0], operands[1], operands[2], operands[3])); else { union loongarch_gen_fn_ptrs generator; generator.fn_7 = gen_atomic_cas_value_add_7_si; loongarch_expand_atomic_qihi (generator, operands[0], operands[1], operands[1], operands[2], operands[3]); } DONE; }) (define_expand "atomic_fetch_sub" [(set (match_operand:SHORT 0 "register_operand" "=&r") (match_operand:SHORT 1 "memory_operand" "+ZB")) (set (match_dup 1) (unspec_volatile:SHORT [(minus:SHORT (match_dup 1) (match_operand:SHORT 2 "reg_or_0_operand" "rJ")) (match_operand:SI 3 "const_int_operand")] ;; model UNSPEC_SYNC_OLD_OP))] "" { union loongarch_gen_fn_ptrs generator; generator.fn_7 = gen_atomic_cas_value_sub_7_si; loongarch_expand_atomic_qihi (generator, operands[0], operands[1], operands[1], operands[2], operands[3]); DONE; }) (define_expand "atomic_fetch_and" [(set (match_operand:SHORT 0 "register_operand" "=&r") (match_operand:SHORT 1 "memory_operand" "+ZB")) (set (match_dup 1) (unspec_volatile:SHORT [(and:SHORT (match_dup 1) (match_operand:SHORT 2 "reg_or_0_operand" "rJ")) (match_operand:SI 3 "const_int_operand")] ;; model UNSPEC_SYNC_OLD_OP))] "" { union loongarch_gen_fn_ptrs generator; generator.fn_7 = gen_atomic_cas_value_and_7_si; loongarch_expand_atomic_qihi (generator, operands[0], operands[1], operands[1], operands[2], operands[3]); DONE; }) (define_expand "atomic_fetch_xor" [(set (match_operand:SHORT 0 "register_operand" "=&r") (match_operand:SHORT 1 "memory_operand" "+ZB")) (set (match_dup 1) (unspec_volatile:SHORT [(xor:SHORT (match_dup 1) (match_operand:SHORT 2 "reg_or_0_operand" "rJ")) (match_operand:SI 3 "const_int_operand")] ;; model UNSPEC_SYNC_OLD_OP))] "" { union loongarch_gen_fn_ptrs generator; generator.fn_7 = gen_atomic_cas_value_xor_7_si; loongarch_expand_atomic_qihi (generator, operands[0], operands[1], operands[1], operands[2], operands[3]); DONE; }) (define_expand "atomic_fetch_or" [(set (match_operand:SHORT 0 "register_operand" "=&r") (match_operand:SHORT 1 "memory_operand" "+ZB")) (set (match_dup 1) (unspec_volatile:SHORT [(ior:SHORT (match_dup 1) (match_operand:SHORT 2 "reg_or_0_operand" "rJ")) (match_operand:SI 3 "const_int_operand")] ;; model UNSPEC_SYNC_OLD_OP))] "" { union loongarch_gen_fn_ptrs generator; generator.fn_7 = gen_atomic_cas_value_or_7_si; loongarch_expand_atomic_qihi (generator, operands[0], operands[1], operands[1], operands[2], operands[3]); DONE; }) (define_expand "atomic_fetch_nand" [(set (match_operand:SHORT 0 "register_operand" "=&r") (match_operand:SHORT 1 "memory_operand" "+ZB")) (set (match_dup 1) (unspec_volatile:SHORT [(not:SHORT (and:SHORT (match_dup 1) (match_operand:SHORT 2 "reg_or_0_operand" "rJ"))) (match_operand:SI 3 "const_int_operand")] ;; model UNSPEC_SYNC_OLD_OP))] "" { union loongarch_gen_fn_ptrs generator; generator.fn_7 = gen_atomic_cas_value_nand_7_si; loongarch_expand_atomic_qihi (generator, operands[0], operands[1], operands[1], operands[2], operands[3]); DONE; })