/* Linux-specific atomic operations for NDS32 Linux. Copyright (C) 2012-2018 Free Software Foundation, Inc. This file is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 3, or (at your option) any later version. This file is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. Under Section 7 of GPL version 3, you are granted additional permissions described in the GCC Runtime Library Exception, version 3.1, as published by the Free Software Foundation. You should have received a copy of the GNU General Public License and a copy of the GCC Runtime Library Exception along with this program; see the files COPYING3 and COPYING.RUNTIME respectively. If not, see . */ /* We implement byte, short and int versions of each atomic operation using the kernel helper defined below. There is no support for 64-bit operations yet. */ /* This function copy form NDS32 Linux-kernal. */ static inline int __kernel_cmpxchg (int oldval, int newval, int *mem) { int temp1, temp2, temp3, offset; asm volatile ("msync\tall\n" "movi\t%0, #0\n" "1:\n" "\tllw\t%1, [%4+%0]\n" "\tsub\t%3, %1, %6\n" "\tcmovz\t%2, %5, %3\n" "\tcmovn\t%2, %1, %3\n" "\tscw\t%2, [%4+%0]\n" "\tbeqz\t%2, 1b\n" : "=&r" (offset), "=&r" (temp3), "=&r" (temp2), "=&r" (temp1) : "r" (mem), "r" (newval), "r" (oldval) : "memory"); return temp1; } #define HIDDEN __attribute__ ((visibility ("hidden"))) #ifdef __NDS32_EL__ #define INVERT_MASK_1 0 #define INVERT_MASK_2 0 #else #define INVERT_MASK_1 24 #define INVERT_MASK_2 16 #endif #define MASK_1 0xffu #define MASK_2 0xffffu #define FETCH_AND_OP_WORD(OP, PFX_OP, INF_OP) \ int HIDDEN \ __sync_fetch_and_##OP##_4 (int *ptr, int val) \ { \ int failure, tmp; \ \ do { \ tmp = __atomic_load_n (ptr, __ATOMIC_SEQ_CST); \ failure = __kernel_cmpxchg (tmp, PFX_OP (tmp INF_OP val), ptr); \ } while (failure != 0); \ \ return tmp; \ } FETCH_AND_OP_WORD (add, , +) FETCH_AND_OP_WORD (sub, , -) FETCH_AND_OP_WORD (or, , |) FETCH_AND_OP_WORD (and, , &) FETCH_AND_OP_WORD (xor, , ^) FETCH_AND_OP_WORD (nand, ~, &) #define NAME_oldval(OP, WIDTH) __sync_fetch_and_##OP##_##WIDTH #define NAME_newval(OP, WIDTH) __sync_##OP##_and_fetch_##WIDTH /* Implement both __sync__and_fetch and __sync_fetch_and_ for subword-sized quantities. */ #define SUBWORD_SYNC_OP(OP, PFX_OP, INF_OP, TYPE, WIDTH, RETURN) \ TYPE HIDDEN \ NAME##_##RETURN (OP, WIDTH) (TYPE *ptr, TYPE val) \ { \ int *wordptr = (int *) ((unsigned long) ptr & ~3); \ unsigned int mask, shift, oldval, newval; \ int failure; \ \ shift = (((unsigned long) ptr & 3) << 3) ^ INVERT_MASK_##WIDTH; \ mask = MASK_##WIDTH << shift; \ \ do { \ oldval = __atomic_load_n (wordptr, __ATOMIC_SEQ_CST); \ newval = ((PFX_OP (((oldval & mask) >> shift) \ INF_OP (unsigned int) val)) << shift) & mask; \ newval |= oldval & ~mask; \ failure = __kernel_cmpxchg (oldval, newval, wordptr); \ } while (failure != 0); \ \ return (RETURN & mask) >> shift; \ } SUBWORD_SYNC_OP (add, , +, unsigned short, 2, oldval) SUBWORD_SYNC_OP (sub, , -, unsigned short, 2, oldval) SUBWORD_SYNC_OP (or, , |, unsigned short, 2, oldval) SUBWORD_SYNC_OP (and, , &, unsigned short, 2, oldval) SUBWORD_SYNC_OP (xor, , ^, unsigned short, 2, oldval) SUBWORD_SYNC_OP (nand, ~, &, unsigned short, 2, oldval) SUBWORD_SYNC_OP (add, , +, unsigned char, 1, oldval) SUBWORD_SYNC_OP (sub, , -, unsigned char, 1, oldval) SUBWORD_SYNC_OP (or, , |, unsigned char, 1, oldval) SUBWORD_SYNC_OP (and, , &, unsigned char, 1, oldval) SUBWORD_SYNC_OP (xor, , ^, unsigned char, 1, oldval) SUBWORD_SYNC_OP (nand, ~, &, unsigned char, 1, oldval) #define OP_AND_FETCH_WORD(OP, PFX_OP, INF_OP) \ int HIDDEN \ __sync_##OP##_and_fetch_4 (int *ptr, int val) \ { \ int tmp, failure; \ \ do { \ tmp = __atomic_load_n (ptr, __ATOMIC_SEQ_CST); \ failure = __kernel_cmpxchg (tmp, PFX_OP (tmp INF_OP val), ptr); \ } while (failure != 0); \ \ return PFX_OP (tmp INF_OP val); \ } OP_AND_FETCH_WORD (add, , +) OP_AND_FETCH_WORD (sub, , -) OP_AND_FETCH_WORD (or, , |) OP_AND_FETCH_WORD (and, , &) OP_AND_FETCH_WORD (xor, , ^) OP_AND_FETCH_WORD (nand, ~, &) SUBWORD_SYNC_OP (add, , +, unsigned short, 2, newval) SUBWORD_SYNC_OP (sub, , -, unsigned short, 2, newval) SUBWORD_SYNC_OP (or, , |, unsigned short, 2, newval) SUBWORD_SYNC_OP (and, , &, unsigned short, 2, newval) SUBWORD_SYNC_OP (xor, , ^, unsigned short, 2, newval) SUBWORD_SYNC_OP (nand, ~, &, unsigned short, 2, newval) SUBWORD_SYNC_OP (add, , +, unsigned char, 1, newval) SUBWORD_SYNC_OP (sub, , -, unsigned char, 1, newval) SUBWORD_SYNC_OP (or, , |, unsigned char, 1, newval) SUBWORD_SYNC_OP (and, , &, unsigned char, 1, newval) SUBWORD_SYNC_OP (xor, , ^, unsigned char, 1, newval) SUBWORD_SYNC_OP (nand, ~, &, unsigned char, 1, newval) int HIDDEN __sync_val_compare_and_swap_4 (int *ptr, int oldval, int newval) { int actual_oldval, fail; while (1) { actual_oldval = __atomic_load_n (ptr, __ATOMIC_SEQ_CST); if (oldval != actual_oldval) return actual_oldval; fail = __kernel_cmpxchg (actual_oldval, newval, ptr); if (!fail) return oldval; } } #define SUBWORD_VAL_CAS(TYPE, WIDTH) \ TYPE HIDDEN \ __sync_val_compare_and_swap_##WIDTH (TYPE *ptr, TYPE oldval, \ TYPE newval) \ { \ int *wordptr = (int *)((unsigned long) ptr & ~3), fail; \ unsigned int mask, shift, actual_oldval, actual_newval; \ \ shift = (((unsigned long) ptr & 3) << 3) ^ INVERT_MASK_##WIDTH; \ mask = MASK_##WIDTH << shift; \ \ while (1) \ { \ actual_oldval = __atomic_load_n (wordptr, __ATOMIC_SEQ_CST); \ \ if (((actual_oldval & mask) >> shift) != (unsigned int) oldval) \ return (actual_oldval & mask) >> shift; \ \ actual_newval = (actual_oldval & ~mask) \ | (((unsigned int) newval << shift) & mask); \ \ fail = __kernel_cmpxchg (actual_oldval, actual_newval, \ wordptr); \ \ if (!fail) \ return oldval; \ } \ } SUBWORD_VAL_CAS (unsigned short, 2) SUBWORD_VAL_CAS (unsigned char, 1) typedef unsigned char bool; bool HIDDEN __sync_bool_compare_and_swap_4 (int *ptr, int oldval, int newval) { int failure = __kernel_cmpxchg (oldval, newval, ptr); return (failure == 0); } #define SUBWORD_BOOL_CAS(TYPE, WIDTH) \ bool HIDDEN \ __sync_bool_compare_and_swap_##WIDTH (TYPE *ptr, TYPE oldval, \ TYPE newval) \ { \ TYPE actual_oldval \ = __sync_val_compare_and_swap_##WIDTH (ptr, oldval, newval); \ return (oldval == actual_oldval); \ } SUBWORD_BOOL_CAS (unsigned short, 2) SUBWORD_BOOL_CAS (unsigned char, 1) int HIDDEN __sync_lock_test_and_set_4 (int *ptr, int val) { int failure, oldval; do { oldval = __atomic_load_n (ptr, __ATOMIC_SEQ_CST); failure = __kernel_cmpxchg (oldval, val, ptr); } while (failure != 0); return oldval; } #define SUBWORD_TEST_AND_SET(TYPE, WIDTH) \ TYPE HIDDEN \ __sync_lock_test_and_set_##WIDTH (TYPE *ptr, TYPE val) \ { \ int failure; \ unsigned int oldval, newval, shift, mask; \ int *wordptr = (int *) ((unsigned long) ptr & ~3); \ \ shift = (((unsigned long) ptr & 3) << 3) ^ INVERT_MASK_##WIDTH; \ mask = MASK_##WIDTH << shift; \ \ do { \ oldval = __atomic_load_n (wordptr, __ATOMIC_SEQ_CST); \ newval = (oldval & ~mask) \ | (((unsigned int) val << shift) & mask); \ failure = __kernel_cmpxchg (oldval, newval, wordptr); \ } while (failure != 0); \ \ return (oldval & mask) >> shift; \ } SUBWORD_TEST_AND_SET (unsigned short, 2) SUBWORD_TEST_AND_SET (unsigned char, 1) #define SYNC_LOCK_RELEASE(TYPE, WIDTH) \ void HIDDEN \ __sync_lock_release_##WIDTH (TYPE *ptr) \ { \ /* All writes before this point must be seen before we release \ the lock itself. */ \ __builtin_nds32_msync_all (); \ *ptr = 0; \ } SYNC_LOCK_RELEASE (int, 4) SYNC_LOCK_RELEASE (short, 2) SYNC_LOCK_RELEASE (char, 1)