#ifndef __CR_ATOMIC_H__ #define __CR_ATOMIC_H__ typedef uint32_t atomic_t; /* Copied from the Linux kernel header arch/arm/include/asm/atomic.h */ #define smp_mb() __asm__ __volatile__ ("dmb" : : : "memory") #define atomic_set(mem,v) (*(mem) = (v)) #define atomic_get(v) (*(volatile uint32_t *)v) static inline unsigned int atomic_add_return(int i, atomic_t *v) { unsigned long tmp; unsigned int result; smp_mb(); __asm__ __volatile__("@ atomic_add_return\n" "1: ldrex %0, [%3]\n" " add %0, %0, %4\n" " strex %1, %0, [%3]\n" " teq %1, #0\n" " bne 1b\n" : "=&r" (result), "=&r" (tmp), "+Qo" (v) : "r" (&v), "Ir" (i) : "cc"); smp_mb(); return result; } static inline unsigned int atomic_sub_return(int i, atomic_t *v) { unsigned long tmp; int result; smp_mb(); __asm__ __volatile__("@ atomic_sub_return\n" "1: ldrex %0, [%3]\n" " sub %0, %0, %4\n" " strex %1, %0, [%3]\n" " teq %1, #0\n" " bne 1b\n" : "=&r" (result), "=&r" (tmp), "+Qo" (v) : "r" (&v), "Ir" (i) : "cc"); smp_mb(); return result; } static inline unsigned int atomic_inc(atomic_t *v) { return atomic_add_return(1, v) - 1; } static inline unsigned int atomic_dec(atomic_t *v) { return atomic_sub_return(1, v) + 1; } /* true if the result is 0, or false for all other cases. */ #define atomic_dec_and_test(v) (atomic_sub_return(1, v) == 0) #endif /* __CR_ATOMIC_H__ */