/*- * BSD LICENSE * * Copyright(c) 2010-2014 Intel Corporation. All rights reserved. * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * * * Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * * Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in * the documentation and/or other materials provided with the * distribution. * * Neither the name of Intel Corporation nor the names of its * contributors may be used to endorse or promote products derived * from this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ #ifndef _RTE_ATOMIC_H_ #define _RTE_ATOMIC_H_ /** * @file * Atomic Operations * * This file defines a generic API for atomic operations. */ #include #ifdef __DOXYGEN__ /** * General memory barrier. * * Guarantees that the LOAD and STORE operations generated before the * barrier occur before the LOAD and STORE operations generated after. * This function is architecture dependent. */ static inline void rte_mb(void); /** * Write memory barrier. * * Guarantees that the STORE operations generated before the barrier * occur before the STORE operations generated after. * This function is architecture dependent. */ static inline void rte_wmb(void); /** * Read memory barrier. * * Guarantees that the LOAD operations generated before the barrier * occur before the LOAD operations generated after. * This function is architecture dependent. */ static inline void rte_rmb(void); /** * General memory barrier between lcores * * Guarantees that the LOAD and STORE operations that precede the * rte_smp_mb() call are globally visible across the lcores * before the the LOAD and STORE operations that follows it. */ static inline void rte_smp_mb(void); /** * Write memory barrier between lcores * * Guarantees that the STORE operations that precede the * rte_smp_wmb() call are globally visible across the lcores * before the the STORE operations that follows it. */ static inline void rte_smp_wmb(void); /** * Read memory barrier between lcores * * Guarantees that the LOAD operations that precede the * rte_smp_rmb() call are globally visible across the lcores * before the the LOAD operations that follows it. */ static inline void rte_smp_rmb(void); #endif /* __DOXYGEN__ */ /** * Compiler barrier. * * Guarantees that operation reordering does not occur at compile time * for operations directly before and after the barrier. */ #define rte_compiler_barrier() do { \ asm volatile ("" : : : "memory"); \ } while(0) /*------------------------- 16 bit atomic operations -------------------------*/ /** * Atomic compare and set. * * (atomic) equivalent to: * if (*dst == exp) * *dst = src (all 16-bit words) * * @param dst * The destination location into which the value will be written. * @param exp * The expected value. * @param src * The new value. * @return * Non-zero on success; 0 on failure. */ static inline int rte_atomic16_cmpset(volatile uint16_t *dst, uint16_t exp, uint16_t src); #ifdef RTE_FORCE_INTRINSICS static inline int rte_atomic16_cmpset(volatile uint16_t *dst, uint16_t exp, uint16_t src) { return __sync_bool_compare_and_swap(dst, exp, src); } #endif /** * The atomic counter structure. */ typedef struct { volatile int16_t cnt; /**< An internal counter value. */ } rte_atomic16_t; /** * Static initializer for an atomic counter. */ #define RTE_ATOMIC16_INIT(val) { (val) } /** * Initialize an atomic counter. * * @param v * A pointer to the atomic counter. */ static inline void rte_atomic16_init(rte_atomic16_t *v) { v->cnt = 0; } /** * Atomically read a 16-bit value from a counter. * * @param v * A pointer to the atomic counter. * @return * The value of the counter. */ static inline int16_t rte_atomic16_read(const rte_atomic16_t *v) { return v->cnt; } /** * Atomically set a counter to a 16-bit value. * * @param v * A pointer to the atomic counter. * @param new_value * The new value for the counter. */ static inline void rte_atomic16_set(rte_atomic16_t *v, int16_t new_value) { v->cnt = new_value; } /** * Atomically add a 16-bit value to an atomic counter. * * @param v * A pointer to the atomic counter. * @param inc * The value to be added to the counter. */ static inline void rte_atomic16_add(rte_atomic16_t *v, int16_t inc) { __sync_fetch_and_add(&v->cnt, inc); } /** * Atomically subtract a 16-bit value from an atomic counter. * * @param v * A pointer to the atomic counter. * @param dec * The value to be subtracted from the counter. */ static inline void rte_atomic16_sub(rte_atomic16_t *v, int16_t dec) { __sync_fetch_and_sub(&v->cnt, dec); } /** * Atomically increment a counter by one. * * @param v * A pointer to the atomic counter. */ static inline void rte_atomic16_inc(rte_atomic16_t *v); #ifdef RTE_FORCE_INTRINSICS static inline void rte_atomic16_inc(rte_atomic16_t *v) { rte_atomic16_add(v, 1); } #endif /** * Atomically decrement a counter by one. * * @param v * A pointer to the atomic counter. */ static inline void rte_atomic16_dec(rte_atomic16_t *v); #ifdef RTE_FORCE_INTRINSICS static inline void rte_atomic16_dec(rte_atomic16_t *v) { rte_atomic16_sub(v, 1); } #endif /** * Atomically add a 16-bit value to a counter and return the result. * * Atomically adds the 16-bits value (inc) to the atomic counter (v) and * returns the value of v after addition. * * @param v * A pointer to the atomic counter. * @param inc * The value to be added to the counter. * @return * The value of v after the addition. */ static inline int16_t rte_atomic16_add_return(rte_atomic16_t *v, int16_t inc) { return __sync_add_and_fetch(&v->cnt, inc); } /** * Atomically subtract a 16-bit value from a counter and return * the result. * * Atomically subtracts the 16-bit value (inc) from the atomic counter * (v) and returns the value of v after the subtraction. * * @param v * A pointer to the atomic counter. * @param dec * The value to be subtracted from the counter. * @return * The value of v after the subtraction. */ static inline int16_t rte_atomic16_sub_return(rte_atomic16_t *v, int16_t dec) { return __sync_sub_and_fetch(&v->cnt, dec); } /** * Atomically increment a 16-bit counter by one and test. * * Atomically increments the atomic counter (v) by one and returns true if * the result is 0, or false in all other cases. * * @param v * A pointer to the atomic counter. * @return * True if the result after the increment operation is 0; false otherwise. */ static inline int rte_atomic16_inc_and_test(rte_atomic16_t *v); #ifdef RTE_FORCE_INTRINSICS static inline int rte_atomic16_inc_and_test(rte_atomic16_t *v) { return __sync_add_and_fetch(&v->cnt, 1) == 0; } #endif /** * Atomically decrement a 16-bit counter by one and test. * * Atomically decrements the atomic counter (v) by one and returns true if * the result is 0, or false in all other cases. * * @param v * A pointer to the atomic counter. * @return * True if the result after the decrement operation is 0; false otherwise. */ static inline int rte_atomic16_dec_and_test(rte_atomic16_t *v); #ifdef RTE_FORCE_INTRINSICS static inline int rte_atomic16_dec_and_test(rte_atomic16_t *v) { return __sync_sub_and_fetch(&v->cnt, 1) == 0; } #endif /** * Atomically test and set a 16-bit atomic counter. * * If the counter value is already set, return 0 (failed). Otherwise, set * the counter value to 1 and return 1 (success). * * @param v * A pointer to the atomic counter. * @return * 0 if failed; else 1, success. */ static inline int rte_atomic16_test_and_set(rte_atomic16_t *v); #ifdef RTE_FORCE_INTRINSICS static inline int rte_atomic16_test_and_set(rte_atomic16_t *v) { return rte_atomic16_cmpset((volatile uint16_t *)&v->cnt, 0, 1); } #endif /** * Atomically set a 16-bit counter to 0. * * @param v * A pointer to the atomic counter. */ static inline void rte_atomic16_clear(rte_atomic16_t *v) { v->cnt = 0; } /*------------------------- 32 bit atomic operations -------------------------*/ /** * Atomic compare and set. * * (atomic) equivalent to: * if (*dst == exp) * *dst = src (all 32-bit words) * * @param dst * The destination location into which the value will be written. * @param exp * The expected value. * @param src * The new value. * @return * Non-zero on success; 0 on failure. */ static inline int rte_atomic32_cmpset(volatile uint32_t *dst, uint32_t exp, uint32_t src); #ifdef RTE_FORCE_INTRINSICS static inline int rte_atomic32_cmpset(volatile uint32_t *dst, uint32_t exp, uint32_t src) { return __sync_bool_compare_and_swap(dst, exp, src); } #endif /** * The atomic counter structure. */ typedef struct { volatile int32_t cnt; /**< An internal counter value. */ } rte_atomic32_t; /** * Static initializer for an atomic counter. */ #define RTE_ATOMIC32_INIT(val) { (val) } /** * Initialize an atomic counter. * * @param v * A pointer to the atomic counter. */ static inline void rte_atomic32_init(rte_atomic32_t *v) { v->cnt = 0; } /** * Atomically read a 32-bit value from a counter. * * @param v * A pointer to the atomic counter. * @return * The value of the counter. */ static inline int32_t rte_atomic32_read(const rte_atomic32_t *v) { return v->cnt; } /** * Atomically set a counter to a 32-bit value. * * @param v * A pointer to the atomic counter. * @param new_value * The new value for the counter. */ static inline void rte_atomic32_set(rte_atomic32_t *v, int32_t new_value) { v->cnt = new_value; } /** * Atomically add a 32-bit value to an atomic counter. * * @param v * A pointer to the atomic counter. * @param inc * The value to be added to the counter. */ static inline void rte_atomic32_add(rte_atomic32_t *v, int32_t inc) { __sync_fetch_and_add(&v->cnt, inc); } /** * Atomically subtract a 32-bit value from an atomic counter. * * @param v * A pointer to the atomic counter. * @param dec * The value to be subtracted from the counter. */ static inline void rte_atomic32_sub(rte_atomic32_t *v, int32_t dec) { __sync_fetch_and_sub(&v->cnt, dec); } /** * Atomically increment a counter by one. * * @param v * A pointer to the atomic counter. */ static inline void rte_atomic32_inc(rte_atomic32_t *v); #ifdef RTE_FORCE_INTRINSICS static inline void rte_atomic32_inc(rte_atomic32_t *v) { rte_atomic32_add(v, 1); } #endif /** * Atomically decrement a counter by one. * * @param v * A pointer to the atomic counter. */ static inline void rte_atomic32_dec(rte_atomic32_t *v); #ifdef RTE_FORCE_INTRINSICS static inline void rte_atomic32_dec(rte_atomic32_t *v) { rte_atomic32_sub(v,1); } #endif /** * Atomically add a 32-bit value to a counter and return the result. * * Atomically adds the 32-bits value (inc) to the atomic counter (v) and * returns the value of v after addition. * * @param v * A pointer to the atomic counter. * @param inc * The value to be added to the counter. * @return * The value of v after the addition. */ static inline int32_t rte_atomic32_add_return(rte_atomic32_t *v, int32_t inc) { return __sync_add_and_fetch(&v->cnt, inc); } /** * Atomically subtract a 32-bit value from a counter and return * the result. * * Atomically subtracts the 32-bit value (inc) from the atomic counter * (v) and returns the value of v after the subtraction. * * @param v * A pointer to the atomic counter. * @param dec * The value to be subtracted from the counter. * @return * The value of v after the subtraction. */ static inline int32_t rte_atomic32_sub_return(rte_atomic32_t *v, int32_t dec) { return __sync_sub_and_fetch(&v->cnt, dec); } /** * Atomically increment a 32-bit counter by one and test. * * Atomically increments the atomic counter (v) by one and returns true if * the result is 0, or false in all other cases. * * @param v * A pointer to the atomic counter. * @return * True if the result after the increment operation is 0; false otherwise. */ static inline int rte_atomic32_inc_and_test(rte_atomic32_t *v); #ifdef RTE_FORCE_INTRINSICS static inline int rte_atomic32_inc_and_test(rte_atomic32_t *v) { return __sync_add_and_fetch(&v->cnt, 1) == 0; } #endif /** * Atomically decrement a 32-bit counter by one and test. * * Atomically decrements the atomic counter (v) by one and returns true if * the result is 0, or false in all other cases. * * @param v * A pointer to the atomic counter. * @return * True if the result after the decrement operation is 0; false otherwise. */ static inline int rte_atomic32_dec_and_test(rte_atomic32_t *v); #ifdef RTE_FORCE_INTRINSICS static inline int rte_atomic32_dec_and_test(rte_atomic32_t *v) { return __sync_sub_and_fetch(&v->cnt, 1) == 0; } #endif /** * Atomically test and set a 32-bit atomic counter. * * If the counter value is already set, return 0 (failed). Otherwise, set * the counter value to 1 and return 1 (success). * * @param v * A pointer to the atomic counter. * @return * 0 if failed; else 1, success. */ static inline int rte_atomic32_test_and_set(rte_atomic32_t *v); #ifdef RTE_FORCE_INTRINSICS static inline int rte_atomic32_test_and_set(rte_atomic32_t *v) { return rte_atomic32_cmpset((volatile uint32_t *)&v->cnt, 0, 1); } #endif /** * Atomically set a 32-bit counter to 0. * * @param v * A pointer to the atomic counter. */ static inline void rte_atomic32_clear(rte_atomic32_t *v) { v->cnt = 0; } /*------------------------- 64 bit atomic operations -------------------------*/ /** * An atomic compare and set function used by the mutex functions. * (atomic) equivalent to: * if (*dst == exp) * *dst = src (all 64-bit words) * * @param dst * The destination into which the value will be written. * @param exp * The expected value. * @param src * The new value. * @return * Non-zero on success; 0 on failure. */ static inline int rte_atomic64_cmpset(volatile uint64_t *dst, uint64_t exp, uint64_t src); #ifdef RTE_FORCE_INTRINSICS static inline int rte_atomic64_cmpset(volatile uint64_t *dst, uint64_t exp, uint64_t src) { return __sync_bool_compare_and_swap(dst, exp, src); } #endif /** * The atomic counter structure. */ typedef struct { volatile int64_t cnt; /**< Internal counter value. */ } rte_atomic64_t; /** * Static initializer for an atomic counter. */ #define RTE_ATOMIC64_INIT(val) { (val) } /** * Initialize the atomic counter. * * @param v * A pointer to the atomic counter. */ static inline void rte_atomic64_init(rte_atomic64_t *v); #ifdef RTE_FORCE_INTRINSICS static inline void rte_atomic64_init(rte_atomic64_t *v) { #ifdef __LP64__ v->cnt = 0; #else int success = 0; uint64_t tmp; while (success == 0) { tmp = v->cnt; success = rte_atomic64_cmpset((volatile uint64_t *)&v->cnt, tmp, 0); } #endif } #endif /** * Atomically read a 64-bit counter. * * @param v * A pointer to the atomic counter. * @return * The value of the counter. */ static inline int64_t rte_atomic64_read(rte_atomic64_t *v); #ifdef RTE_FORCE_INTRINSICS static inline int64_t rte_atomic64_read(rte_atomic64_t *v) { #ifdef __LP64__ return v->cnt; #else int success = 0; uint64_t tmp; while (success == 0) { tmp = v->cnt; /* replace the value by itself */ success = rte_atomic64_cmpset((volatile uint64_t *)&v->cnt, tmp, tmp); } return tmp; #endif } #endif /** * Atomically set a 64-bit counter. * * @param v * A pointer to the atomic counter. * @param new_value * The new value of the counter. */ static inline void rte_atomic64_set(rte_atomic64_t *v, int64_t new_value); #ifdef RTE_FORCE_INTRINSICS static inline void rte_atomic64_set(rte_atomic64_t *v, int64_t new_value) { #ifdef __LP64__ v->cnt = new_value; #else int success = 0; uint64_t tmp; while (success == 0) { tmp = v->cnt; success = rte_atomic64_cmpset((volatile uint64_t *)&v->cnt, tmp, new_value); } #endif } #endif /** * Atomically add a 64-bit value to a counter. * * @param v * A pointer to the atomic counter. * @param inc * The value to be added to the counter. */ static inline void rte_atomic64_add(rte_atomic64_t *v, int64_t inc); #ifdef RTE_FORCE_INTRINSICS static inline void rte_atomic64_add(rte_atomic64_t *v, int64_t inc) { __sync_fetch_and_add(&v->cnt, inc); } #endif /** * Atomically subtract a 64-bit value from a counter. * * @param v * A pointer to the atomic counter. * @param dec * The value to be subtracted from the counter. */ static inline void rte_atomic64_sub(rte_atomic64_t *v, int64_t dec); #ifdef RTE_FORCE_INTRINSICS static inline void rte_atomic64_sub(rte_atomic64_t *v, int64_t dec) { __sync_fetch_and_sub(&v->cnt, dec); } #endif /** * Atomically increment a 64-bit counter by one and test. * * @param v * A pointer to the atomic counter. */ static inline void rte_atomic64_inc(rte_atomic64_t *v); #ifdef RTE_FORCE_INTRINSICS static inline void rte_atomic64_inc(rte_atomic64_t *v) { rte_atomic64_add(v, 1); } #endif /** * Atomically decrement a 64-bit counter by one and test. * * @param v * A pointer to the atomic counter. */ static inline void rte_atomic64_dec(rte_atomic64_t *v); #ifdef RTE_FORCE_INTRINSICS static inline void rte_atomic64_dec(rte_atomic64_t *v) { rte_atomic64_sub(v, 1); } #endif /** * Add a 64-bit value to an atomic counter and return the result. * * Atomically adds the 64-bit value (inc) to the atomic counter (v) and * returns the value of v after the addition. * * @param v * A pointer to the atomic counter. * @param inc * The value to be added to the counter. * @return * The value of v after the addition. */ static inline int64_t rte_atomic64_add_return(rte_atomic64_t *v, int64_t inc); #ifdef RTE_FORCE_INTRINSICS static inline int64_t rte_atomic64_add_return(rte_atomic64_t *v, int64_t inc) { return __sync_add_and_fetch(&v->cnt, inc); } #endif /** * Subtract a 64-bit value from an atomic counter and return the result. * * Atomically subtracts the 64-bit value (dec) from the atomic counter (v) * and returns the value of v after the subtraction. * * @param v * A pointer to the atomic counter. * @param dec * The value to be subtracted from the counter. * @return * The value of v after the subtraction. */ static inline int64_t rte_atomic64_sub_return(rte_atomic64_t *v, int64_t dec); #ifdef RTE_FORCE_INTRINSICS static inline int64_t rte_atomic64_sub_return(rte_atomic64_t *v, int64_t dec) { return __sync_sub_and_fetch(&v->cnt, dec); } #endif /** * Atomically increment a 64-bit counter by one and test. * * Atomically increments the atomic counter (v) by one and returns * true if the result is 0, or false in all other cases. * * @param v * A pointer to the atomic counter. * @return * True if the result after the addition is 0; false otherwise. */ static inline int rte_atomic64_inc_and_test(rte_atomic64_t *v); #ifdef RTE_FORCE_INTRINSICS static inline int rte_atomic64_inc_and_test(rte_atomic64_t *v) { return rte_atomic64_add_return(v, 1) == 0; } #endif /** * Atomically decrement a 64-bit counter by one and test. * * Atomically decrements the atomic counter (v) by one and returns true if * the result is 0, or false in all other cases. * * @param v * A pointer to the atomic counter. * @return * True if the result after subtraction is 0; false otherwise. */ static inline int rte_atomic64_dec_and_test(rte_atomic64_t *v); #ifdef RTE_FORCE_INTRINSICS static inline int rte_atomic64_dec_and_test(rte_atomic64_t *v) { return rte_atomic64_sub_return(v, 1) == 0; } #endif /** * Atomically test and set a 64-bit atomic counter. * * If the counter value is already set, return 0 (failed). Otherwise, set * the counter value to 1 and return 1 (success). * * @param v * A pointer to the atomic counter. * @return * 0 if failed; else 1, success. */ static inline int rte_atomic64_test_and_set(rte_atomic64_t *v); #ifdef RTE_FORCE_INTRINSICS static inline int rte_atomic64_test_and_set(rte_atomic64_t *v) { return rte_atomic64_cmpset((volatile uint64_t *)&v->cnt, 0, 1); } #endif /** * Atomically set a 64-bit counter to 0. * * @param v * A pointer to the atomic counter. */ static inline void rte_atomic64_clear(rte_atomic64_t *v); #ifdef RTE_FORCE_INTRINSICS static inline void rte_atomic64_clear(rte_atomic64_t *v) { rte_atomic64_set(v, 0); } #endif #endif /* _RTE_ATOMIC_H_ */