diff options
Diffstat (limited to 'src/framework/common/include/generic/dmm_atomic.h')
-rw-r--r-- | src/framework/common/include/generic/dmm_atomic.h | 177 |
1 files changed, 177 insertions, 0 deletions
diff --git a/src/framework/common/include/generic/dmm_atomic.h b/src/framework/common/include/generic/dmm_atomic.h new file mode 100644 index 0000000..637306b --- /dev/null +++ b/src/framework/common/include/generic/dmm_atomic.h @@ -0,0 +1,177 @@ +/* +* +* Copyright (c) 2018 Huawei Technologies Co.,Ltd. +* Licensed under the Apache License, Version 2.0 (the "License"); +* you may not use this file except in compliance with the License. +* You may obtain a copy of the License at: +* +* http://www.apache.org/licenses/LICENSE-2.0 +* +* Unless required by applicable law or agreed to in writing, software +* distributed under the License is distributed on an "AS IS" BASIS, +* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +* See the License for the specific language governing permissions and +* limitations under the License. +*/ +#ifndef _DMM_ATOMIC_H_ +#define _DMM_ATOMIC_H_ + +/* atomic 32 bit operation */ + +typedef struct +{ + volatile int cnt; +} dmm_atomic_t; + +inline static int +dmm_atomic_get (dmm_atomic_t * a) +{ + return a->cnt; +} + +inline static int +dmm_atomic_add (dmm_atomic_t * a, int n) +{ + return __sync_fetch_and_add (&a->cnt, n); +} + +inline static int +dmm_atomic_sub (dmm_atomic_t * a, int n) +{ + return __sync_fetch_and_sub (&a->cnt, n); +} + +inline static int +dmm_atomic_and (dmm_atomic_t * a, int n) +{ + return __sync_fetch_and_and (&a->cnt, n); +} + +inline static int +dmm_atomic_or (dmm_atomic_t * a, int n) +{ + return __sync_fetch_and_or (&a->cnt, n); +} + +inline static int +dmm_atomic_xor (dmm_atomic_t * a, int n) +{ + return __sync_fetch_and_xor (&a->cnt, n); +} + +inline static int +dmm_atomic_swap (dmm_atomic_t * a, int o, int n) +{ + return __sync_val_compare_and_swap (&a->cnt, o, n); +} + +inline static int +dmm_atomic_add_return (dmm_atomic_t * a, int n) +{ + return __sync_add_and_fetch (&a->cnt, n); +} + +inline static int +dmm_atomic_sub_return (dmm_atomic_t * a, int n) +{ + return __sync_sub_and_fetch (&a->cnt, n); +} + +inline static int +dmm_atomic_and_return (dmm_atomic_t * a, int n) +{ + return __sync_and_and_fetch (&a->cnt, n); +} + +inline static int +dmm_atomic_or_return (dmm_atomic_t * a, int n) +{ + return __sync_or_and_fetch (&a->cnt, n); +} + +inline static int +dmm_atomic_xor_return (dmm_atomic_t * a, int n) +{ + return __sync_xor_and_fetch (&a->cnt, n); +} + +/* atomit 64bit operation */ + +typedef struct +{ + volatile long long int cnt; +} dmm_atomic64_t; + +inline static long long int +dmm_atomic64_get (dmm_atomic64_t * a) +{ + return a->cnt; +} + +inline static long long int +dmm_atomic64_add (dmm_atomic64_t * a, int n) +{ + return __sync_fetch_and_add (&a->cnt, n); +} + +inline static long long int +dmm_atomic64_sub (dmm_atomic64_t * a, int n) +{ + return __sync_fetch_and_sub (&a->cnt, n); +} + +inline static long long int +dmm_atomic64_and (dmm_atomic64_t * a, int n) +{ + return __sync_fetch_and_and (&a->cnt, n); +} + +inline static long long int +dmm_atomic64_or (dmm_atomic64_t * a, int n) +{ + return __sync_fetch_and_or (&a->cnt, n); +} + +inline static long long int +dmm_atomic64_xor (dmm_atomic64_t * a, int n) +{ + return __sync_fetch_and_xor (&a->cnt, n); +} + +inline static long long int +dmm_atomic64_swap (dmm_atomic_t * a, int o, int n) +{ + return __sync_val_compare_and_swap (&a->cnt, o, n); +} + +inline static long long int +dmm_atomic64_add_return (dmm_atomic64_t * a, int n) +{ + return __sync_add_and_fetch (&a->cnt, n); +} + +inline static long long int +dmm_atomic64_sub_return (dmm_atomic64_t * a, int n) +{ + return __sync_sub_and_fetch (&a->cnt, n); +} + +inline static long long int +dmm_atomic64_and_return (dmm_atomic64_t * a, int n) +{ + return __sync_and_and_fetch (&a->cnt, n); +} + +inline static long long int +dmm_atomic64_or_return (dmm_atomic64_t * a, int n) +{ + return __sync_or_and_fetch (&a->cnt, n); +} + +inline static long long int +dmm_atomic64_xor_return (dmm_atomic64_t * a, int n) +{ + return __sync_xor_and_fetch (&a->cnt, n); +} + +#endif /* #ifndef _DMM_ATOMIC_H_ */ |