| /* atomic operations */ |
| |
| /* |
| * Copyright (c) 1997-2015, Wind River Systems, Inc. |
| * Copyright (c) 2023 Nordic Semiconductor ASA |
| * |
| * SPDX-License-Identifier: Apache-2.0 |
| */ |
| |
| #ifndef ZEPHYR_INCLUDE_SYS_ATOMIC_BUILTIN_H_ |
| #define ZEPHYR_INCLUDE_SYS_ATOMIC_BUILTIN_H_ |
| |
| #include <stdbool.h> |
| #include <zephyr/sys/atomic_types.h> |
| |
| #ifdef __cplusplus |
| extern "C" { |
| #endif |
| |
| /* Included from <atomic.h> */ |
| |
| static inline bool atomic_cas(atomic_t *target, atomic_val_t old_value, |
| atomic_val_t new_value) |
| { |
| return __atomic_compare_exchange_n(target, &old_value, new_value, |
| 0, __ATOMIC_SEQ_CST, |
| __ATOMIC_SEQ_CST); |
| } |
| |
| static inline bool atomic_ptr_cas(atomic_ptr_t *target, atomic_ptr_val_t old_value, |
| atomic_ptr_val_t new_value) |
| { |
| return __atomic_compare_exchange_n(target, &old_value, new_value, |
| 0, __ATOMIC_SEQ_CST, |
| __ATOMIC_SEQ_CST); |
| } |
| |
| static inline atomic_val_t atomic_add(atomic_t *target, atomic_val_t value) |
| { |
| return __atomic_fetch_add(target, value, __ATOMIC_SEQ_CST); |
| } |
| |
| static inline atomic_val_t atomic_sub(atomic_t *target, atomic_val_t value) |
| { |
| return __atomic_fetch_sub(target, value, __ATOMIC_SEQ_CST); |
| } |
| |
| static inline atomic_val_t atomic_inc(atomic_t *target) |
| { |
| return atomic_add(target, 1); |
| } |
| |
| static inline atomic_val_t atomic_dec(atomic_t *target) |
| { |
| return atomic_sub(target, 1); |
| } |
| |
| static inline atomic_val_t atomic_get(const atomic_t *target) |
| { |
| return __atomic_load_n(target, __ATOMIC_SEQ_CST); |
| } |
| |
| static inline atomic_ptr_val_t atomic_ptr_get(const atomic_ptr_t *target) |
| { |
| return __atomic_load_n(target, __ATOMIC_SEQ_CST); |
| } |
| |
| static inline atomic_val_t atomic_set(atomic_t *target, atomic_val_t value) |
| { |
| /* This builtin, as described by Intel, is not a traditional |
| * test-and-set operation, but rather an atomic exchange operation. It |
| * writes value into *ptr, and returns the previous contents of *ptr. |
| */ |
| return __atomic_exchange_n(target, value, __ATOMIC_SEQ_CST); |
| } |
| |
| static inline atomic_ptr_val_t atomic_ptr_set(atomic_ptr_t *target, atomic_ptr_val_t value) |
| { |
| return __atomic_exchange_n(target, value, __ATOMIC_SEQ_CST); |
| } |
| |
| static inline atomic_val_t atomic_clear(atomic_t *target) |
| { |
| return atomic_set(target, 0); |
| } |
| |
| static inline atomic_ptr_val_t atomic_ptr_clear(atomic_ptr_t *target) |
| { |
| return atomic_ptr_set(target, NULL); |
| } |
| |
| static inline atomic_val_t atomic_or(atomic_t *target, atomic_val_t value) |
| { |
| return __atomic_fetch_or(target, value, __ATOMIC_SEQ_CST); |
| } |
| |
| static inline atomic_val_t atomic_xor(atomic_t *target, atomic_val_t value) |
| { |
| return __atomic_fetch_xor(target, value, __ATOMIC_SEQ_CST); |
| } |
| |
| static inline atomic_val_t atomic_and(atomic_t *target, atomic_val_t value) |
| { |
| return __atomic_fetch_and(target, value, __ATOMIC_SEQ_CST); |
| } |
| |
| static inline atomic_val_t atomic_nand(atomic_t *target, atomic_val_t value) |
| { |
| return __atomic_fetch_nand(target, value, __ATOMIC_SEQ_CST); |
| } |
| |
| #ifdef __cplusplus |
| } |
| #endif |
| |
| #endif /* ZEPHYR_INCLUDE_SYS_ATOMIC_BUILTIN_H_ */ |