| /* |
| * Copyright (c) 2010-2016 Wind River Systems, Inc. |
| * |
| * SPDX-License-Identifier: Apache-2.0 |
| */ |
| |
| /** |
| * @brief fixed-size stack object |
| */ |
| |
| #include <zephyr/kernel.h> |
| #include <zephyr/kernel_structs.h> |
| |
| #include <zephyr/toolchain.h> |
| #include <ksched.h> |
| #include <wait_q.h> |
| #include <zephyr/sys/check.h> |
| #include <zephyr/init.h> |
| #include <zephyr/internal/syscall_handler.h> |
| #include <kernel_internal.h> |
| |
| #ifdef CONFIG_OBJ_CORE_STACK |
| static struct k_obj_type obj_type_stack; |
| #endif /* CONFIG_OBJ_CORE_STACK */ |
| |
| void k_stack_init(struct k_stack *stack, stack_data_t *buffer, |
| uint32_t num_entries) |
| { |
| z_waitq_init(&stack->wait_q); |
| stack->lock = (struct k_spinlock) {}; |
| stack->next = stack->base = buffer; |
| stack->top = stack->base + num_entries; |
| |
| SYS_PORT_TRACING_OBJ_INIT(k_stack, stack); |
| k_object_init(stack); |
| |
| #ifdef CONFIG_OBJ_CORE_STACK |
| k_obj_core_init_and_link(K_OBJ_CORE(stack), &obj_type_stack); |
| #endif /* CONFIG_OBJ_CORE_STACK */ |
| } |
| |
| int32_t z_impl_k_stack_alloc_init(struct k_stack *stack, uint32_t num_entries) |
| { |
| void *buffer; |
| int32_t ret; |
| |
| SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_stack, alloc_init, stack); |
| |
| buffer = z_thread_malloc(num_entries * sizeof(stack_data_t)); |
| if (buffer != NULL) { |
| k_stack_init(stack, buffer, num_entries); |
| stack->flags = K_STACK_FLAG_ALLOC; |
| ret = (int32_t)0; |
| } else { |
| ret = -ENOMEM; |
| } |
| |
| SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_stack, alloc_init, stack, ret); |
| |
| return ret; |
| } |
| |
| #ifdef CONFIG_USERSPACE |
| static inline int32_t z_vrfy_k_stack_alloc_init(struct k_stack *stack, |
| uint32_t num_entries) |
| { |
| K_OOPS(K_SYSCALL_OBJ_NEVER_INIT(stack, K_OBJ_STACK)); |
| K_OOPS(K_SYSCALL_VERIFY(num_entries > 0)); |
| return z_impl_k_stack_alloc_init(stack, num_entries); |
| } |
| #include <syscalls/k_stack_alloc_init_mrsh.c> |
| #endif /* CONFIG_USERSPACE */ |
| |
| int k_stack_cleanup(struct k_stack *stack) |
| { |
| SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_stack, cleanup, stack); |
| |
| CHECKIF(z_waitq_head(&stack->wait_q) != NULL) { |
| SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_stack, cleanup, stack, -EAGAIN); |
| |
| return -EAGAIN; |
| } |
| |
| if ((stack->flags & K_STACK_FLAG_ALLOC) != (uint8_t)0) { |
| k_free(stack->base); |
| stack->base = NULL; |
| stack->flags &= ~K_STACK_FLAG_ALLOC; |
| } |
| |
| SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_stack, cleanup, stack, 0); |
| |
| return 0; |
| } |
| |
| int z_impl_k_stack_push(struct k_stack *stack, stack_data_t data) |
| { |
| struct k_thread *first_pending_thread; |
| int ret = 0; |
| k_spinlock_key_t key = k_spin_lock(&stack->lock); |
| |
| SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_stack, push, stack); |
| |
| CHECKIF(stack->next == stack->top) { |
| ret = -ENOMEM; |
| goto out; |
| } |
| |
| first_pending_thread = z_unpend_first_thread(&stack->wait_q); |
| |
| if (first_pending_thread != NULL) { |
| z_thread_return_value_set_with_data(first_pending_thread, |
| 0, (void *)data); |
| |
| z_ready_thread(first_pending_thread); |
| z_reschedule(&stack->lock, key); |
| goto end; |
| } else { |
| *(stack->next) = data; |
| stack->next++; |
| goto out; |
| } |
| |
| out: |
| k_spin_unlock(&stack->lock, key); |
| |
| end: |
| SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_stack, push, stack, ret); |
| |
| return ret; |
| } |
| |
| #ifdef CONFIG_USERSPACE |
| static inline int z_vrfy_k_stack_push(struct k_stack *stack, stack_data_t data) |
| { |
| K_OOPS(K_SYSCALL_OBJ(stack, K_OBJ_STACK)); |
| |
| return z_impl_k_stack_push(stack, data); |
| } |
| #include <syscalls/k_stack_push_mrsh.c> |
| #endif /* CONFIG_USERSPACE */ |
| |
| int z_impl_k_stack_pop(struct k_stack *stack, stack_data_t *data, |
| k_timeout_t timeout) |
| { |
| k_spinlock_key_t key; |
| int result; |
| |
| key = k_spin_lock(&stack->lock); |
| |
| SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_stack, pop, stack, timeout); |
| |
| if (likely(stack->next > stack->base)) { |
| stack->next--; |
| *data = *(stack->next); |
| k_spin_unlock(&stack->lock, key); |
| |
| SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_stack, pop, stack, timeout, 0); |
| |
| return 0; |
| } |
| |
| SYS_PORT_TRACING_OBJ_FUNC_BLOCKING(k_stack, pop, stack, timeout); |
| |
| if (K_TIMEOUT_EQ(timeout, K_NO_WAIT)) { |
| k_spin_unlock(&stack->lock, key); |
| |
| SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_stack, pop, stack, timeout, -EBUSY); |
| |
| return -EBUSY; |
| } |
| |
| result = z_pend_curr(&stack->lock, key, &stack->wait_q, timeout); |
| if (result == -EAGAIN) { |
| SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_stack, pop, stack, timeout, -EAGAIN); |
| |
| return -EAGAIN; |
| } |
| |
| *data = (stack_data_t)_current->base.swap_data; |
| |
| SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_stack, pop, stack, timeout, 0); |
| |
| return 0; |
| } |
| |
| #ifdef CONFIG_USERSPACE |
| static inline int z_vrfy_k_stack_pop(struct k_stack *stack, |
| stack_data_t *data, k_timeout_t timeout) |
| { |
| K_OOPS(K_SYSCALL_OBJ(stack, K_OBJ_STACK)); |
| K_OOPS(K_SYSCALL_MEMORY_WRITE(data, sizeof(stack_data_t))); |
| return z_impl_k_stack_pop(stack, data, timeout); |
| } |
| #include <syscalls/k_stack_pop_mrsh.c> |
| #endif /* CONFIG_USERSPACE */ |
| |
| #ifdef CONFIG_OBJ_CORE_STACK |
| static int init_stack_obj_core_list(void) |
| { |
| /* Initialize stack object type */ |
| |
| z_obj_type_init(&obj_type_stack, K_OBJ_TYPE_STACK_ID, |
| offsetof(struct k_stack, obj_core)); |
| |
| /* Initialize and link statically defined stacks */ |
| |
| STRUCT_SECTION_FOREACH(k_stack, stack) { |
| k_obj_core_init_and_link(K_OBJ_CORE(stack), &obj_type_stack); |
| } |
| |
| return 0; |
| } |
| |
| SYS_INIT(init_stack_obj_core_list, PRE_KERNEL_1, |
| CONFIG_KERNEL_INIT_PRIORITY_OBJECTS); |
| #endif /* CONFIG_OBJ_CORE_STACK */ |