|  | /* | 
|  | * Copyright (c) 2016 Wind River Systems, Inc. | 
|  | * | 
|  | * SPDX-License-Identifier: Apache-2.0 | 
|  | */ | 
|  |  | 
|  | /** | 
|  | * @file @brief mutex kernel services | 
|  | * | 
|  | * This module contains routines for handling mutex locking and unlocking. | 
|  | * | 
|  | * Mutexes implement a priority inheritance algorithm that boosts the priority | 
|  | * level of the owning thread to match the priority level of the highest | 
|  | * priority thread waiting on the mutex. | 
|  | * | 
|  | * Each mutex that contributes to priority inheritance must be released in the | 
|  | * reverse order in which it was acquired.  Furthermore each subsequent mutex | 
|  | * that contributes to raising the owning thread's priority level must be | 
|  | * acquired at a point after the most recent "bumping" of the priority level. | 
|  | * | 
|  | * For example, if thread A has two mutexes contributing to the raising of its | 
|  | * priority level, the second mutex M2 must be acquired by thread A after | 
|  | * thread A's priority level was bumped due to owning the first mutex M1. | 
|  | * When releasing the mutex, thread A must release M2 before it releases M1. | 
|  | * Failure to follow this nested model may result in threads running at | 
|  | * unexpected priority levels (too high, or too low). | 
|  | */ | 
|  |  | 
|  | #include <kernel.h> | 
|  | #include <kernel_structs.h> | 
|  | #include <toolchain.h> | 
|  | #include <linker/sections.h> | 
|  | #include <wait_q.h> | 
|  | #include <misc/dlist.h> | 
|  | #include <debug/object_tracing_common.h> | 
|  | #include <errno.h> | 
|  | #include <init.h> | 
|  | #include <syscall_handler.h> | 
|  | #include <tracing.h> | 
|  |  | 
|  | #define RECORD_STATE_CHANGE(mutex) do { } while (false) | 
|  | #define RECORD_CONFLICT(mutex) do { } while (false) | 
|  |  | 
|  |  | 
|  | extern struct k_mutex _k_mutex_list_start[]; | 
|  | extern struct k_mutex _k_mutex_list_end[]; | 
|  |  | 
|  | #ifdef CONFIG_OBJECT_TRACING | 
|  |  | 
|  | struct k_mutex *_trace_list_k_mutex; | 
|  |  | 
|  | /* | 
|  | * Complete initialization of statically defined mutexes. | 
|  | */ | 
|  | static int init_mutex_module(struct device *dev) | 
|  | { | 
|  | ARG_UNUSED(dev); | 
|  |  | 
|  | struct k_mutex *mutex; | 
|  |  | 
|  | for (mutex = _k_mutex_list_start; mutex < _k_mutex_list_end; mutex++) { | 
|  | SYS_TRACING_OBJ_INIT(k_mutex, mutex); | 
|  | } | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | SYS_INIT(init_mutex_module, PRE_KERNEL_1, CONFIG_KERNEL_INIT_PRIORITY_OBJECTS); | 
|  |  | 
|  | #endif /* CONFIG_OBJECT_TRACING */ | 
|  |  | 
|  | void _impl_k_mutex_init(struct k_mutex *mutex) | 
|  | { | 
|  | mutex->owner = NULL; | 
|  | mutex->lock_count = 0; | 
|  |  | 
|  | sys_trace_void(SYS_TRACE_ID_MUTEX_INIT); | 
|  |  | 
|  | /* initialized upon first use */ | 
|  | /* mutex->owner_orig_prio = 0; */ | 
|  |  | 
|  | _waitq_init(&mutex->wait_q); | 
|  |  | 
|  |  | 
|  | SYS_TRACING_OBJ_INIT(k_mutex, mutex); | 
|  | _k_object_init(mutex); | 
|  | sys_trace_end_call(SYS_TRACE_ID_MUTEX_INIT); | 
|  | } | 
|  |  | 
|  | #ifdef CONFIG_USERSPACE | 
|  | Z_SYSCALL_HANDLER(k_mutex_init, mutex) | 
|  | { | 
|  | Z_OOPS(Z_SYSCALL_OBJ_INIT(mutex, K_OBJ_MUTEX)); | 
|  | _impl_k_mutex_init((struct k_mutex *)mutex); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  | #endif | 
|  |  | 
|  | static int new_prio_for_inheritance(int target, int limit) | 
|  | { | 
|  | int new_prio = _is_prio_higher(target, limit) ? target : limit; | 
|  |  | 
|  | new_prio = _get_new_prio_with_ceiling(new_prio); | 
|  |  | 
|  | return new_prio; | 
|  | } | 
|  |  | 
|  | static void adjust_owner_prio(struct k_mutex *mutex, int new_prio) | 
|  | { | 
|  | if (mutex->owner->base.prio != new_prio) { | 
|  |  | 
|  | K_DEBUG("%p (ready (y/n): %c) prio changed to %d (was %d)\n", | 
|  | mutex->owner, _is_thread_ready(mutex->owner) ? | 
|  | 'y' : 'n', | 
|  | new_prio, mutex->owner->base.prio); | 
|  |  | 
|  | _thread_priority_set(mutex->owner, new_prio); | 
|  | } | 
|  | } | 
|  |  | 
|  | int _impl_k_mutex_lock(struct k_mutex *mutex, s32_t timeout) | 
|  | { | 
|  | int new_prio; | 
|  | unsigned int key; | 
|  |  | 
|  | sys_trace_void(SYS_TRACE_ID_MUTEX_LOCK); | 
|  | _sched_lock(); | 
|  |  | 
|  | if (likely(mutex->lock_count == 0 || mutex->owner == _current)) { | 
|  |  | 
|  | RECORD_STATE_CHANGE(); | 
|  |  | 
|  | mutex->owner_orig_prio = mutex->lock_count == 0 ? | 
|  | _current->base.prio : | 
|  | mutex->owner_orig_prio; | 
|  |  | 
|  | mutex->lock_count++; | 
|  | mutex->owner = _current; | 
|  |  | 
|  | K_DEBUG("%p took mutex %p, count: %d, orig prio: %d\n", | 
|  | _current, mutex, mutex->lock_count, | 
|  | mutex->owner_orig_prio); | 
|  |  | 
|  | k_sched_unlock(); | 
|  | sys_trace_end_call(SYS_TRACE_ID_MUTEX_LOCK); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | RECORD_CONFLICT(); | 
|  |  | 
|  | if (unlikely(timeout == K_NO_WAIT)) { | 
|  | k_sched_unlock(); | 
|  | sys_trace_end_call(SYS_TRACE_ID_MUTEX_LOCK); | 
|  | return -EBUSY; | 
|  | } | 
|  |  | 
|  | new_prio = new_prio_for_inheritance(_current->base.prio, | 
|  | mutex->owner->base.prio); | 
|  |  | 
|  | key = irq_lock(); | 
|  |  | 
|  | K_DEBUG("adjusting prio up on mutex %p\n", mutex); | 
|  |  | 
|  | if (_is_prio_higher(new_prio, mutex->owner->base.prio)) { | 
|  | adjust_owner_prio(mutex, new_prio); | 
|  | } | 
|  |  | 
|  | int got_mutex = _pend_current_thread(key, &mutex->wait_q, timeout); | 
|  |  | 
|  | K_DEBUG("on mutex %p got_mutex value: %d\n", mutex, got_mutex); | 
|  |  | 
|  | K_DEBUG("%p got mutex %p (y/n): %c\n", _current, mutex, | 
|  | got_mutex ? 'y' : 'n'); | 
|  |  | 
|  | if (got_mutex == 0) { | 
|  | k_sched_unlock(); | 
|  | sys_trace_end_call(SYS_TRACE_ID_MUTEX_LOCK); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* timed out */ | 
|  |  | 
|  | K_DEBUG("%p timeout on mutex %p\n", _current, mutex); | 
|  |  | 
|  | struct k_thread *waiter = _waitq_head(&mutex->wait_q); | 
|  |  | 
|  | new_prio = mutex->owner_orig_prio; | 
|  | new_prio = waiter ? new_prio_for_inheritance(waiter->base.prio, | 
|  | new_prio) : new_prio; | 
|  |  | 
|  | K_DEBUG("adjusting prio down on mutex %p\n", mutex); | 
|  |  | 
|  | key = irq_lock(); | 
|  | adjust_owner_prio(mutex, new_prio); | 
|  | irq_unlock(key); | 
|  |  | 
|  | k_sched_unlock(); | 
|  |  | 
|  | sys_trace_end_call(SYS_TRACE_ID_MUTEX_LOCK); | 
|  | return -EAGAIN; | 
|  | } | 
|  |  | 
|  | #ifdef CONFIG_USERSPACE | 
|  | Z_SYSCALL_HANDLER(k_mutex_lock, mutex, timeout) | 
|  | { | 
|  | Z_OOPS(Z_SYSCALL_OBJ(mutex, K_OBJ_MUTEX)); | 
|  | return _impl_k_mutex_lock((struct k_mutex *)mutex, (s32_t)timeout); | 
|  | } | 
|  | #endif | 
|  |  | 
|  | void _impl_k_mutex_unlock(struct k_mutex *mutex) | 
|  | { | 
|  | unsigned int key; | 
|  |  | 
|  | __ASSERT(mutex->lock_count > 0, ""); | 
|  | __ASSERT(mutex->owner == _current, ""); | 
|  |  | 
|  | sys_trace_void(SYS_TRACE_ID_MUTEX_UNLOCK); | 
|  | _sched_lock(); | 
|  |  | 
|  | RECORD_STATE_CHANGE(); | 
|  |  | 
|  | mutex->lock_count--; | 
|  |  | 
|  | K_DEBUG("mutex %p lock_count: %d\n", mutex, mutex->lock_count); | 
|  |  | 
|  | if (mutex->lock_count != 0) { | 
|  | k_sched_unlock(); | 
|  | return; | 
|  | } | 
|  |  | 
|  | key = irq_lock(); | 
|  |  | 
|  | adjust_owner_prio(mutex, mutex->owner_orig_prio); | 
|  |  | 
|  | struct k_thread *new_owner = _unpend_first_thread(&mutex->wait_q); | 
|  |  | 
|  | mutex->owner = new_owner; | 
|  |  | 
|  | K_DEBUG("new owner of mutex %p: %p (prio: %d)\n", | 
|  | mutex, new_owner, new_owner ? new_owner->base.prio : -1000); | 
|  |  | 
|  | if (new_owner != NULL) { | 
|  | _ready_thread(new_owner); | 
|  |  | 
|  | irq_unlock(key); | 
|  |  | 
|  | _set_thread_return_value(new_owner, 0); | 
|  |  | 
|  | /* | 
|  | * new owner is already of higher or equal prio than first | 
|  | * waiter since the wait queue is priority-based: no need to | 
|  | * ajust its priority | 
|  | */ | 
|  | mutex->lock_count++; | 
|  | mutex->owner_orig_prio = new_owner->base.prio; | 
|  | } | 
|  |  | 
|  | irq_unlock(key); | 
|  |  | 
|  | k_sched_unlock(); | 
|  | } | 
|  |  | 
|  | #ifdef CONFIG_USERSPACE | 
|  | Z_SYSCALL_HANDLER(k_mutex_unlock, mutex) | 
|  | { | 
|  | Z_OOPS(Z_SYSCALL_OBJ(mutex, K_OBJ_MUTEX)); | 
|  | Z_OOPS(Z_SYSCALL_VERIFY(((struct k_mutex *)mutex)->lock_count > 0)); | 
|  | Z_OOPS(Z_SYSCALL_VERIFY(((struct k_mutex *)mutex)->owner == _current)); | 
|  | _impl_k_mutex_unlock((struct k_mutex *)mutex); | 
|  | return 0; | 
|  | } | 
|  | #endif |