| /* |
| * Copyright (c) 2020 Nordic Semiconductor ASA |
| * |
| * SPDX-License-Identifier: Apache-2.0 |
| */ |
| |
| /** |
| * @file |
| * |
| * Second generation work queue implementation |
| */ |
| |
| #include <zephyr/kernel.h> |
| #include <zephyr/kernel_structs.h> |
| #include <wait_q.h> |
| #include <zephyr/spinlock.h> |
| #include <errno.h> |
| #include <ksched.h> |
| #include <zephyr/sys/printk.h> |
| |
| static inline void flag_clear(uint32_t *flagp, |
| uint32_t bit) |
| { |
| *flagp &= ~BIT(bit); |
| } |
| |
| static inline void flag_set(uint32_t *flagp, |
| uint32_t bit) |
| { |
| *flagp |= BIT(bit); |
| } |
| |
| static inline bool flag_test(const uint32_t *flagp, |
| uint32_t bit) |
| { |
| return (*flagp & BIT(bit)) != 0U; |
| } |
| |
| static inline bool flag_test_and_clear(uint32_t *flagp, |
| int bit) |
| { |
| bool ret = flag_test(flagp, bit); |
| |
| flag_clear(flagp, bit); |
| |
| return ret; |
| } |
| |
| static inline void flags_set(uint32_t *flagp, |
| uint32_t flags) |
| { |
| *flagp = flags; |
| } |
| |
| static inline uint32_t flags_get(const uint32_t *flagp) |
| { |
| return *flagp; |
| } |
| |
| /* Lock to protect the internal state of all work items, work queues, |
| * and pending_cancels. |
| */ |
| static struct k_spinlock lock; |
| |
| /* Invoked by work thread */ |
| static void handle_flush(struct k_work *work) { } |
| |
| static inline void init_flusher(struct z_work_flusher *flusher) |
| { |
| struct k_work *work = &flusher->work; |
| k_sem_init(&flusher->sem, 0, 1); |
| k_work_init(&flusher->work, handle_flush); |
| flag_set(&work->flags, K_WORK_FLUSHING_BIT); |
| } |
| |
| /* List of pending cancellations. */ |
| static sys_slist_t pending_cancels; |
| |
| /* Initialize a canceler record and add it to the list of pending |
| * cancels. |
| * |
| * Invoked with work lock held. |
| * |
| * @param canceler the structure used to notify a waiting process. |
| * @param work the work structure that is to be canceled |
| */ |
| static inline void init_work_cancel(struct z_work_canceller *canceler, |
| struct k_work *work) |
| { |
| k_sem_init(&canceler->sem, 0, 1); |
| canceler->work = work; |
| sys_slist_append(&pending_cancels, &canceler->node); |
| } |
| |
| /* Complete flushing of a work item. |
| * |
| * Invoked with work lock held. |
| * |
| * Invoked from a work queue thread. |
| * |
| * Reschedules. |
| * |
| * @param work the work structure that has completed flushing. |
| */ |
| static void finalize_flush_locked(struct k_work *work) |
| { |
| struct z_work_flusher *flusher |
| = CONTAINER_OF(work, struct z_work_flusher, work); |
| |
| flag_clear(&work->flags, K_WORK_FLUSHING_BIT); |
| |
| k_sem_give(&flusher->sem); |
| }; |
| |
| /* Complete cancellation of a work item and unlock held lock. |
| * |
| * Invoked with work lock held. |
| * |
| * Invoked from a work queue thread. |
| * |
| * Reschedules. |
| * |
| * @param work the work structure that has completed cancellation |
| */ |
| static void finalize_cancel_locked(struct k_work *work) |
| { |
| struct z_work_canceller *wc, *tmp; |
| sys_snode_t *prev = NULL; |
| |
| /* Clear this first, so released high-priority threads don't |
| * see it when doing things. |
| */ |
| flag_clear(&work->flags, K_WORK_CANCELING_BIT); |
| |
| /* Search for and remove the matching container, and release |
| * what's waiting for the completion. The same work item can |
| * appear multiple times in the list if multiple threads |
| * attempt to cancel it. |
| */ |
| SYS_SLIST_FOR_EACH_CONTAINER_SAFE(&pending_cancels, wc, tmp, node) { |
| if (wc->work == work) { |
| sys_slist_remove(&pending_cancels, prev, &wc->node); |
| k_sem_give(&wc->sem); |
| break; |
| } |
| prev = &wc->node; |
| } |
| } |
| |
| void k_work_init(struct k_work *work, |
| k_work_handler_t handler) |
| { |
| __ASSERT_NO_MSG(work != NULL); |
| __ASSERT_NO_MSG(handler != NULL); |
| |
| *work = (struct k_work)Z_WORK_INITIALIZER(handler); |
| |
| SYS_PORT_TRACING_OBJ_INIT(k_work, work); |
| } |
| |
| static inline int work_busy_get_locked(const struct k_work *work) |
| { |
| return flags_get(&work->flags) & K_WORK_MASK; |
| } |
| |
| int k_work_busy_get(const struct k_work *work) |
| { |
| k_spinlock_key_t key = k_spin_lock(&lock); |
| int ret = work_busy_get_locked(work); |
| |
| k_spin_unlock(&lock, key); |
| |
| return ret; |
| } |
| |
| /* Add a flusher work item to the queue. |
| * |
| * Invoked with work lock held. |
| * |
| * Caller must notify queue of pending work. |
| * |
| * @param queue queue on which a work item may appear. |
| * @param work the work item that is either queued or running on @p |
| * queue |
| * @param flusher an uninitialized/unused flusher object |
| */ |
| static void queue_flusher_locked(struct k_work_q *queue, |
| struct k_work *work, |
| struct z_work_flusher *flusher) |
| { |
| bool in_list = false; |
| struct k_work *wn; |
| |
| /* Determine whether the work item is still queued. */ |
| SYS_SLIST_FOR_EACH_CONTAINER(&queue->pending, wn, node) { |
| if (wn == work) { |
| in_list = true; |
| break; |
| } |
| } |
| |
| init_flusher(flusher); |
| if (in_list) { |
| sys_slist_insert(&queue->pending, &work->node, |
| &flusher->work.node); |
| } else { |
| sys_slist_prepend(&queue->pending, &flusher->work.node); |
| } |
| } |
| |
| /* Try to remove a work item from the given queue. |
| * |
| * Invoked with work lock held. |
| * |
| * @param queue the queue from which the work should be removed |
| * @param work work that may be on the queue |
| */ |
| static inline void queue_remove_locked(struct k_work_q *queue, |
| struct k_work *work) |
| { |
| if (flag_test_and_clear(&work->flags, K_WORK_QUEUED_BIT)) { |
| (void)sys_slist_find_and_remove(&queue->pending, &work->node); |
| } |
| } |
| |
| /* Potentially notify a queue that it needs to look for pending work. |
| * |
| * This may make the work queue thread ready, but as the lock is held it |
| * will not be a reschedule point. Callers should yield after the lock is |
| * released where appropriate (generally if this returns true). |
| * |
| * @param queue to be notified. If this is null no notification is required. |
| * |
| * @return true if and only if the queue was notified and woken, i.e. a |
| * reschedule is pending. |
| */ |
| static inline bool notify_queue_locked(struct k_work_q *queue) |
| { |
| bool rv = false; |
| |
| if (queue != NULL) { |
| rv = z_sched_wake(&queue->notifyq, 0, NULL); |
| } |
| |
| return rv; |
| } |
| |
| /* Submit an work item to a queue if queue state allows new work. |
| * |
| * Submission is rejected if no queue is provided, or if the queue is |
| * draining and the work isn't being submitted from the queue's |
| * thread (chained submission). |
| * |
| * Invoked with work lock held. |
| * Conditionally notifies queue. |
| * |
| * @param queue the queue to which work should be submitted. This may |
| * be null, in which case the submission will fail. |
| * |
| * @param work to be submitted |
| * |
| * @retval 1 if successfully queued |
| * @retval -EINVAL if no queue is provided |
| * @retval -ENODEV if the queue is not started |
| * @retval -EBUSY if the submission was rejected (draining, plugged) |
| */ |
| static inline int queue_submit_locked(struct k_work_q *queue, |
| struct k_work *work) |
| { |
| if (queue == NULL) { |
| return -EINVAL; |
| } |
| |
| int ret; |
| bool chained = (_current == &queue->thread) && !k_is_in_isr(); |
| bool draining = flag_test(&queue->flags, K_WORK_QUEUE_DRAIN_BIT); |
| bool plugged = flag_test(&queue->flags, K_WORK_QUEUE_PLUGGED_BIT); |
| |
| /* Test for acceptability, in priority order: |
| * |
| * * -ENODEV if the queue isn't running. |
| * * -EBUSY if draining and not chained |
| * * -EBUSY if plugged and not draining |
| * * otherwise OK |
| */ |
| if (!flag_test(&queue->flags, K_WORK_QUEUE_STARTED_BIT)) { |
| ret = -ENODEV; |
| } else if (draining && !chained) { |
| ret = -EBUSY; |
| } else if (plugged && !draining) { |
| ret = -EBUSY; |
| } else { |
| sys_slist_append(&queue->pending, &work->node); |
| ret = 1; |
| (void)notify_queue_locked(queue); |
| } |
| |
| return ret; |
| } |
| |
| /* Attempt to submit work to a queue. |
| * |
| * The submission can fail if: |
| * * the work is cancelling, |
| * * no candidate queue can be identified; |
| * * the candidate queue rejects the submission. |
| * |
| * Invoked with work lock held. |
| * Conditionally notifies queue. |
| * |
| * @param work the work structure to be submitted |
| |
| * @param queuep pointer to a queue reference. On input this should |
| * dereference to the proposed queue (which may be null); after completion it |
| * will be null if the work was not submitted or if submitted will reference |
| * the queue it was submitted to. That may or may not be the queue provided |
| * on input. |
| * |
| * @retval 0 if work was already submitted to a queue |
| * @retval 1 if work was not submitted and has been queued to @p queue |
| * @retval 2 if work was running and has been queued to the queue that was |
| * running it |
| * @retval -EBUSY if canceling or submission was rejected by queue |
| * @retval -EINVAL if no queue is provided |
| * @retval -ENODEV if the queue is not started |
| */ |
| static int submit_to_queue_locked(struct k_work *work, |
| struct k_work_q **queuep) |
| { |
| int ret = 0; |
| |
| if (flag_test(&work->flags, K_WORK_CANCELING_BIT)) { |
| /* Disallowed */ |
| ret = -EBUSY; |
| } else if (!flag_test(&work->flags, K_WORK_QUEUED_BIT)) { |
| /* Not currently queued */ |
| ret = 1; |
| |
| /* If no queue specified resubmit to last queue. |
| */ |
| if (*queuep == NULL) { |
| *queuep = work->queue; |
| } |
| |
| /* If the work is currently running we have to use the |
| * queue it's running on to prevent handler |
| * re-entrancy. |
| */ |
| if (flag_test(&work->flags, K_WORK_RUNNING_BIT)) { |
| __ASSERT_NO_MSG(work->queue != NULL); |
| *queuep = work->queue; |
| ret = 2; |
| } |
| |
| int rc = queue_submit_locked(*queuep, work); |
| |
| if (rc < 0) { |
| ret = rc; |
| } else { |
| flag_set(&work->flags, K_WORK_QUEUED_BIT); |
| work->queue = *queuep; |
| } |
| } else { |
| /* Already queued, do nothing. */ |
| } |
| |
| if (ret <= 0) { |
| *queuep = NULL; |
| } |
| |
| return ret; |
| } |
| |
| /* Submit work to a queue but do not yield the current thread. |
| * |
| * Intended for internal use. |
| * |
| * See also submit_to_queue_locked(). |
| * |
| * @param queuep pointer to a queue reference. |
| * @param work the work structure to be submitted |
| * |
| * @retval see submit_to_queue_locked() |
| */ |
| int z_work_submit_to_queue(struct k_work_q *queue, |
| struct k_work *work) |
| { |
| __ASSERT_NO_MSG(work != NULL); |
| __ASSERT_NO_MSG(work->handler != NULL); |
| |
| k_spinlock_key_t key = k_spin_lock(&lock); |
| |
| int ret = submit_to_queue_locked(work, &queue); |
| |
| k_spin_unlock(&lock, key); |
| |
| return ret; |
| } |
| |
| int k_work_submit_to_queue(struct k_work_q *queue, |
| struct k_work *work) |
| { |
| SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work, submit_to_queue, queue, work); |
| |
| int ret = z_work_submit_to_queue(queue, work); |
| |
| /* submit_to_queue_locked() won't reschedule on its own |
| * (really it should, otherwise this process will result in |
| * spurious calls to z_swap() due to the race), so do it here |
| * if the queue state changed. |
| */ |
| if (ret > 0) { |
| z_reschedule_unlocked(); |
| } |
| |
| SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work, submit_to_queue, queue, work, ret); |
| |
| return ret; |
| } |
| |
| int k_work_submit(struct k_work *work) |
| { |
| SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work, submit, work); |
| |
| int ret = k_work_submit_to_queue(&k_sys_work_q, work); |
| |
| SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work, submit, work, ret); |
| |
| return ret; |
| } |
| |
| /* Flush the work item if necessary. |
| * |
| * Flushing is necessary only if the work is either queued or running. |
| * |
| * Invoked with work lock held by key. |
| * Sleeps. |
| * |
| * @param work the work item that is to be flushed |
| * @param flusher state used to synchronize the flush |
| * |
| * @retval true if work is queued or running. If this happens the |
| * caller must take the flusher semaphore after releasing the lock. |
| * |
| * @retval false otherwise. No wait required. |
| */ |
| static bool work_flush_locked(struct k_work *work, |
| struct z_work_flusher *flusher) |
| { |
| bool need_flush = (flags_get(&work->flags) |
| & (K_WORK_QUEUED | K_WORK_RUNNING)) != 0U; |
| |
| if (need_flush) { |
| struct k_work_q *queue = work->queue; |
| |
| __ASSERT_NO_MSG(queue != NULL); |
| |
| queue_flusher_locked(queue, work, flusher); |
| notify_queue_locked(queue); |
| } |
| |
| return need_flush; |
| } |
| |
| bool k_work_flush(struct k_work *work, |
| struct k_work_sync *sync) |
| { |
| __ASSERT_NO_MSG(work != NULL); |
| __ASSERT_NO_MSG(!flag_test(&work->flags, K_WORK_DELAYABLE_BIT)); |
| __ASSERT_NO_MSG(!k_is_in_isr()); |
| __ASSERT_NO_MSG(sync != NULL); |
| #ifdef CONFIG_KERNEL_COHERENCE |
| __ASSERT_NO_MSG(arch_mem_coherent(sync)); |
| #endif /* CONFIG_KERNEL_COHERENCE */ |
| |
| SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work, flush, work); |
| |
| struct z_work_flusher *flusher = &sync->flusher; |
| k_spinlock_key_t key = k_spin_lock(&lock); |
| |
| bool need_flush = work_flush_locked(work, flusher); |
| |
| k_spin_unlock(&lock, key); |
| |
| /* If necessary wait until the flusher item completes */ |
| if (need_flush) { |
| SYS_PORT_TRACING_OBJ_FUNC_BLOCKING(k_work, flush, work, K_FOREVER); |
| |
| k_sem_take(&flusher->sem, K_FOREVER); |
| } |
| |
| SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work, flush, work, need_flush); |
| |
| return need_flush; |
| } |
| |
| /* Execute the non-waiting steps necessary to cancel a work item. |
| * |
| * Invoked with work lock held. |
| * |
| * @param work the work item to be canceled. |
| * |
| * @retval true if we need to wait for the work item to finish canceling |
| * @retval false if the work item is idle |
| * |
| * @return k_busy_wait() captured under lock |
| */ |
| static int cancel_async_locked(struct k_work *work) |
| { |
| /* If we haven't already started canceling, do it now. */ |
| if (!flag_test(&work->flags, K_WORK_CANCELING_BIT)) { |
| /* Remove it from the queue, if it's queued. */ |
| queue_remove_locked(work->queue, work); |
| } |
| |
| /* If it's still busy after it's been dequeued, then flag it |
| * as canceling. |
| */ |
| int ret = work_busy_get_locked(work); |
| |
| if (ret != 0) { |
| flag_set(&work->flags, K_WORK_CANCELING_BIT); |
| ret = work_busy_get_locked(work); |
| } |
| |
| return ret; |
| } |
| |
| /* Complete cancellation necessary, release work lock, and wait if |
| * necessary. |
| * |
| * Invoked with work lock held by key. |
| * Sleeps. |
| * |
| * @param work work that is being canceled |
| * @param canceller state used to synchronize the cancellation |
| * @param key used by work lock |
| * |
| * @retval true if and only if the work was still active on entry. The caller |
| * must wait on the canceller semaphore after releasing the lock. |
| * |
| * @retval false if work was idle on entry. The caller need not wait. |
| */ |
| static bool cancel_sync_locked(struct k_work *work, |
| struct z_work_canceller *canceller) |
| { |
| bool ret = flag_test(&work->flags, K_WORK_CANCELING_BIT); |
| |
| /* If something's still running then we have to wait for |
| * completion, which is indicated when finish_cancel() gets |
| * invoked. |
| */ |
| if (ret) { |
| init_work_cancel(canceller, work); |
| } |
| |
| return ret; |
| } |
| |
| int k_work_cancel(struct k_work *work) |
| { |
| __ASSERT_NO_MSG(work != NULL); |
| __ASSERT_NO_MSG(!flag_test(&work->flags, K_WORK_DELAYABLE_BIT)); |
| |
| SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work, cancel, work); |
| |
| k_spinlock_key_t key = k_spin_lock(&lock); |
| int ret = cancel_async_locked(work); |
| |
| k_spin_unlock(&lock, key); |
| |
| SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work, cancel, work, ret); |
| |
| return ret; |
| } |
| |
| bool k_work_cancel_sync(struct k_work *work, |
| struct k_work_sync *sync) |
| { |
| __ASSERT_NO_MSG(work != NULL); |
| __ASSERT_NO_MSG(sync != NULL); |
| __ASSERT_NO_MSG(!flag_test(&work->flags, K_WORK_DELAYABLE_BIT)); |
| __ASSERT_NO_MSG(!k_is_in_isr()); |
| #ifdef CONFIG_KERNEL_COHERENCE |
| __ASSERT_NO_MSG(arch_mem_coherent(sync)); |
| #endif /* CONFIG_KERNEL_COHERENCE */ |
| |
| SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work, cancel_sync, work, sync); |
| |
| struct z_work_canceller *canceller = &sync->canceller; |
| k_spinlock_key_t key = k_spin_lock(&lock); |
| bool pending = (work_busy_get_locked(work) != 0U); |
| bool need_wait = false; |
| |
| if (pending) { |
| (void)cancel_async_locked(work); |
| need_wait = cancel_sync_locked(work, canceller); |
| } |
| |
| k_spin_unlock(&lock, key); |
| |
| if (need_wait) { |
| SYS_PORT_TRACING_OBJ_FUNC_BLOCKING(k_work, cancel_sync, work, sync); |
| |
| k_sem_take(&canceller->sem, K_FOREVER); |
| } |
| |
| SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work, cancel_sync, work, sync, pending); |
| return pending; |
| } |
| |
| /* Loop executed by a work queue thread. |
| * |
| * @param workq_ptr pointer to the work queue structure |
| */ |
| static void work_queue_main(void *workq_ptr, void *p2, void *p3) |
| { |
| ARG_UNUSED(p2); |
| ARG_UNUSED(p3); |
| |
| struct k_work_q *queue = (struct k_work_q *)workq_ptr; |
| |
| while (true) { |
| sys_snode_t *node; |
| struct k_work *work = NULL; |
| k_work_handler_t handler = NULL; |
| k_spinlock_key_t key = k_spin_lock(&lock); |
| bool yield; |
| |
| /* Check for and prepare any new work. */ |
| node = sys_slist_get(&queue->pending); |
| if (node != NULL) { |
| /* Mark that there's some work active that's |
| * not on the pending list. |
| */ |
| flag_set(&queue->flags, K_WORK_QUEUE_BUSY_BIT); |
| work = CONTAINER_OF(node, struct k_work, node); |
| flag_set(&work->flags, K_WORK_RUNNING_BIT); |
| flag_clear(&work->flags, K_WORK_QUEUED_BIT); |
| |
| /* Static code analysis tool can raise a false-positive violation |
| * in the line below that 'work' is checked for null after being |
| * dereferenced. |
| * |
| * The work is figured out by CONTAINER_OF, as a container |
| * of type struct k_work that contains the node. |
| * The only way for it to be NULL is if node would be a member |
| * of struct k_work object that has been placed at address NULL, |
| * which should never happen, even line 'if (work != NULL)' |
| * ensures that. |
| * This means that if node is not NULL, then work will not be NULL. |
| */ |
| handler = work->handler; |
| } else if (flag_test_and_clear(&queue->flags, |
| K_WORK_QUEUE_DRAIN_BIT)) { |
| /* Not busy and draining: move threads waiting for |
| * drain to ready state. The held spinlock inhibits |
| * immediate reschedule; released threads get their |
| * chance when this invokes z_sched_wait() below. |
| * |
| * We don't touch K_WORK_QUEUE_PLUGGABLE, so getting |
| * here doesn't mean that the queue will allow new |
| * submissions. |
| */ |
| (void)z_sched_wake_all(&queue->drainq, 1, NULL); |
| } else { |
| /* No work is available and no queue state requires |
| * special handling. |
| */ |
| ; |
| } |
| |
| if (work == NULL) { |
| /* Nothing's had a chance to add work since we took |
| * the lock, and we didn't find work nor got asked to |
| * stop. Just go to sleep: when something happens the |
| * work thread will be woken and we can check again. |
| */ |
| |
| (void)z_sched_wait(&lock, key, &queue->notifyq, |
| K_FOREVER, NULL); |
| continue; |
| } |
| |
| k_spin_unlock(&lock, key); |
| |
| __ASSERT_NO_MSG(handler != NULL); |
| handler(work); |
| |
| /* Mark the work item as no longer running and deal |
| * with any cancellation and flushing issued while it |
| * was running. Clear the BUSY flag and optionally |
| * yield to prevent starving other threads. |
| */ |
| key = k_spin_lock(&lock); |
| |
| flag_clear(&work->flags, K_WORK_RUNNING_BIT); |
| if (flag_test(&work->flags, K_WORK_FLUSHING_BIT)) { |
| finalize_flush_locked(work); |
| } |
| if (flag_test(&work->flags, K_WORK_CANCELING_BIT)) { |
| finalize_cancel_locked(work); |
| } |
| |
| flag_clear(&queue->flags, K_WORK_QUEUE_BUSY_BIT); |
| yield = !flag_test(&queue->flags, K_WORK_QUEUE_NO_YIELD_BIT); |
| k_spin_unlock(&lock, key); |
| |
| /* Optionally yield to prevent the work queue from |
| * starving other threads. |
| */ |
| if (yield) { |
| k_yield(); |
| } |
| } |
| } |
| |
| void k_work_queue_init(struct k_work_q *queue) |
| { |
| __ASSERT_NO_MSG(queue != NULL); |
| |
| *queue = (struct k_work_q) { |
| .flags = 0, |
| }; |
| |
| SYS_PORT_TRACING_OBJ_INIT(k_work_queue, queue); |
| } |
| |
| void k_work_queue_start(struct k_work_q *queue, |
| k_thread_stack_t *stack, |
| size_t stack_size, |
| int prio, |
| const struct k_work_queue_config *cfg) |
| { |
| __ASSERT_NO_MSG(queue); |
| __ASSERT_NO_MSG(stack); |
| __ASSERT_NO_MSG(!flag_test(&queue->flags, K_WORK_QUEUE_STARTED_BIT)); |
| uint32_t flags = K_WORK_QUEUE_STARTED; |
| |
| SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work_queue, start, queue); |
| |
| sys_slist_init(&queue->pending); |
| z_waitq_init(&queue->notifyq); |
| z_waitq_init(&queue->drainq); |
| |
| if ((cfg != NULL) && cfg->no_yield) { |
| flags |= K_WORK_QUEUE_NO_YIELD; |
| } |
| |
| /* It hasn't actually been started yet, but all the state is in place |
| * so we can submit things and once the thread gets control it's ready |
| * to roll. |
| */ |
| flags_set(&queue->flags, flags); |
| |
| (void)k_thread_create(&queue->thread, stack, stack_size, |
| work_queue_main, queue, NULL, NULL, |
| prio, 0, K_FOREVER); |
| |
| if ((cfg != NULL) && (cfg->name != NULL)) { |
| k_thread_name_set(&queue->thread, cfg->name); |
| } |
| |
| if ((cfg != NULL) && (cfg->essential)) { |
| queue->thread.base.user_options |= K_ESSENTIAL; |
| } |
| |
| k_thread_start(&queue->thread); |
| |
| SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work_queue, start, queue); |
| } |
| |
| int k_work_queue_drain(struct k_work_q *queue, |
| bool plug) |
| { |
| __ASSERT_NO_MSG(queue); |
| __ASSERT_NO_MSG(!k_is_in_isr()); |
| |
| SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work_queue, drain, queue); |
| |
| int ret = 0; |
| k_spinlock_key_t key = k_spin_lock(&lock); |
| |
| if (((flags_get(&queue->flags) |
| & (K_WORK_QUEUE_BUSY | K_WORK_QUEUE_DRAIN)) != 0U) |
| || plug |
| || !sys_slist_is_empty(&queue->pending)) { |
| flag_set(&queue->flags, K_WORK_QUEUE_DRAIN_BIT); |
| if (plug) { |
| flag_set(&queue->flags, K_WORK_QUEUE_PLUGGED_BIT); |
| } |
| |
| notify_queue_locked(queue); |
| ret = z_sched_wait(&lock, key, &queue->drainq, |
| K_FOREVER, NULL); |
| } else { |
| k_spin_unlock(&lock, key); |
| } |
| |
| SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work_queue, drain, queue, ret); |
| |
| return ret; |
| } |
| |
| int k_work_queue_unplug(struct k_work_q *queue) |
| { |
| __ASSERT_NO_MSG(queue); |
| |
| SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work_queue, unplug, queue); |
| |
| int ret = -EALREADY; |
| k_spinlock_key_t key = k_spin_lock(&lock); |
| |
| if (flag_test_and_clear(&queue->flags, K_WORK_QUEUE_PLUGGED_BIT)) { |
| ret = 0; |
| } |
| |
| k_spin_unlock(&lock, key); |
| |
| SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work_queue, unplug, queue, ret); |
| |
| return ret; |
| } |
| |
| #ifdef CONFIG_SYS_CLOCK_EXISTS |
| |
| /* Timeout handler for delayable work. |
| * |
| * Invoked by timeout infrastructure. |
| * Takes and releases work lock. |
| * Conditionally reschedules. |
| */ |
| static void work_timeout(struct _timeout *to) |
| { |
| struct k_work_delayable *dw |
| = CONTAINER_OF(to, struct k_work_delayable, timeout); |
| struct k_work *wp = &dw->work; |
| k_spinlock_key_t key = k_spin_lock(&lock); |
| struct k_work_q *queue = NULL; |
| |
| /* If the work is still marked delayed (should be) then clear that |
| * state and submit it to the queue. If successful the queue will be |
| * notified of new work at the next reschedule point. |
| * |
| * If not successful there is no notification that the work has been |
| * abandoned. Sorry. |
| */ |
| if (flag_test_and_clear(&wp->flags, K_WORK_DELAYED_BIT)) { |
| queue = dw->queue; |
| (void)submit_to_queue_locked(wp, &queue); |
| } |
| |
| k_spin_unlock(&lock, key); |
| } |
| |
| void k_work_init_delayable(struct k_work_delayable *dwork, |
| k_work_handler_t handler) |
| { |
| __ASSERT_NO_MSG(dwork != NULL); |
| __ASSERT_NO_MSG(handler != NULL); |
| |
| *dwork = (struct k_work_delayable){ |
| .work = { |
| .handler = handler, |
| .flags = K_WORK_DELAYABLE, |
| }, |
| }; |
| z_init_timeout(&dwork->timeout); |
| |
| SYS_PORT_TRACING_OBJ_INIT(k_work_delayable, dwork); |
| } |
| |
| static inline int work_delayable_busy_get_locked(const struct k_work_delayable *dwork) |
| { |
| return flags_get(&dwork->work.flags) & K_WORK_MASK; |
| } |
| |
| int k_work_delayable_busy_get(const struct k_work_delayable *dwork) |
| { |
| k_spinlock_key_t key = k_spin_lock(&lock); |
| int ret = work_delayable_busy_get_locked(dwork); |
| |
| k_spin_unlock(&lock, key); |
| return ret; |
| } |
| |
| /* Attempt to schedule a work item for future (maybe immediate) |
| * submission. |
| * |
| * Invoked with work lock held. |
| * |
| * See also submit_to_queue_locked(), which implements this for a no-wait |
| * delay. |
| * |
| * Invoked with work lock held. |
| * |
| * @param queuep pointer to a pointer to a queue. On input this |
| * should dereference to the proposed queue (which may be null); after |
| * completion it will be null if the work was not submitted or if |
| * submitted will reference the queue it was submitted to. That may |
| * or may not be the queue provided on input. |
| * |
| * @param dwork the delayed work structure |
| * |
| * @param delay the delay to use before scheduling. |
| * |
| * @retval from submit_to_queue_locked() if delay is K_NO_WAIT; otherwise |
| * @retval 1 to indicate successfully scheduled. |
| */ |
| static int schedule_for_queue_locked(struct k_work_q **queuep, |
| struct k_work_delayable *dwork, |
| k_timeout_t delay) |
| { |
| int ret = 1; |
| struct k_work *work = &dwork->work; |
| |
| if (K_TIMEOUT_EQ(delay, K_NO_WAIT)) { |
| return submit_to_queue_locked(work, queuep); |
| } |
| |
| flag_set(&work->flags, K_WORK_DELAYED_BIT); |
| dwork->queue = *queuep; |
| |
| /* Add timeout */ |
| z_add_timeout(&dwork->timeout, work_timeout, delay); |
| |
| return ret; |
| } |
| |
| /* Unschedule delayable work. |
| * |
| * If the work is delayed, cancel the timeout and clear the delayed |
| * flag. |
| * |
| * Invoked with work lock held. |
| * |
| * @param dwork pointer to delayable work structure. |
| * |
| * @return true if and only if work had been delayed so the timeout |
| * was cancelled. |
| */ |
| static inline bool unschedule_locked(struct k_work_delayable *dwork) |
| { |
| bool ret = false; |
| struct k_work *work = &dwork->work; |
| |
| /* If scheduled, try to cancel. If it fails, that means the |
| * callback has been dequeued and will inevitably run (or has |
| * already run), so treat that as "undelayed" and return |
| * false. |
| */ |
| if (flag_test_and_clear(&work->flags, K_WORK_DELAYED_BIT)) { |
| ret = z_abort_timeout(&dwork->timeout) == 0; |
| } |
| |
| return ret; |
| } |
| |
| /* Full cancellation of a delayable work item. |
| * |
| * Unschedules the delayed part then delegates to standard work |
| * cancellation. |
| * |
| * Invoked with work lock held. |
| * |
| * @param dwork delayable work item |
| * |
| * @return k_work_busy_get() flags |
| */ |
| static int cancel_delayable_async_locked(struct k_work_delayable *dwork) |
| { |
| (void)unschedule_locked(dwork); |
| |
| return cancel_async_locked(&dwork->work); |
| } |
| |
| int k_work_schedule_for_queue(struct k_work_q *queue, |
| struct k_work_delayable *dwork, |
| k_timeout_t delay) |
| { |
| __ASSERT_NO_MSG(dwork != NULL); |
| |
| SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work, schedule_for_queue, queue, dwork, delay); |
| |
| struct k_work *work = &dwork->work; |
| int ret = 0; |
| k_spinlock_key_t key = k_spin_lock(&lock); |
| |
| /* Schedule the work item if it's idle or running. */ |
| if ((work_busy_get_locked(work) & ~K_WORK_RUNNING) == 0U) { |
| ret = schedule_for_queue_locked(&queue, dwork, delay); |
| } |
| |
| k_spin_unlock(&lock, key); |
| |
| SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work, schedule_for_queue, queue, dwork, delay, ret); |
| |
| return ret; |
| } |
| |
| int k_work_schedule(struct k_work_delayable *dwork, |
| k_timeout_t delay) |
| { |
| SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work, schedule, dwork, delay); |
| |
| int ret = k_work_schedule_for_queue(&k_sys_work_q, dwork, delay); |
| |
| SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work, schedule, dwork, delay, ret); |
| |
| return ret; |
| } |
| |
| int k_work_reschedule_for_queue(struct k_work_q *queue, |
| struct k_work_delayable *dwork, |
| k_timeout_t delay) |
| { |
| __ASSERT_NO_MSG(dwork != NULL); |
| |
| SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work, reschedule_for_queue, queue, dwork, delay); |
| |
| int ret; |
| k_spinlock_key_t key = k_spin_lock(&lock); |
| |
| /* Remove any active scheduling. */ |
| (void)unschedule_locked(dwork); |
| |
| /* Schedule the work item with the new parameters. */ |
| ret = schedule_for_queue_locked(&queue, dwork, delay); |
| |
| k_spin_unlock(&lock, key); |
| |
| SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work, reschedule_for_queue, queue, dwork, delay, ret); |
| |
| return ret; |
| } |
| |
| int k_work_reschedule(struct k_work_delayable *dwork, |
| k_timeout_t delay) |
| { |
| SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work, reschedule, dwork, delay); |
| |
| int ret = k_work_reschedule_for_queue(&k_sys_work_q, dwork, delay); |
| |
| SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work, reschedule, dwork, delay, ret); |
| |
| return ret; |
| } |
| |
| int k_work_cancel_delayable(struct k_work_delayable *dwork) |
| { |
| __ASSERT_NO_MSG(dwork != NULL); |
| |
| SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work, cancel_delayable, dwork); |
| |
| k_spinlock_key_t key = k_spin_lock(&lock); |
| int ret = cancel_delayable_async_locked(dwork); |
| |
| k_spin_unlock(&lock, key); |
| |
| SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work, cancel_delayable, dwork, ret); |
| |
| return ret; |
| } |
| |
| bool k_work_cancel_delayable_sync(struct k_work_delayable *dwork, |
| struct k_work_sync *sync) |
| { |
| __ASSERT_NO_MSG(dwork != NULL); |
| __ASSERT_NO_MSG(sync != NULL); |
| __ASSERT_NO_MSG(!k_is_in_isr()); |
| #ifdef CONFIG_KERNEL_COHERENCE |
| __ASSERT_NO_MSG(arch_mem_coherent(sync)); |
| #endif /* CONFIG_KERNEL_COHERENCE */ |
| |
| SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work, cancel_delayable_sync, dwork, sync); |
| |
| struct z_work_canceller *canceller = &sync->canceller; |
| k_spinlock_key_t key = k_spin_lock(&lock); |
| bool pending = (work_delayable_busy_get_locked(dwork) != 0U); |
| bool need_wait = false; |
| |
| if (pending) { |
| (void)cancel_delayable_async_locked(dwork); |
| need_wait = cancel_sync_locked(&dwork->work, canceller); |
| } |
| |
| k_spin_unlock(&lock, key); |
| |
| if (need_wait) { |
| k_sem_take(&canceller->sem, K_FOREVER); |
| } |
| |
| SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work, cancel_delayable_sync, dwork, sync, pending); |
| return pending; |
| } |
| |
| bool k_work_flush_delayable(struct k_work_delayable *dwork, |
| struct k_work_sync *sync) |
| { |
| __ASSERT_NO_MSG(dwork != NULL); |
| __ASSERT_NO_MSG(sync != NULL); |
| __ASSERT_NO_MSG(!k_is_in_isr()); |
| #ifdef CONFIG_KERNEL_COHERENCE |
| __ASSERT_NO_MSG(arch_mem_coherent(sync)); |
| #endif /* CONFIG_KERNEL_COHERENCE */ |
| |
| SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work, flush_delayable, dwork, sync); |
| |
| struct k_work *work = &dwork->work; |
| struct z_work_flusher *flusher = &sync->flusher; |
| k_spinlock_key_t key = k_spin_lock(&lock); |
| |
| /* If it's idle release the lock and return immediately. */ |
| if (work_busy_get_locked(work) == 0U) { |
| k_spin_unlock(&lock, key); |
| |
| SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work, flush_delayable, dwork, sync, false); |
| |
| return false; |
| } |
| |
| /* If unscheduling did something then submit it. Ignore a |
| * failed submission (e.g. when cancelling). |
| */ |
| if (unschedule_locked(dwork)) { |
| struct k_work_q *queue = dwork->queue; |
| |
| (void)submit_to_queue_locked(work, &queue); |
| } |
| |
| /* Wait for it to finish */ |
| bool need_flush = work_flush_locked(work, flusher); |
| |
| k_spin_unlock(&lock, key); |
| |
| /* If necessary wait until the flusher item completes */ |
| if (need_flush) { |
| k_sem_take(&flusher->sem, K_FOREVER); |
| } |
| |
| SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work, flush_delayable, dwork, sync, need_flush); |
| |
| return need_flush; |
| } |
| |
| #endif /* CONFIG_SYS_CLOCK_EXISTS */ |