| /* |
| * Copyright (c) 2010-2015 Wind River Systems, Inc. |
| * |
| * Redistribution and use in source and binary forms, with or without |
| * modification, are permitted provided that the following conditions are met: |
| * |
| * 1) Redistributions of source code must retain the above copyright notice, |
| * this list of conditions and the following disclaimer. |
| * |
| * 2) Redistributions in binary form must reproduce the above copyright notice, |
| * this list of conditions and the following disclaimer in the documentation |
| * and/or other materials provided with the distribution. |
| * |
| * 3) Neither the name of Wind River Systems nor the names of its contributors |
| * may be used to endorse or promote products derived from this software without |
| * specific prior written permission. |
| * |
| * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" |
| * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
| * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE |
| * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE |
| * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR |
| * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF |
| * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS |
| * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN |
| * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) |
| * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE |
| * POSSIBILITY OF SUCH DAMAGE. |
| */ |
| |
| /** @file |
| * |
| * @brief Nanokernel dynamic-size LIFO queue object |
| * |
| * This module provides the nanokernel LIFO object implementation, including |
| * the following APIs: |
| * |
| * nano_lifo_init |
| * nano_fiber_lifo_put, nano_task_lifo_put, nano_isr_lifo_put |
| * nano_fiber_lifo_get, nano_task_lifo_get, nano_isr_lifo_get |
| * nano_fiber_lifo_get_wait, nano_task_lifo_get_wait |
| */ |
| |
| /** INTERNAL |
| * |
| * In some cases the compiler "alias" attribute is used to map two or more |
| * APIs to the same function, since they have identical implementations. |
| */ |
| |
| #include <nano_private.h> |
| #include <toolchain.h> |
| #include <sections.h> |
| #include <wait_q.h> |
| |
| /** INTERNAL |
| * |
| * Although the existing implementation will support invocation from an ISR |
| * context, for future flexibility, this API will be restricted from ISR |
| * level invocation. |
| */ |
| void nano_lifo_init(struct nano_lifo *lifo) |
| { |
| lifo->list = (void *) 0; |
| _nano_wait_q_init(&lifo->wait_q); |
| } |
| |
| FUNC_ALIAS(_lifo_put_non_preemptible, nano_isr_lifo_put, void); |
| FUNC_ALIAS(_lifo_put_non_preemptible, nano_fiber_lifo_put, void); |
| |
| /** INTERNAL |
| * |
| * This function is capable of supporting invocations from both a fiber and an |
| * ISR context. However, the nano_isr_lifo_put and nano_fiber_lifo_put aliases |
| * are created to support any required implementation differences in the future |
| * without introducing a source code migration issue. |
| */ |
| void _lifo_put_non_preemptible(struct nano_lifo *lifo, void *data) |
| { |
| struct tcs *tcs; |
| unsigned int imask; |
| |
| imask = irq_lock(); |
| tcs = _nano_wait_q_remove(&lifo->wait_q); |
| if (tcs) { |
| _nano_timeout_abort(tcs); |
| fiberRtnValueSet(tcs, (unsigned int) data); |
| } else { |
| *(void **) data = lifo->list; |
| lifo->list = data; |
| } |
| |
| irq_unlock(imask); |
| } |
| |
| void nano_task_lifo_put(struct nano_lifo *lifo, void *data) |
| { |
| struct tcs *tcs; |
| unsigned int imask; |
| |
| imask = irq_lock(); |
| tcs = _nano_wait_q_remove(&lifo->wait_q); |
| if (tcs) { |
| _nano_timeout_abort(tcs); |
| fiberRtnValueSet(tcs, (unsigned int) data); |
| _Swap(imask); |
| return; |
| } else { |
| *(void **) data = lifo->list; |
| lifo->list = data; |
| } |
| |
| irq_unlock(imask); |
| } |
| |
| FUNC_ALIAS(_lifo_get, nano_isr_lifo_get, void *); |
| FUNC_ALIAS(_lifo_get, nano_fiber_lifo_get, void *); |
| FUNC_ALIAS(_lifo_get, nano_task_lifo_get, void *); |
| |
| /** INTERNAL |
| * |
| * This function is capable of supporting invocations from fiber, task, and ISR |
| * contexts. However, the nano_isr_lifo_get, nano_task_lifo_get, and |
| * nano_fiber_lifo_get aliases are created to support any required |
| * implementation differences in the future without introducing a source code |
| * migration issue. |
| */ |
| void *_lifo_get(struct nano_lifo *lifo) |
| { |
| void *data; |
| unsigned int imask; |
| |
| imask = irq_lock(); |
| |
| data = lifo->list; |
| if (data) { |
| lifo->list = *(void **) data; |
| } |
| |
| irq_unlock(imask); |
| |
| return data; |
| } |
| |
| /** INTERNAL |
| * |
| * There exists a separate nano_task_lifo_get_wait() implementation since a |
| * task cannot pend on a nanokernel object. Instead, tasks will poll |
| * the lifo object. |
| */ |
| void *nano_fiber_lifo_get_wait(struct nano_lifo *lifo ) |
| { |
| void *data; |
| unsigned int imask; |
| |
| imask = irq_lock(); |
| |
| if (!lifo->list) { |
| _nano_wait_q_put(&lifo->wait_q); |
| data = (void *) _Swap(imask); |
| } else { |
| data = lifo->list; |
| lifo->list = *(void **) data; |
| irq_unlock(imask); |
| } |
| |
| return data; |
| } |
| |
| void *nano_task_lifo_get_wait(struct nano_lifo *lifo) |
| { |
| void *data; |
| unsigned int imask; |
| |
| /* spin until data is put onto the LIFO */ |
| |
| while (1) { |
| imask = irq_lock(); |
| |
| /* |
| * Predict that the branch will be taken to break out of the loop. |
| * There is little cost to a misprediction since that leads to idle. |
| */ |
| |
| if (likely(lifo->list)) |
| break; |
| |
| /* see explanation in nano_stack.c:nano_task_stack_pop_wait() */ |
| |
| nano_cpu_atomic_idle(imask); |
| } |
| |
| data = lifo->list; |
| lifo->list = *(void **) data; |
| |
| irq_unlock(imask); |
| |
| return data; |
| } |
| |
| /* |
| * @brief Get first element from lifo and panic if NULL |
| * |
| * Get the first element from the specified lifo but generate a fatal error |
| * if the element is NULL. |
| * |
| * @param lifo LIFO from which to receive. |
| * |
| * @return Pointer to first element in the list |
| * |
| * \NOMANUAL |
| */ |
| void *_nano_fiber_lifo_get_panic(struct nano_lifo *lifo) |
| { |
| void *element; |
| |
| element = nano_fiber_lifo_get(lifo); |
| |
| if (element == NULL) { |
| _NanoFatalErrorHandler(_NANO_ERR_ALLOCATION_FAIL, &_default_esf); |
| } |
| |
| return element; |
| } |
| |
| #ifdef CONFIG_NANO_TIMEOUTS |
| |
| void *nano_fiber_lifo_get_wait_timeout(struct nano_lifo *lifo, |
| int32_t timeout_in_ticks) |
| { |
| unsigned int key = irq_lock(); |
| void *data; |
| |
| if (!lifo->list) { |
| if (unlikely(TICKS_NONE == timeout_in_ticks)) { |
| irq_unlock(key); |
| return NULL; |
| } |
| if (likely(timeout_in_ticks != TICKS_UNLIMITED)) { |
| _nano_timeout_add(_nanokernel.current, &lifo->wait_q, |
| timeout_in_ticks); |
| } |
| _nano_wait_q_put(&lifo->wait_q); |
| data = (void *)_Swap(key); |
| } else { |
| data = lifo->list; |
| lifo->list = *(void **)data; |
| irq_unlock(key); |
| } |
| |
| return data; |
| } |
| |
| void *nano_task_lifo_get_wait_timeout(struct nano_lifo *lifo, |
| int32_t timeout_in_ticks) |
| { |
| int64_t cur_ticks, limit; |
| unsigned int key; |
| void *data; |
| |
| if (unlikely(TICKS_UNLIMITED == timeout_in_ticks)) { |
| return nano_task_lifo_get_wait(lifo); |
| } |
| |
| if (unlikely(TICKS_NONE == timeout_in_ticks)) { |
| return nano_task_lifo_get(lifo); |
| } |
| |
| key = irq_lock(); |
| cur_ticks = nano_tick_get(); |
| limit = cur_ticks + timeout_in_ticks; |
| |
| while (cur_ticks < limit) { |
| |
| /* |
| * Predict that the branch will be taken to break out of the loop. |
| * There is little cost to a misprediction since that leads to idle. |
| */ |
| |
| if (likely(lifo->list)) { |
| data = lifo->list; |
| lifo->list = *(void **)data; |
| irq_unlock(key); |
| return data; |
| } |
| |
| /* see explanation in nano_stack.c:nano_task_stack_pop_wait() */ |
| |
| nano_cpu_atomic_idle(key); |
| |
| key = irq_lock(); |
| cur_ticks = nano_tick_get(); |
| } |
| |
| irq_unlock(key); |
| return NULL; |
| } |
| #endif /* CONFIG_NANO_TIMEOUTS */ |