blob: 786bc8fc156167da485004d4377c67d34f085493 [file] [log] [blame]
/* ring_buffer.c: Simple ring buffer API */
/*
* Copyright (c) 2015 Intel Corporation
*
* SPDX-License-Identifier: Apache-2.0
*/
#include <sys/ring_buffer.h>
#include <string.h>
/* LCOV_EXCL_START */
/* The weak function used to allow overwriting it in the test and trigger
* rewinding earlier.
*/
uint32_t __weak ring_buf_get_rewind_threshold(void)
{
return RING_BUFFER_MAX_SIZE;
}
/* LCOV_EXCL_STOP */
/**
* Internal data structure for a buffer header.
*
* We want all of this to fit in a single uint32_t. Every item stored in the
* ring buffer will be one of these headers plus any extra data supplied
*/
struct ring_element {
uint32_t type :16; /**< Application-specific */
uint32_t length :8; /**< length in 32-bit chunks */
uint32_t value :8; /**< Room for small integral values */
};
static uint32_t mod(struct ring_buf *buf, uint32_t val)
{
return likely(buf->mask) ? val & buf->mask : val % buf->size;
}
/* Check if indexes did not progress too far (too close to 32-bit wrapping).
* If so, then reduce all indexes by an arbitrary value.
*/
static void item_indexes_rewind(struct ring_buf *buf)
{
uint32_t rewind;
uint32_t threshold = ring_buf_get_rewind_threshold();
if (buf->head < threshold) {
return;
}
rewind = buf->size * (threshold / buf->size);
k_spinlock_key_t key = k_spin_lock(&buf->lock);
buf->tail -= rewind;
buf->head -= rewind;
k_spin_unlock(&buf->lock, key);
}
/* Check if indexes did not progresses too far (too close to 32-bit wrapping).
* If so, then rewind all indexes by an arbitrary value. For byte mode temporary
* indexes must also be reduced.
*/
static void byte_indexes_rewind(struct ring_buf *buf)
{
uint32_t rewind;
uint32_t threshold = ring_buf_get_rewind_threshold();
/* Checking head since it is the smallest index. */
if (buf->head < threshold) {
return;
}
rewind = buf->size * (threshold / buf->size);
k_spinlock_key_t key = k_spin_lock(&buf->lock);
buf->tail -= rewind;
buf->head -= rewind;
buf->misc.byte_mode.tmp_head -= rewind;
buf->misc.byte_mode.tmp_tail -= rewind;
k_spin_unlock(&buf->lock, key);
}
int ring_buf_item_put(struct ring_buf *buf, uint16_t type, uint8_t value,
uint32_t *data, uint8_t size32)
{
uint32_t i, space, index, rc;
space = ring_buf_space_get(buf);
if (space >= (size32 + 1)) {
struct ring_element *header =
(struct ring_element *)&buf->buf.buf32[mod(buf, buf->tail)];
header->type = type;
header->length = size32;
header->value = value;
if (likely(buf->mask)) {
for (i = 0U; i < size32; ++i) {
index = (i + buf->tail + 1) & buf->mask;
buf->buf.buf32[index] = data[i];
}
} else {
for (i = 0U; i < size32; ++i) {
index = (i + buf->tail + 1) % buf->size;
buf->buf.buf32[index] = data[i];
}
}
buf->tail = buf->tail + size32 + 1;
rc = 0U;
} else {
buf->misc.item_mode.dropped_put_count++;
rc = -EMSGSIZE;
}
return rc;
}
int ring_buf_item_get(struct ring_buf *buf, uint16_t *type, uint8_t *value,
uint32_t *data, uint8_t *size32)
{
struct ring_element *header;
uint32_t i, index;
if (ring_buf_is_empty(buf)) {
return -EAGAIN;
}
header = (struct ring_element *) &buf->buf.buf32[mod(buf, buf->head)];
if (header->length > *size32) {
*size32 = header->length;
return -EMSGSIZE;
}
*size32 = header->length;
*type = header->type;
*value = header->value;
if (likely(buf->mask)) {
for (i = 0U; i < header->length; ++i) {
index = (i + buf->head + 1) & buf->mask;
data[i] = buf->buf.buf32[index];
}
} else {
for (i = 0U; i < header->length; ++i) {
index = (i + buf->head + 1) % buf->size;
data[i] = buf->buf.buf32[index];
}
}
buf->head = buf->head + header->length + 1;
item_indexes_rewind(buf);
return 0;
}
/** @brief Wraps index if it exceeds the limit.
*
* @param val Value
* @param max Max.
*
* @return value % max.
*/
static inline uint32_t wrap(uint32_t val, uint32_t max)
{
return val >= max ? (val - max) : val;
}
uint32_t ring_buf_put_claim(struct ring_buf *buf, uint8_t **data, uint32_t size)
{
uint32_t space, trail_size, allocated, tmp_trail_mod;
tmp_trail_mod = mod(buf, buf->misc.byte_mode.tmp_tail);
space = (buf->head + buf->size) - buf->misc.byte_mode.tmp_tail;
trail_size = buf->size - tmp_trail_mod;
/* Limit requested size to available size. */
size = MIN(size, space);
trail_size = buf->size - (tmp_trail_mod);
/* Limit allocated size to trail size. */
allocated = MIN(trail_size, size);
*data = &buf->buf.buf8[tmp_trail_mod];
buf->misc.byte_mode.tmp_tail =
buf->misc.byte_mode.tmp_tail + allocated;
return allocated;
}
int ring_buf_put_finish(struct ring_buf *buf, uint32_t size)
{
if ((buf->tail + size) > (buf->head + buf->size)) {
return -EINVAL;
}
buf->tail += size;
buf->misc.byte_mode.tmp_tail = buf->tail;
return 0;
}
uint32_t ring_buf_put(struct ring_buf *buf, const uint8_t *data, uint32_t size)
{
uint8_t *dst;
uint32_t partial_size;
uint32_t total_size = 0U;
int err;
do {
partial_size = ring_buf_put_claim(buf, &dst, size);
memcpy(dst, data, partial_size);
total_size += partial_size;
size -= partial_size;
data += partial_size;
} while (size && partial_size);
err = ring_buf_put_finish(buf, total_size);
__ASSERT_NO_MSG(err == 0);
return total_size;
}
uint32_t ring_buf_get_claim(struct ring_buf *buf, uint8_t **data, uint32_t size)
{
uint32_t space, granted_size, trail_size, tmp_head_mod;
tmp_head_mod = mod(buf, buf->misc.byte_mode.tmp_head);
space = buf->tail - buf->misc.byte_mode.tmp_head;
trail_size = buf->size - tmp_head_mod;
/* Limit requested size to available size. */
granted_size = MIN(size, space);
/* Limit allocated size to trail size. */
granted_size = MIN(trail_size, granted_size);
*data = &buf->buf.buf8[tmp_head_mod];
buf->misc.byte_mode.tmp_head += granted_size;
return granted_size;
}
int ring_buf_get_finish(struct ring_buf *buf, uint32_t size)
{
if ((buf->head + size) > buf->tail) {
return -EINVAL;
}
buf->head += size;
buf->misc.byte_mode.tmp_head = buf->head;
byte_indexes_rewind(buf);
return 0;
}
uint32_t ring_buf_get(struct ring_buf *buf, uint8_t *data, uint32_t size)
{
uint8_t *src;
uint32_t partial_size;
uint32_t total_size = 0U;
int err;
do {
partial_size = ring_buf_get_claim(buf, &src, size);
memcpy(data, src, partial_size);
total_size += partial_size;
size -= partial_size;
data += partial_size;
} while (size && partial_size);
err = ring_buf_get_finish(buf, total_size);
__ASSERT_NO_MSG(err == 0);
return total_size;
}