blob: 46807bf0739f3da6a466ad22069e5f7dcfecc0e0 [file] [log] [blame]
/*
* Copyright (c) 2021 Andes Technology Corporation
*
* SPDX-License-Identifier: Apache-2.0
*/
#include <zephyr/init.h>
#include <zephyr/kernel.h>
#include <zephyr/arch/cpu.h>
#include <zephyr/linker/linker-defs.h>
#include <zephyr/arch/riscv/csr.h>
#ifndef CONFIG_ASSERT
#define LOG_LEVEL CONFIG_LOG_DEFAULT_LEVEL
#include <zephyr/logging/log.h>
LOG_MODULE_REGISTER(pma_init, LOG_LEVEL);
#endif
/* Programmable PMA mechanism is supported */
#define MMSC_CFG_PPMA BIT(30)
/*
* PMA Configuration (PMACFG) bitfields
*/
/* ETYPE: Entry address matching mode */
#define PMACFG_ETYPE_MASK BIT_MASK(2)
#define PMACFG_ETYPE_OFF 0
#define PMACFG_ETYPE_TOR 1
#define PMACFG_ETYPE_NA4 2
#define PMACFG_ETYPE_NAPOT 3
/* MTYPE: Memory type attribute */
#define PMACFG_MTYPE_MASK (0xF << 2)
/* non-cacheable attributes (bufferable or not) */
#define PMACFG_MTYPE_MEMORY_NOCACHE_BUFFERABLE (3 << 2)
/* cacheable attributes (write-through/back, no/read/write/RW-allocate) */
#define PMACFG_MTYPE_MEMORY_WBCACHE_RWALLOC (11 << 2)
/* pmaaddr is 4-byte granularity in each mode */
#define TO_PMA_ADDR(addr) ((addr) >> 2)
/* The base address is aligned to size */
#define NAPOT_BASE(start, size) TO_PMA_ADDR((start) & ~((size) - 1))
/* The encoding of size is 0b01...1, (change the leading bit of bitmask to 0) */
#define NAPOT_SIZE(size) TO_PMA_ADDR(((size) - 1) >> 1)
#define NA4_ENCODING(start) TO_PMA_ADDR(start)
#define NAPOT_ENCODING(start, size) (NAPOT_BASE(start, size) \
| NAPOT_SIZE(size))
#ifdef CONFIG_64BIT
/* In riscv64, CSR pmacfg number are even number (0, 2, ...) */
# define PMACFG_NUM(index) ((index / RV_REGSIZE) * 2)
#else
# define PMACFG_NUM(index) (index / RV_REGSIZE)
#endif
#define PMACFG_SHIFT(index) ((index % RV_REGSIZE) * 8)
struct pma_region_attr {
/* Attributes belonging to pmacfg{i} */
uint8_t pmacfg;
};
struct pma_region {
unsigned long start;
unsigned long size;
struct pma_region_attr attr;
};
/*
* Write value to CSRs pmaaddr{i}
*/
static void write_pmaaddr_csr(const uint32_t index, unsigned long value)
{
#define SWITCH_CASE_PMAADDR_WRITE(x) \
case (x): \
csr_write(NDS_PMAADDR##x, value); break;
switch (index) {
FOR_EACH(SWITCH_CASE_PMAADDR_WRITE, (;), 0, 1, 2, 3, 4, 5, 6, 7,
8, 9, 10, 11, 12, 13, 14, 15);
}
}
/*
* Write value to pma{i}cfg entry which are packed into CSRs pmacfg{j}
*/
static void write_pmacfg_entry(const uint32_t entry_index, uint8_t entry_value)
{
/* 1-byte pma{i}cfg entries are packed into XLEN-byte CSRs pmacfg{j} */
uint32_t index = PMACFG_NUM(entry_index);
uint8_t shift = PMACFG_SHIFT(entry_index);
unsigned long pmacfg = 0;
#define SWITCH_CASE_PMACFG_READ(x) \
case (x): \
pmacfg = csr_read(NDS_PMACFG##x); break;
switch (index) {
FOR_EACH(SWITCH_CASE_PMACFG_READ, (;), 0, 1, 2, 3);
}
/* clear old value in pmacfg entry */
pmacfg &= ~(0xFF << shift);
/* set new value to pmacfg entry value */
pmacfg |= entry_value << shift;
#define SWITCH_CASE_PMACFG_WRITE(x) \
case (x): \
csr_write(NDS_PMACFG##x, pmacfg); break;
switch (index) {
FOR_EACH(SWITCH_CASE_PMACFG_WRITE, (;), 0, 1, 2, 3);
}
}
/*
* This internal function performs PMA region initialization.
*
* Note:
* The caller must provide a valid region index.
*/
static void region_init(const uint32_t index,
const struct pma_region *region_conf)
{
unsigned long pmaaddr;
uint8_t pmacfg;
if (region_conf->size == 4) {
pmaaddr = NA4_ENCODING(region_conf->start);
pmacfg = region_conf->attr.pmacfg | PMACFG_ETYPE_NA4;
} else {
pmaaddr = NAPOT_ENCODING(region_conf->start, region_conf->size);
pmacfg = region_conf->attr.pmacfg | PMACFG_ETYPE_NAPOT;
}
write_pmaaddr_csr(index, pmaaddr);
write_pmacfg_entry(index, pmacfg);
}
/*
* This internal function performs run-time sanity check for
* PMA region start address and size.
*/
static int pma_region_is_valid(const struct pma_region *region)
{
/* Region size must greater or equal to the minimum PMA region size */
if (region->size < CONFIG_SOC_ANDES_V5_PMA_REGION_MIN_ALIGN_AND_SIZE) {
return -EINVAL;
}
/* Region size must be power-of-two */
if (region->size & (region->size - 1)) {
return -EINVAL;
}
/* Start address of the region must align with size */
if (region->start & (region->size - 1)) {
return -EINVAL;
}
return 0;
}
#ifdef CONFIG_NOCACHE_MEMORY
static void configure_nocache_region(void)
{
const struct pma_region nocache_region = {
.start = (unsigned long)&_nocache_ram_start,
.size = (unsigned long)&_nocache_ram_size,
.attr = {PMACFG_MTYPE_MEMORY_NOCACHE_BUFFERABLE},
};
if (pma_region_is_valid(&nocache_region)) {
/* Skip PMA configuration if nocache region size is 0 */
if (nocache_region.size != 0) {
__ASSERT(0, "Configuring PMA region of nocache region "
"failed\n");
}
} else {
/* Initialize nocache region at PMA region 0 */
region_init(0, &nocache_region);
}
}
#endif /* CONFIG_NOCACHE_MEMORY */
/*
* @brief Init PMA CSRs of each CPU core
*
* In SMP, each CPU has it's own PMA CSR and PMA CSR only affect one CPU.
* We should configure CSRs of all CPUs to make memory attribute
* (e.g. uncacheable) affects all CPUs.
*/
void pma_init_per_core(void)
{
#ifdef CONFIG_NOCACHE_MEMORY
configure_nocache_region();
#endif /* CONFIG_NOCACHE_MEMORY */
}
static int pma_init(void)
{
unsigned long mmsc_cfg;
mmsc_cfg = csr_read(NDS_MMSC_CFG);
if (!(mmsc_cfg & MMSC_CFG_PPMA)) {
/* This CPU doesn't support PMA */
__ASSERT(0, "CPU doesn't support PMA. "
"Please disable CONFIG_SOC_ANDES_V5_PMA\n");
#ifndef CONFIG_ASSERT
LOG_ERR("CPU doesn't support PMA. "
"Please disable CONFIG_SOC_ANDES_V5_PMA");
#endif
return -ENODEV;
}
pma_init_per_core();
return 0;
}
SYS_INIT(pma_init, PRE_KERNEL_2, CONFIG_KERNEL_INIT_PRIORITY_DEFAULT);