blob: eaffb6ce763b378dc70de4e9e2e668c6915b57fb [file] [log] [blame]
Benjamin Walsh456c6da2016-09-02 18:55:39 -04001/*
2 * Copyright (c) 2010-2014 Wind River Systems, Inc.
3 *
David B. Kinderac74d8b2017-01-18 17:01:01 -08004 * SPDX-License-Identifier: Apache-2.0
Benjamin Walsh456c6da2016-09-02 18:55:39 -04005 */
6
7/**
8 * @file
Anas Nashifcb888e62016-12-18 09:42:55 -05009 * @brief Kernel initialization module
Benjamin Walsh456c6da2016-09-02 18:55:39 -040010 *
Anas Nashifdc3d73b2016-12-19 20:25:56 -050011 * This module contains routines that are used to initialize the kernel.
Benjamin Walsh456c6da2016-09-02 18:55:39 -040012 */
13
Benjamin Walshf6ca7de2016-11-08 10:36:50 -050014#include <offsets_short.h>
Gerard Marull-Paretascffefc82022-05-06 11:04:23 +020015#include <zephyr/kernel.h>
16#include <zephyr/sys/printk.h>
17#include <zephyr/debug/stack.h>
Flavio Ceoline7bd10a2023-10-06 22:38:53 +000018#include <zephyr/random/random.h>
Gerard Marull-Paretascffefc82022-05-06 11:04:23 +020019#include <zephyr/linker/sections.h>
20#include <zephyr/toolchain.h>
21#include <zephyr/kernel_structs.h>
22#include <zephyr/device.h>
23#include <zephyr/init.h>
24#include <zephyr/linker/linker-defs.h>
Benjamin Walshb4b108d2016-10-13 10:31:48 -040025#include <ksched.h>
Anas Nashif3ca50f52024-02-23 10:53:01 -050026#include <kthread.h>
Mahavir Jainacea2412016-12-02 21:48:39 +053027#include <string.h>
Gerard Marull-Paretascffefc82022-05-06 11:04:23 +020028#include <zephyr/sys/dlist.h>
Andy Ross245b54e2018-02-08 09:10:46 -080029#include <kernel_internal.h>
Gerard Marull-Paretascffefc82022-05-06 11:04:23 +020030#include <zephyr/drivers/entropy.h>
31#include <zephyr/logging/log_ctrl.h>
32#include <zephyr/tracing/tracing.h>
Flavio Ceolinb3d92022018-09-17 15:56:06 -070033#include <stdbool.h>
Gerard Marull-Paretascffefc82022-05-06 11:04:23 +020034#include <zephyr/debug/gcov.h>
Andrew Boie468efad2020-05-12 16:20:14 -070035#include <kswap.h>
Gerard Marull-Paretascffefc82022-05-06 11:04:23 +020036#include <zephyr/timing/timing.h>
37#include <zephyr/logging/log.h>
Jordan Yatesdb3d51b2022-03-12 21:10:42 +100038#include <zephyr/pm/device_runtime.h>
Ederson de Souzaeeebb4d2024-01-05 13:29:20 -080039#include <zephyr/internal/syscall_handler.h>
Krzysztof Chruscinski3ed80832020-11-26 19:32:34 +010040LOG_MODULE_REGISTER(os, CONFIG_KERNEL_LOG_LEVEL);
Anas Nashif57554052018-03-03 02:31:05 -060041
Kumar Gala0722b6f2022-10-21 09:34:08 -050042BUILD_ASSERT(CONFIG_MP_NUM_CPUS == CONFIG_MP_MAX_NUM_CPUS,
43 "CONFIG_MP_NUM_CPUS and CONFIG_MP_MAX_NUM_CPUS need to be set the same");
44
Krzysztof Chruscinski7dcff6e2021-04-16 15:16:00 +020045/* the only struct z_kernel instance */
Qipeng Zhafa973d12023-05-15 13:29:18 +080046__pinned_bss
Krzysztof Chruscinski7dcff6e2021-04-16 15:16:00 +020047struct z_kernel _kernel;
48
Flavio Ceolin4f299302023-05-30 11:49:45 -070049__pinned_bss
50atomic_t _cpus_active;
51
Benjamin Walsh456c6da2016-09-02 18:55:39 -040052/* init/main and idle threads */
Daniel Leungebbfde92021-07-12 13:47:48 -070053K_THREAD_PINNED_STACK_DEFINE(z_main_stack, CONFIG_MAIN_STACK_SIZE);
Andrew Boiefe031612019-09-21 17:54:37 -070054struct k_thread z_main_thread;
Andrew Boie80a0d9d2020-03-12 15:37:29 -070055
56#ifdef CONFIG_MULTITHREADING
Daniel Leung660d1472021-03-25 16:05:15 -070057__pinned_bss
Kumar Galac778eb22022-10-12 10:55:36 -050058struct k_thread z_idle_threads[CONFIG_MP_MAX_NUM_CPUS];
Daniel Leung660d1472021-03-25 16:05:15 -070059
60static K_KERNEL_PINNED_STACK_ARRAY_DEFINE(z_idle_stacks,
Kumar Galac778eb22022-10-12 10:55:36 -050061 CONFIG_MP_MAX_NUM_CPUS,
Daniel Leung660d1472021-03-25 16:05:15 -070062 CONFIG_IDLE_STACK_SIZE);
Anas Nashif3ca50f52024-02-23 10:53:01 -050063
64static void z_init_static_threads(void)
65{
66 STRUCT_SECTION_FOREACH(_static_thread_data, thread_data) {
67 z_setup_new_thread(
68 thread_data->init_thread,
69 thread_data->init_stack,
70 thread_data->init_stack_size,
71 thread_data->init_entry,
72 thread_data->init_p1,
73 thread_data->init_p2,
74 thread_data->init_p3,
75 thread_data->init_prio,
76 thread_data->init_options,
77 thread_data->init_name);
78
79 thread_data->init_thread->init_data = thread_data;
80 }
81
82#ifdef CONFIG_USERSPACE
83 STRUCT_SECTION_FOREACH(k_object_assignment, pos) {
84 for (int i = 0; pos->objects[i] != NULL; i++) {
85 k_object_access_grant(pos->objects[i],
86 pos->thread);
87 }
88 }
Simon Heinbcd1d192024-03-08 12:00:10 +010089#endif /* CONFIG_USERSPACE */
Anas Nashif3ca50f52024-02-23 10:53:01 -050090
91 /*
92 * Non-legacy static threads may be started immediately or
93 * after a previously specified delay. Even though the
94 * scheduler is locked, ticks can still be delivered and
95 * processed. Take a sched lock to prevent them from running
96 * until they are all started.
97 *
98 * Note that static threads defined using the legacy API have a
99 * delay of K_FOREVER.
100 */
101 k_sched_lock();
102 STRUCT_SECTION_FOREACH(_static_thread_data, thread_data) {
103 k_timeout_t init_delay = Z_THREAD_INIT_DELAY(thread_data);
104
105 if (!K_TIMEOUT_EQ(init_delay, K_FOREVER)) {
106 thread_schedule_new(thread_data->init_thread,
107 init_delay);
108 }
109 }
110 k_sched_unlock();
111}
112#else
113#define z_init_static_threads() do { } while (false)
Andrew Boie80a0d9d2020-03-12 15:37:29 -0700114#endif /* CONFIG_MULTITHREADING */
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400115
Gerard Marull-Paretas83123932022-10-04 11:25:37 +0200116extern const struct init_entry __init_start[];
117extern const struct init_entry __init_EARLY_start[];
118extern const struct init_entry __init_PRE_KERNEL_1_start[];
119extern const struct init_entry __init_PRE_KERNEL_2_start[];
120extern const struct init_entry __init_POST_KERNEL_start[];
121extern const struct init_entry __init_APPLICATION_start[];
122extern const struct init_entry __init_end[];
123
Gerard Marull-Paretas495245a2022-10-04 11:52:18 +0200124enum init_level {
125 INIT_LEVEL_EARLY = 0,
126 INIT_LEVEL_PRE_KERNEL_1,
127 INIT_LEVEL_PRE_KERNEL_2,
128 INIT_LEVEL_POST_KERNEL,
129 INIT_LEVEL_APPLICATION,
130#ifdef CONFIG_SMP
131 INIT_LEVEL_SMP,
Simon Heinbcd1d192024-03-08 12:00:10 +0100132#endif /* CONFIG_SMP */
Gerard Marull-Paretas495245a2022-10-04 11:52:18 +0200133};
134
Gerard Marull-Paretas83123932022-10-04 11:25:37 +0200135#ifdef CONFIG_SMP
136extern const struct init_entry __init_SMP_start[];
Simon Heinbcd1d192024-03-08 12:00:10 +0100137#endif /* CONFIG_SMP */
Gerard Marull-Paretas83123932022-10-04 11:25:37 +0200138
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400139/*
140 * storage space for the interrupt stack
141 *
Anas Nashifdc3d73b2016-12-19 20:25:56 -0500142 * Note: This area is used as the system stack during kernel initialization,
143 * since the kernel hasn't yet set up its own stack areas. The dual purposing
144 * of this area is safe since interrupts are disabled until the kernel context
145 * switches to the init thread.
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400146 */
Daniel Leung660d1472021-03-25 16:05:15 -0700147K_KERNEL_PINNED_STACK_ARRAY_DEFINE(z_interrupt_stacks,
Kumar Galac778eb22022-10-12 10:55:36 -0500148 CONFIG_MP_MAX_NUM_CPUS,
Daniel Leung660d1472021-03-25 16:05:15 -0700149 CONFIG_ISR_STACK_SIZE);
Andy Ross780ba232018-01-29 09:20:18 -0800150
Peter Mitsis96cb05c2016-09-15 12:37:58 -0400151extern void idle(void *unused1, void *unused2, void *unused3);
152
Peter Mitsis6df8efe2023-05-11 14:06:46 -0400153#ifdef CONFIG_OBJ_CORE_SYSTEM
154static struct k_obj_type obj_type_cpu;
155static struct k_obj_type obj_type_kernel;
Peter Mitsise6f10902023-06-01 12:16:40 -0400156
157#ifdef CONFIG_OBJ_CORE_STATS_SYSTEM
158static struct k_obj_core_stats_desc cpu_stats_desc = {
159 .raw_size = sizeof(struct k_cycle_stats),
160 .query_size = sizeof(struct k_thread_runtime_stats),
161 .raw = z_cpu_stats_raw,
162 .query = z_cpu_stats_query,
163 .reset = NULL,
164 .disable = NULL,
165 .enable = NULL,
166};
167
168static struct k_obj_core_stats_desc kernel_stats_desc = {
169 .raw_size = sizeof(struct k_cycle_stats) * CONFIG_MP_MAX_NUM_CPUS,
170 .query_size = sizeof(struct k_thread_runtime_stats),
171 .raw = z_kernel_stats_raw,
172 .query = z_kernel_stats_query,
173 .reset = NULL,
174 .disable = NULL,
175 .enable = NULL,
176};
Simon Heinbcd1d192024-03-08 12:00:10 +0100177#endif /* CONFIG_OBJ_CORE_STATS_SYSTEM */
178#endif /* CONFIG_OBJ_CORE_SYSTEM */
Carles Cuficb0cf9f2017-01-10 10:57:38 +0100179
Andrew Boiefe228a82019-06-11 12:49:32 -0700180/* LCOV_EXCL_START
181 *
182 * This code is called so early in the boot process that code coverage
183 * doesn't work properly. In addition, not all arches call this code,
184 * some like x86 do this with optimized assembly
185 */
186
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400187/**
Nicolas Pitre678b76e2022-02-10 13:54:49 -0500188 * @brief equivalent of memset() for early boot usage
189 *
190 * Architectures that can't safely use the regular (optimized) memset very
191 * early during boot because e.g. hardware isn't yet sufficiently initialized
192 * may override this with their own safe implementation.
193 */
194__boot_func
195void __weak z_early_memset(void *dst, int c, size_t n)
196{
197 (void) memset(dst, c, n);
198}
199
200/**
201 * @brief equivalent of memcpy() for early boot usage
202 *
203 * Architectures that can't safely use the regular (optimized) memcpy very
204 * early during boot because e.g. hardware isn't yet sufficiently initialized
205 * may override this with their own safe implementation.
206 */
207__boot_func
208void __weak z_early_memcpy(void *dst, const void *src, size_t n)
209{
210 (void) memcpy(dst, src, n);
211}
212
213/**
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400214 * @brief Clear BSS
215 *
216 * This routine clears the BSS region, so all bytes are 0.
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400217 */
Daniel Leung660d1472021-03-25 16:05:15 -0700218__boot_func
Patrik Flykt4344e272019-03-08 14:19:05 -0700219void z_bss_zero(void)
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400220{
Alexander Razinkovd2c101d2023-10-27 18:33:15 +0300221 if (IS_ENABLED(CONFIG_SKIP_BSS_CLEAR)) {
Andy Ross5722da72022-08-09 17:55:14 -0700222 return;
223 }
224
Nicolas Pitre678b76e2022-02-10 13:54:49 -0500225 z_early_memset(__bss_start, 0, __bss_end - __bss_start);
Martí Bolívar6e8775f2020-05-11 11:56:08 -0700226#if DT_NODE_HAS_STATUS(DT_CHOSEN(zephyr_ccm), okay)
Nicolas Pitre678b76e2022-02-10 13:54:49 -0500227 z_early_memset(&__ccm_bss_start, 0,
228 (uintptr_t) &__ccm_bss_end
229 - (uintptr_t) &__ccm_bss_start);
Erwin Rol1dc41d12017-10-05 01:22:32 +0200230#endif
Martí Bolívar6e8775f2020-05-11 11:56:08 -0700231#if DT_NODE_HAS_STATUS(DT_CHOSEN(zephyr_dtcm), okay)
Nicolas Pitre678b76e2022-02-10 13:54:49 -0500232 z_early_memset(&__dtcm_bss_start, 0,
233 (uintptr_t) &__dtcm_bss_end
234 - (uintptr_t) &__dtcm_bss_start);
Alexander Wachterb4c5f4b2019-07-03 14:19:29 +0200235#endif
Immo Birnbaumda288292022-01-21 12:38:30 +0100236#if DT_NODE_HAS_STATUS(DT_CHOSEN(zephyr_ocm), okay)
Nicolas Pitre36173982022-02-22 00:00:52 -0500237 z_early_memset(&__ocm_bss_start, 0,
238 (uintptr_t) &__ocm_bss_end
239 - (uintptr_t) &__ocm_bss_start);
Immo Birnbaumda288292022-01-21 12:38:30 +0100240#endif
Adithya Baglody91c5b842018-11-13 16:57:45 +0530241#ifdef CONFIG_CODE_DATA_RELOCATION
242 extern void bss_zeroing_relocation(void);
243
244 bss_zeroing_relocation();
245#endif /* CONFIG_CODE_DATA_RELOCATION */
Adithya Baglody71e90f92018-08-29 16:44:16 +0530246#ifdef CONFIG_COVERAGE_GCOV
Nicolas Pitre678b76e2022-02-10 13:54:49 -0500247 z_early_memset(&__gcov_bss_start, 0,
248 ((uintptr_t) &__gcov_bss_end - (uintptr_t) &__gcov_bss_start));
Simon Heinbcd1d192024-03-08 12:00:10 +0100249#endif /* CONFIG_COVERAGE_GCOV */
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400250}
251
Daniel Leungd8127282021-02-24 10:18:34 -0800252#ifdef CONFIG_LINKER_USE_BOOT_SECTION
253/**
254 * @brief Clear BSS within the bot region
255 *
256 * This routine clears the BSS within the boot region.
257 * This is separate from z_bss_zero() as boot region may
258 * contain symbols required for the boot process before
259 * paging is initialized.
260 */
261__boot_func
262void z_bss_zero_boot(void)
263{
Nicolas Pitre678b76e2022-02-10 13:54:49 -0500264 z_early_memset(&lnkr_boot_bss_start, 0,
265 (uintptr_t)&lnkr_boot_bss_end
266 - (uintptr_t)&lnkr_boot_bss_start);
Daniel Leungd8127282021-02-24 10:18:34 -0800267}
268#endif /* CONFIG_LINKER_USE_BOOT_SECTION */
269
Daniel Leung1310ad62021-02-23 13:33:38 -0800270#ifdef CONFIG_LINKER_USE_PINNED_SECTION
271/**
272 * @brief Clear BSS within the pinned region
273 *
274 * This routine clears the BSS within the pinned region.
275 * This is separate from z_bss_zero() as pinned region may
276 * contain symbols required for the boot process before
277 * paging is initialized.
278 */
279#ifdef CONFIG_LINKER_USE_BOOT_SECTION
280__boot_func
281#else
282__pinned_func
Simon Heinbcd1d192024-03-08 12:00:10 +0100283#endif /* CONFIG_LINKER_USE_BOOT_SECTION */
Daniel Leung1310ad62021-02-23 13:33:38 -0800284void z_bss_zero_pinned(void)
285{
Nicolas Pitre678b76e2022-02-10 13:54:49 -0500286 z_early_memset(&lnkr_pinned_bss_start, 0,
287 (uintptr_t)&lnkr_pinned_bss_end
288 - (uintptr_t)&lnkr_pinned_bss_start);
Daniel Leung1310ad62021-02-23 13:33:38 -0800289}
290#endif /* CONFIG_LINKER_USE_PINNED_SECTION */
291
Andrew Boie01100ea2019-02-21 15:02:22 -0800292#ifdef CONFIG_STACK_CANARIES
Flavio Ceolind16c5b92023-08-01 15:07:57 -0700293#ifdef CONFIG_STACK_CANARIES_TLS
294extern __thread volatile uintptr_t __stack_chk_guard;
295#else
Andrew Boie01100ea2019-02-21 15:02:22 -0800296extern volatile uintptr_t __stack_chk_guard;
Simon Heinbcd1d192024-03-08 12:00:10 +0100297#endif /* CONFIG_STACK_CANARIES_TLS */
Andrew Boie01100ea2019-02-21 15:02:22 -0800298#endif /* CONFIG_STACK_CANARIES */
299
Andrew Boiefe228a82019-06-11 12:49:32 -0700300/* LCOV_EXCL_STOP */
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400301
Daniel Leung660d1472021-03-25 16:05:15 -0700302__pinned_bss
Peter Bigot74ef3952019-12-23 11:48:43 -0600303bool z_sys_post_kernel;
Daniel Leung660d1472021-03-25 16:05:15 -0700304
Ederson de Souzaeeebb4d2024-01-05 13:29:20 -0800305static int do_device_init(const struct init_entry *entry)
306{
307 const struct device *dev = entry->dev;
308 int rc = 0;
309
310 if (entry->init_fn.dev != NULL) {
311 rc = entry->init_fn.dev(dev);
312 /* Mark device initialized. If initialization
313 * failed, record the error condition.
314 */
315 if (rc != 0) {
316 if (rc < 0) {
317 rc = -rc;
318 }
319 if (rc > UINT8_MAX) {
320 rc = UINT8_MAX;
321 }
322 dev->state->init_res = rc;
323 }
324 }
325
326 dev->state->initialized = true;
327
328 if (rc == 0) {
329 /* Run automatic device runtime enablement */
330 (void)pm_device_runtime_auto_enable(dev);
331 }
332
333 return rc;
334}
335
Gerard Marull-Paretas83123932022-10-04 11:25:37 +0200336/**
337 * @brief Execute all the init entry initialization functions at a given level
338 *
339 * @details Invokes the initialization routine for each init entry object
340 * created by the INIT_ENTRY_DEFINE() macro using the specified level.
341 * The linker script places the init entry objects in memory in the order
342 * they need to be invoked, with symbols indicating where one level leaves
343 * off and the next one begins.
344 *
345 * @param level init level to run.
346 */
Gerard Marull-Paretas495245a2022-10-04 11:52:18 +0200347static void z_sys_init_run_level(enum init_level level)
Gerard Marull-Paretas83123932022-10-04 11:25:37 +0200348{
349 static const struct init_entry *levels[] = {
350 __init_EARLY_start,
351 __init_PRE_KERNEL_1_start,
352 __init_PRE_KERNEL_2_start,
353 __init_POST_KERNEL_start,
354 __init_APPLICATION_start,
355#ifdef CONFIG_SMP
356 __init_SMP_start,
Simon Heinbcd1d192024-03-08 12:00:10 +0100357#endif /* CONFIG_SMP */
Gerard Marull-Paretas83123932022-10-04 11:25:37 +0200358 /* End marker */
359 __init_end,
360 };
361 const struct init_entry *entry;
362
363 for (entry = levels[level]; entry < levels[level+1]; entry++) {
364 const struct device *dev = entry->dev;
Gerard Marull-Paretas83123932022-10-04 11:25:37 +0200365
366 if (dev != NULL) {
Ederson de Souzaeeebb4d2024-01-05 13:29:20 -0800367 do_device_init(entry);
Gerard Marull-Paretasa5fd0d12022-10-19 09:33:44 +0200368 } else {
369 (void)entry->init_fn.sys();
Gerard Marull-Paretas83123932022-10-04 11:25:37 +0200370 }
371 }
372}
373
Ederson de Souzaeeebb4d2024-01-05 13:29:20 -0800374
375int z_impl_device_init(const struct device *dev)
376{
377 if (dev == NULL) {
378 return -ENOENT;
379 }
380
381 STRUCT_SECTION_FOREACH_ALTERNATE(_deferred_init, init_entry, entry) {
382 if (entry->dev == dev) {
383 return do_device_init(entry);
384 }
385 }
386
387 return -ENOENT;
388}
389
390#ifdef CONFIG_USERSPACE
391static inline int z_vrfy_device_init(const struct device *dev)
392{
393 K_OOPS(K_SYSCALL_OBJ_INIT(dev, K_OBJ_ANY));
394
395 return z_impl_device_init(dev);
396}
397#include <syscalls/device_init_mrsh.c>
398#endif
399
Anas Nashif4b593122020-08-27 09:08:40 -0400400extern void boot_banner(void);
Peter Bigot74ef3952019-12-23 11:48:43 -0600401
Anas Nashif3ca50f52024-02-23 10:53:01 -0500402
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400403/**
Leandro Pereiraa1ae8452018-03-06 15:08:55 -0800404 * @brief Mainline for kernel's background thread
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400405 *
406 * This routine completes kernel initialization by invoking the remaining
407 * init functions, then invokes application's main() routine.
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400408 */
Daniel Leung660d1472021-03-25 16:05:15 -0700409__boot_func
Leandro Pereiraa1ae8452018-03-06 15:08:55 -0800410static void bg_thread_main(void *unused1, void *unused2, void *unused3)
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400411{
412 ARG_UNUSED(unused1);
413 ARG_UNUSED(unused2);
414 ARG_UNUSED(unused3);
415
Andrew Boiee35f1792020-12-09 12:18:40 -0800416#ifdef CONFIG_MMU
417 /* Invoked here such that backing store or eviction algorithms may
418 * initialize kernel objects, and that all POST_KERNEL and later tasks
419 * may perform memory management tasks (except for z_phys_map() which
420 * is allowed at any time)
421 */
422 z_mem_manage_init();
423#endif /* CONFIG_MMU */
Peter Bigot74ef3952019-12-23 11:48:43 -0600424 z_sys_post_kernel = true;
425
Gerard Marull-Paretas495245a2022-10-04 11:52:18 +0200426 z_sys_init_run_level(INIT_LEVEL_POST_KERNEL);
Hess Nathan527e7122024-04-29 14:08:04 +0200427#if defined(CONFIG_STACK_POINTER_RANDOM) && (CONFIG_STACK_POINTER_RANDOM != 0)
Andrew Boie538754c2018-05-23 15:25:23 -0700428 z_stack_adjust_initialized = 1;
Simon Heinbcd1d192024-03-08 12:00:10 +0100429#endif /* CONFIG_STACK_POINTER_RANDOM */
Anas Nashif4b593122020-08-27 09:08:40 -0400430 boot_banner();
Andrew Boie0b474ee2016-11-08 11:06:55 -0800431
Stephanos Ioannidis4a64bfe2022-12-09 06:16:44 +0900432#if defined(CONFIG_CPP)
Evgeniy Paltsev497cb2e2021-04-26 16:15:49 +0300433 void z_cpp_init_static(void);
434 z_cpp_init_static();
Simon Heinbcd1d192024-03-08 12:00:10 +0100435#endif /* CONFIG_CPP */
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400436
Peter Bigotc308f802020-05-08 09:28:44 -0500437 /* Final init level before app starts */
Gerard Marull-Paretas495245a2022-10-04 11:52:18 +0200438 z_sys_init_run_level(INIT_LEVEL_APPLICATION);
Peter Bigotc308f802020-05-08 09:28:44 -0500439
Patrik Flykt4344e272019-03-08 14:19:05 -0700440 z_init_static_threads();
Anas Nashif83088a22017-08-24 04:27:51 -0400441
Anas Nashif39f632e2020-12-07 13:15:42 -0500442#ifdef CONFIG_KERNEL_COHERENCE
Daniel Leung079bc642021-02-02 16:12:15 -0800443 __ASSERT_NO_MSG(arch_mem_coherent(&_kernel));
Simon Heinbcd1d192024-03-08 12:00:10 +0100444#endif /* CONFIG_KERNEL_COHERENCE */
Andy Rossf6d32ab2020-05-13 15:34:04 +0000445
Andy Rosseb258702018-04-12 12:10:10 -0700446#ifdef CONFIG_SMP
Andy Ross2b210cb2022-01-17 11:56:54 -0800447 if (!IS_ENABLED(CONFIG_SMP_BOOT_DELAY)) {
448 z_smp_init();
449 }
Gerard Marull-Paretas495245a2022-10-04 11:52:18 +0200450 z_sys_init_run_level(INIT_LEVEL_SMP);
Simon Heinbcd1d192024-03-08 12:00:10 +0100451#endif /* CONFIG_SMP */
Inaky Perez-Gonzalezc51f73f2017-06-20 17:01:09 -0700452
Daniel Leunge88afd22021-07-15 13:15:29 -0700453#ifdef CONFIG_MMU
454 z_mem_manage_boot_finish();
455#endif /* CONFIG_MMU */
456
Stephanos Ioannidisfa5fd412022-11-05 00:22:25 +0900457 extern int main(void);
Andrew Boief1c373c2016-10-28 12:45:05 -0700458
Stephanos Ioannidisfa5fd412022-11-05 00:22:25 +0900459 (void)main();
Allan Stephens073442e2016-11-09 07:46:56 -0600460
Anas Nashif87910122024-02-22 22:24:36 -0500461 /* Mark non-essential since main() has no more work to do */
462 z_thread_essential_clear(&z_main_thread);
Andrew Boie8e053332019-06-11 12:58:16 -0700463
Anas Nashif471ffbe2020-01-30 08:44:10 -0500464#ifdef CONFIG_COVERAGE_DUMP
Adithya Baglody71e90f92018-08-29 16:44:16 +0530465 /* Dump coverage data once the main() has exited. */
466 gcov_coverage_dump();
Simon Heinbcd1d192024-03-08 12:00:10 +0100467#endif /* CONFIG_COVERAGE_DUMP */
Andrew Boie8e053332019-06-11 12:58:16 -0700468} /* LCOV_EXCL_LINE ... because we just dumped final coverage data */
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400469
Andy Ross780ba232018-01-29 09:20:18 -0800470#if defined(CONFIG_MULTITHREADING)
Daniel Leung660d1472021-03-25 16:05:15 -0700471__boot_func
Andrew Boiec24673e2020-03-16 09:44:28 -0700472static void init_idle_thread(int i)
Andy Ross780ba232018-01-29 09:20:18 -0800473{
Andrew Boiec24673e2020-03-16 09:44:28 -0700474 struct k_thread *thread = &z_idle_threads[i];
475 k_thread_stack_t *stack = z_idle_stacks[i];
476
477#ifdef CONFIG_THREAD_NAME
Andrew Boiec24673e2020-03-16 09:44:28 -0700478
Kumar Gala4f458ba2022-10-18 11:11:46 -0500479#if CONFIG_MP_MAX_NUM_CPUS > 1
Trond Einar Snekvik6224ecb2022-03-25 12:46:32 +0100480 char tname[8];
Andrew Boiec24673e2020-03-16 09:44:28 -0700481 snprintk(tname, 8, "idle %02d", i);
482#else
Trond Einar Snekvik6224ecb2022-03-25 12:46:32 +0100483 char *tname = "idle";
Simon Heinbcd1d192024-03-08 12:00:10 +0100484#endif /* CONFIG_MP_MAX_NUM_CPUS */
Trond Einar Snekvik6224ecb2022-03-25 12:46:32 +0100485
486#else
Andrew Boiec24673e2020-03-16 09:44:28 -0700487 char *tname = NULL;
488#endif /* CONFIG_THREAD_NAME */
489
Anas Nashif9e3e7f62019-12-19 08:19:45 -0500490 z_setup_new_thread(thread, stack,
Andrew Boief5a7e1a2020-09-02 09:20:38 -0700491 CONFIG_IDLE_STACK_SIZE, idle, &_kernel.cpus[i],
Andy Ross851d14a2021-05-13 15:46:43 -0700492 NULL, NULL, K_IDLE_PRIO, K_ESSENTIAL,
Andrew Boief5a7e1a2020-09-02 09:20:38 -0700493 tname);
Anas Nashif9e3e7f62019-12-19 08:19:45 -0500494 z_mark_thread_as_started(thread);
Andy Ross6c283ca2019-08-16 22:09:30 -0700495
496#ifdef CONFIG_SMP
Anas Nashif9e3e7f62019-12-19 08:19:45 -0500497 thread->base.is_idle = 1U;
Simon Heinbcd1d192024-03-08 12:00:10 +0100498#endif /* CONFIG_SMP */
Andy Ross780ba232018-01-29 09:20:18 -0800499}
Andy Ross780ba232018-01-29 09:20:18 -0800500
Andy Ross2b210cb2022-01-17 11:56:54 -0800501void z_init_cpu(int id)
Andy Rossc6d077e2021-08-18 06:28:11 -0700502{
Andy Ross2b210cb2022-01-17 11:56:54 -0800503 init_idle_thread(id);
504 _kernel.cpus[id].idle_thread = &z_idle_threads[id];
505 _kernel.cpus[id].id = id;
506 _kernel.cpus[id].irq_stack =
Daniel Leungb69d2482024-03-22 12:56:12 -0700507 (K_KERNEL_STACK_BUFFER(z_interrupt_stacks[id]) +
Andy Ross2b210cb2022-01-17 11:56:54 -0800508 K_KERNEL_STACK_SIZEOF(z_interrupt_stacks[id]));
509#ifdef CONFIG_SCHED_THREAD_USAGE_ALL
Peter Mitsis9bedfd82023-05-23 18:36:04 -0400510 _kernel.cpus[id].usage = &_kernel.usage[id];
511 _kernel.cpus[id].usage->track_usage =
Andy Ross2b210cb2022-01-17 11:56:54 -0800512 CONFIG_SCHED_THREAD_USAGE_AUTO_ENABLE;
513#endif
Flavio Ceolin4f299302023-05-30 11:49:45 -0700514
515 /*
516 * Increment number of CPUs active. The pm subsystem
517 * will keep track of this from here.
518 */
519 atomic_inc(&_cpus_active);
Peter Mitsis6df8efe2023-05-11 14:06:46 -0400520
521#ifdef CONFIG_OBJ_CORE_SYSTEM
522 k_obj_core_init_and_link(K_OBJ_CORE(&_kernel.cpus[id]), &obj_type_cpu);
Peter Mitsise6f10902023-06-01 12:16:40 -0400523#ifdef CONFIG_OBJ_CORE_STATS_SYSTEM
524 k_obj_core_stats_register(K_OBJ_CORE(&_kernel.cpus[id]),
525 _kernel.cpus[id].usage,
526 sizeof(struct k_cycle_stats));
527#endif
Peter Mitsis6df8efe2023-05-11 14:06:46 -0400528#endif
Andy Rossc6d077e2021-08-18 06:28:11 -0700529}
530
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400531/**
532 *
Anas Nashifdc3d73b2016-12-19 20:25:56 -0500533 * @brief Initializes kernel data structures
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400534 *
Anas Nashifdc3d73b2016-12-19 20:25:56 -0500535 * This routine initializes various kernel data structures, including
536 * the init and idle threads and any architecture-specific initialization.
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400537 *
Benjamin Walshf6ca7de2016-11-08 10:36:50 -0500538 * Note that all fields of "_kernel" are set to zero on entry, which may
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400539 * be all the initialization many of them require.
540 *
Andrew Boiee4cc84a2020-04-24 11:29:47 -0700541 * @return initial stack pointer for the main thread
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400542 */
Daniel Leung660d1472021-03-25 16:05:15 -0700543__boot_func
Andrew Boiee4cc84a2020-04-24 11:29:47 -0700544static char *prepare_multithreading(void)
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400545{
Andrew Boiee4cc84a2020-04-24 11:29:47 -0700546 char *stack_ptr;
547
Benjamin Walshf6ca7de2016-11-08 10:36:50 -0500548 /* _kernel.ready_q is all zeroes */
Patrik Flykt4344e272019-03-08 14:19:05 -0700549 z_sched_init();
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400550
Andy Ross2724fd12018-01-29 14:55:20 -0800551#ifndef CONFIG_SMP
Benjamin Walsh88b36912016-12-02 10:37:27 -0500552 /*
553 * prime the cache with the main thread since:
554 *
555 * - the cache can never be NULL
556 * - the main thread will be the one to run first
557 * - no other thread is initialized yet and thus their priority fields
558 * contain garbage, which would prevent the cache loading algorithm
559 * to work as intended
560 */
Andrew Boiefe031612019-09-21 17:54:37 -0700561 _kernel.ready_q.cache = &z_main_thread;
Simon Heinbcd1d192024-03-08 12:00:10 +0100562#endif /* CONFIG_SMP */
Andrew Boiee4cc84a2020-04-24 11:29:47 -0700563 stack_ptr = z_setup_new_thread(&z_main_thread, z_main_stack,
564 CONFIG_MAIN_STACK_SIZE, bg_thread_main,
565 NULL, NULL, NULL,
566 CONFIG_MAIN_THREAD_PRIORITY,
Ioannis Glaropoulos40aab322021-01-28 21:46:28 +0100567 K_ESSENTIAL, "main");
Andrew Boiefe031612019-09-21 17:54:37 -0700568 z_mark_thread_as_started(&z_main_thread);
569 z_ready_thread(&z_main_thread);
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400570
Andy Ross2b210cb2022-01-17 11:56:54 -0800571 z_init_cpu(0);
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400572
Andrew Boiee4cc84a2020-04-24 11:29:47 -0700573 return stack_ptr;
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400574}
575
Daniel Leung660d1472021-03-25 16:05:15 -0700576__boot_func
Andrew Boiee4cc84a2020-04-24 11:29:47 -0700577static FUNC_NORETURN void switch_to_main_thread(char *stack_ptr)
Benjamin Walshc742d7e2016-10-05 17:50:54 -0400578{
Benjamin Walsh296a2342016-11-20 11:04:31 -0500579#ifdef CONFIG_ARCH_HAS_CUSTOM_SWAP_TO_MAIN
Andrew Boiee4cc84a2020-04-24 11:29:47 -0700580 arch_switch_to_main_thread(&z_main_thread, stack_ptr, bg_thread_main);
Benjamin Walsh296a2342016-11-20 11:04:31 -0500581#else
Andrew Boiee4cc84a2020-04-24 11:29:47 -0700582 ARG_UNUSED(stack_ptr);
Benjamin Walshc742d7e2016-10-05 17:50:54 -0400583 /*
584 * Context switch to main task (entry function is _main()): the
585 * current fake thread is not on a wait queue or ready queue, so it
586 * will never be rescheduled in.
587 */
Patrik Flykt4344e272019-03-08 14:19:05 -0700588 z_swap_unlocked();
Simon Heinbcd1d192024-03-08 12:00:10 +0100589#endif /* CONFIG_ARCH_HAS_CUSTOM_SWAP_TO_MAIN */
Andrew Boiec5164f32019-06-11 13:33:32 -0700590 CODE_UNREACHABLE; /* LCOV_EXCL_LINE */
Benjamin Walshc742d7e2016-10-05 17:50:54 -0400591}
Anas Nashifc0ea5052019-01-30 09:58:41 -0500592#endif /* CONFIG_MULTITHREADING */
Benjamin Walshc742d7e2016-10-05 17:50:54 -0400593
Daniel Leung660d1472021-03-25 16:05:15 -0700594__boot_func
Flavio Ceolin974e3362023-10-09 15:25:42 -0700595void __weak z_early_rand_get(uint8_t *buf, size_t length)
Leandro Pereira389c3642018-05-23 13:38:52 -0700596{
Flavio Ceolin991ff6f2023-10-10 10:31:35 -0700597 static uint64_t state = (uint64_t)CONFIG_TIMER_RANDOM_INITIAL_STATE;
Leandro Pereira389c3642018-05-23 13:38:52 -0700598 int rc;
599
Flavio Ceolin6c0fad22023-10-09 20:52:07 -0700600#ifdef CONFIG_ENTROPY_HAS_DRIVER
601 const struct device *const entropy = DEVICE_DT_GET_OR_NULL(DT_CHOSEN(zephyr_entropy));
602
603 if ((entropy != NULL) && device_is_ready(entropy)) {
604 /* Try to see if driver provides an ISR-specific API */
605 rc = entropy_get_entropy_isr(entropy, buf, length, ENTROPY_BUSYWAIT);
606 if (rc > 0) {
607 length -= rc;
608 buf += rc;
609 }
Leandro Pereira389c3642018-05-23 13:38:52 -0700610 }
Simon Heinbcd1d192024-03-08 12:00:10 +0100611#endif /* CONFIG_ENTROPY_HAS_DRIVER */
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400612
Flavio Ceolin6c0fad22023-10-09 20:52:07 -0700613 while (length > 0) {
614 uint32_t val;
615
616 state = state + k_cycle_get_32();
617 state = state * 2862933555777941757ULL + 3037000493ULL;
618 val = (uint32_t)(state >> 32);
619 rc = MIN(length, sizeof(val));
620 z_early_memcpy((void *)buf, &val, rc);
621
622 length -= rc;
623 buf += rc;
624 }
Flavio Ceolin394f66b2019-08-09 16:31:33 -0700625}
626
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400627/**
628 *
Anas Nashifdc3d73b2016-12-19 20:25:56 -0500629 * @brief Initialize kernel
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400630 *
631 * This routine is invoked when the system is ready to run C code. The
632 * processor must be running in 32-bit mode, and the BSS must have been
633 * cleared/zeroed.
634 *
635 * @return Does not return
636 */
Daniel Leung660d1472021-03-25 16:05:15 -0700637__boot_func
Daniel Leung256db602022-12-15 15:54:56 -0800638FUNC_NO_STACK_PROTECTOR
Patrik Flykt4344e272019-03-08 14:19:05 -0700639FUNC_NORETURN void z_cstart(void)
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400640{
Adithya Baglody71e90f92018-08-29 16:44:16 +0530641 /* gcov hook needed to get the coverage report.*/
642 gcov_static_init();
643
Gerard Marull-Paretase42f58e2022-10-11 17:17:18 +0200644 /* initialize early init calls */
Gerard Marull-Paretas495245a2022-10-04 11:52:18 +0200645 z_sys_init_run_level(INIT_LEVEL_EARLY);
Gerard Marull-Paretase42f58e2022-10-11 17:17:18 +0200646
Andrew Boie982d5c82018-05-23 13:30:34 -0700647 /* perform any architecture-specific initialization */
Andrew Boie4f77c2a2019-11-07 12:43:29 -0800648 arch_kernel_init();
Andrew Boie982d5c82018-05-23 13:30:34 -0700649
Ederson de Souza75fd4522022-01-25 10:47:05 -0800650 LOG_CORE_INIT();
651
Andrew Boie468efad2020-05-12 16:20:14 -0700652#if defined(CONFIG_MULTITHREADING)
Andy Rossfd340eb2024-04-19 15:03:09 -0700653 z_dummy_thread_init(&_thread_dummy);
Simon Heinbcd1d192024-03-08 12:00:10 +0100654#endif /* CONFIG_MULTITHREADING */
Peter Bigot1cadd8b2021-02-02 10:07:18 -0600655 /* do any necessary initialization of static devices */
656 z_device_state_init();
657
Andrew Boie0b474ee2016-11-08 11:06:55 -0800658 /* perform basic hardware initialization */
Gerard Marull-Paretas495245a2022-10-04 11:52:18 +0200659 z_sys_init_run_level(INIT_LEVEL_PRE_KERNEL_1);
660 z_sys_init_run_level(INIT_LEVEL_PRE_KERNEL_2);
Andrew Boie0b474ee2016-11-08 11:06:55 -0800661
Mazen NEIFERe2bbad92017-02-07 10:01:12 +0100662#ifdef CONFIG_STACK_CANARIES
Andrew Boie468efad2020-05-12 16:20:14 -0700663 uintptr_t stack_guard;
664
Flavio Ceolinf9c7a5e2023-10-09 15:22:18 -0700665 z_early_rand_get((uint8_t *)&stack_guard, sizeof(stack_guard));
Flavio Ceolin394f66b2019-08-09 16:31:33 -0700666 __stack_chk_guard = stack_guard;
667 __stack_chk_guard <<= 8;
668#endif /* CONFIG_STACK_CANARIES */
Leandro Pereira389c3642018-05-23 13:38:52 -0700669
Daniel Leung15597122021-03-31 13:40:01 -0700670#ifdef CONFIG_TIMING_FUNCTIONS_NEED_AT_BOOT
Daniel Leungfd7a68d2020-10-14 12:17:12 -0700671 timing_init();
672 timing_start();
Simon Heinbcd1d192024-03-08 12:00:10 +0100673#endif /* CONFIG_TIMING_FUNCTIONS_NEED_AT_BOOT */
Daniel Leungfd7a68d2020-10-14 12:17:12 -0700674
Andy Ross3d146152018-06-13 10:51:42 -0700675#ifdef CONFIG_MULTITHREADING
Andrew Boiee4cc84a2020-04-24 11:29:47 -0700676 switch_to_main_thread(prepare_multithreading());
Andy Ross3d146152018-06-13 10:51:42 -0700677#else
Ioannis Glaropoulos60bd51a2020-08-03 11:11:19 +0200678#ifdef ARCH_SWITCH_TO_MAIN_NO_MULTITHREADING
679 /* Custom ARCH-specific routine to switch to main()
680 * in the case of no multi-threading.
681 */
682 ARCH_SWITCH_TO_MAIN_NO_MULTITHREADING(bg_thread_main,
683 NULL, NULL, NULL);
684#else
Andy Ross3d146152018-06-13 10:51:42 -0700685 bg_thread_main(NULL, NULL, NULL);
686
Andrew Boiec5164f32019-06-11 13:33:32 -0700687 /* LCOV_EXCL_START
688 * We've already dumped coverage data at this point.
689 */
Andy Ross8daafd42018-08-30 09:45:12 -0700690 irq_lock();
Flavio Ceolinb3d92022018-09-17 15:56:06 -0700691 while (true) {
Andy Ross3d146152018-06-13 10:51:42 -0700692 }
Andrew Boiec5164f32019-06-11 13:33:32 -0700693 /* LCOV_EXCL_STOP */
Simon Heinbcd1d192024-03-08 12:00:10 +0100694#endif /* ARCH_SWITCH_TO_MAIN_NO_MULTITHREADING */
Ioannis Glaropoulos60bd51a2020-08-03 11:11:19 +0200695#endif /* CONFIG_MULTITHREADING */
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400696
697 /*
698 * Compiler can't tell that the above routines won't return and issues
699 * a warning unless we explicitly tell it that control never gets this
700 * far.
701 */
702
Andrew Boiec5164f32019-06-11 13:33:32 -0700703 CODE_UNREACHABLE; /* LCOV_EXCL_LINE */
Benjamin Walsh456c6da2016-09-02 18:55:39 -0400704}
Peter Mitsis6df8efe2023-05-11 14:06:46 -0400705
706#ifdef CONFIG_OBJ_CORE_SYSTEM
707static int init_cpu_obj_core_list(void)
708{
709 /* Initialize CPU object type */
710
711 z_obj_type_init(&obj_type_cpu, K_OBJ_TYPE_CPU_ID,
712 offsetof(struct _cpu, obj_core));
713
Peter Mitsise6f10902023-06-01 12:16:40 -0400714#ifdef CONFIG_OBJ_CORE_STATS_SYSTEM
715 k_obj_type_stats_init(&obj_type_cpu, &cpu_stats_desc);
Simon Heinbcd1d192024-03-08 12:00:10 +0100716#endif /* CONFIG_OBJ_CORE_STATS_SYSTEM */
Peter Mitsise6f10902023-06-01 12:16:40 -0400717
Peter Mitsis6df8efe2023-05-11 14:06:46 -0400718 return 0;
719}
720
721static int init_kernel_obj_core_list(void)
722{
723 /* Initialize kernel object type */
724
725 z_obj_type_init(&obj_type_kernel, K_OBJ_TYPE_KERNEL_ID,
726 offsetof(struct z_kernel, obj_core));
727
Peter Mitsise6f10902023-06-01 12:16:40 -0400728#ifdef CONFIG_OBJ_CORE_STATS_SYSTEM
729 k_obj_type_stats_init(&obj_type_kernel, &kernel_stats_desc);
Simon Heinbcd1d192024-03-08 12:00:10 +0100730#endif /* CONFIG_OBJ_CORE_STATS_SYSTEM */
Peter Mitsise6f10902023-06-01 12:16:40 -0400731
Peter Mitsis6df8efe2023-05-11 14:06:46 -0400732 k_obj_core_init_and_link(K_OBJ_CORE(&_kernel), &obj_type_kernel);
Peter Mitsise6f10902023-06-01 12:16:40 -0400733#ifdef CONFIG_OBJ_CORE_STATS_SYSTEM
734 k_obj_core_stats_register(K_OBJ_CORE(&_kernel), _kernel.usage,
735 sizeof(_kernel.usage));
Simon Heinbcd1d192024-03-08 12:00:10 +0100736#endif /* CONFIG_OBJ_CORE_STATS_SYSTEM */
Peter Mitsis6df8efe2023-05-11 14:06:46 -0400737
738 return 0;
739}
740
741SYS_INIT(init_cpu_obj_core_list, PRE_KERNEL_1,
742 CONFIG_KERNEL_INIT_PRIORITY_OBJECTS);
743
744SYS_INIT(init_kernel_obj_core_list, PRE_KERNEL_1,
745 CONFIG_KERNEL_INIT_PRIORITY_OBJECTS);
Simon Heinbcd1d192024-03-08 12:00:10 +0100746#endif /* CONFIG_OBJ_CORE_SYSTEM */