| /* |
| * Copyright (c) 2017 Erwin Rol <erwin@erwinrol.com> |
| * Copyright (c) 2020 Alexander Kozhinov <AlexanderKozhinov@yandex.com> |
| * Copyright (c) 2021 Carbon Robotics |
| * SPDX-License-Identifier: Apache-2.0 |
| */ |
| |
| #define DT_DRV_COMPAT st_stm32_ethernet |
| |
| #define LOG_MODULE_NAME eth_stm32_hal |
| #define LOG_LEVEL CONFIG_ETHERNET_LOG_LEVEL |
| |
| #include <zephyr/logging/log.h> |
| LOG_MODULE_REGISTER(LOG_MODULE_NAME); |
| |
| #include <zephyr/kernel.h> |
| #include <zephyr/device.h> |
| #include <zephyr/sys/__assert.h> |
| #include <zephyr/sys/util.h> |
| #include <zephyr/sys/crc.h> |
| #include <errno.h> |
| #include <stdbool.h> |
| #include <zephyr/net/net_pkt.h> |
| #include <zephyr/net/net_if.h> |
| #include <zephyr/net/ethernet.h> |
| #include <ethernet/eth_stats.h> |
| #include <soc.h> |
| #include <zephyr/sys/printk.h> |
| #include <zephyr/drivers/clock_control.h> |
| #include <zephyr/drivers/clock_control/stm32_clock_control.h> |
| #include <zephyr/drivers/pinctrl.h> |
| #include <zephyr/irq.h> |
| #include <zephyr/net/lldp.h> |
| #include <zephyr/drivers/hwinfo.h> |
| |
| #if defined(CONFIG_PTP_CLOCK_STM32_HAL) |
| #include <zephyr/drivers/ptp_clock.h> |
| #endif /* CONFIG_PTP_CLOCK_STM32_HAL */ |
| |
| #include "eth.h" |
| #include "eth_stm32_hal_priv.h" |
| |
| #if defined(CONFIG_ETH_STM32_HAL_RANDOM_MAC) || DT_INST_PROP(0, zephyr_random_mac_address) |
| #define ETH_STM32_RANDOM_MAC |
| #endif |
| |
| #if defined(CONFIG_ETH_STM32_HAL_USE_DTCM_FOR_DMA_BUFFER) && \ |
| !DT_NODE_HAS_STATUS(DT_CHOSEN(zephyr_dtcm), okay) |
| #error DTCM for DMA buffer is activated but zephyr,dtcm is not present in dts |
| #endif |
| |
| #define PHY_ADDR CONFIG_ETH_STM32_HAL_PHY_ADDRESS |
| |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) |
| |
| #define PHY_BSR ((uint16_t)0x0001U) /*!< Transceiver Basic Status Register */ |
| #define PHY_LINKED_STATUS ((uint16_t)0x0004U) /*!< Valid link established */ |
| |
| #define IS_ETH_DMATXDESC_OWN(dma_tx_desc) (dma_tx_desc->DESC3 & \ |
| ETH_DMATXNDESCRF_OWN) |
| |
| #define ETH_RXBUFNB ETH_RX_DESC_CNT |
| #define ETH_TXBUFNB ETH_TX_DESC_CNT |
| |
| #define ETH_MEDIA_INTERFACE_MII HAL_ETH_MII_MODE |
| #define ETH_MEDIA_INTERFACE_RMII HAL_ETH_RMII_MODE |
| |
| /* Only one tx_buffer is sufficient to pass only 1 dma_buffer */ |
| #define ETH_TXBUF_DEF_NB 1U |
| #else |
| |
| #define IS_ETH_DMATXDESC_OWN(dma_tx_desc) (dma_tx_desc->Status & \ |
| ETH_DMATXDESC_OWN) |
| |
| #endif /* CONFIG_SOC_SERIES_STM32H7X */ |
| |
| #define ETH_DMA_TX_TIMEOUT_MS 20U /* transmit timeout in milliseconds */ |
| |
| #if defined(CONFIG_ETH_STM32_HAL_USE_DTCM_FOR_DMA_BUFFER) && \ |
| DT_NODE_HAS_STATUS(DT_CHOSEN(zephyr_dtcm), okay) |
| #define __eth_stm32_desc __dtcm_noinit_section |
| #define __eth_stm32_buf __dtcm_noinit_section |
| #elif defined(CONFIG_SOC_SERIES_STM32H7X) |
| #define __eth_stm32_desc __attribute__((section(".eth_stm32_desc"))) |
| #define __eth_stm32_buf __attribute__((section(".eth_stm32_buf"))) |
| #elif defined(CONFIG_NOCACHE_MEMORY) |
| #define __eth_stm32_desc __nocache __aligned(4) |
| #define __eth_stm32_buf __nocache __aligned(4) |
| #else |
| #define __eth_stm32_desc __aligned(4) |
| #define __eth_stm32_buf __aligned(4) |
| #endif |
| |
| static ETH_DMADescTypeDef dma_rx_desc_tab[ETH_RXBUFNB] __eth_stm32_desc; |
| static ETH_DMADescTypeDef dma_tx_desc_tab[ETH_TXBUFNB] __eth_stm32_desc; |
| static uint8_t dma_rx_buffer[ETH_RXBUFNB][ETH_STM32_RX_BUF_SIZE] __eth_stm32_buf; |
| static uint8_t dma_tx_buffer[ETH_TXBUFNB][ETH_STM32_TX_BUF_SIZE] __eth_stm32_buf; |
| |
| #if defined(CONFIG_ETH_STM32_MULTICAST_FILTER) |
| |
| static struct net_if_mcast_monitor mcast_monitor; |
| |
| static K_MUTEX_DEFINE(multicast_addr_lock); |
| |
| #if defined(CONFIG_NET_NATIVE_IPV6) |
| static struct in6_addr multicast_ipv6_joined_addrs[NET_IF_MAX_IPV6_MADDR] = {0}; |
| #endif /* CONFIG_NET_NATIVE_IPV6 */ |
| |
| #if defined(CONFIG_NET_NATIVE_IPV4) |
| static struct in_addr multicast_ipv4_joined_addrs[NET_IF_MAX_IPV4_MADDR] = {0}; |
| #endif /* CONFIG_NET_NATIVE_IPV4 */ |
| |
| #endif /* CONFIG_ETH_STM32_MULTICAST_FILTER */ |
| |
| #if defined(CONFIG_ETH_STM32_HAL_API_V2) |
| |
| BUILD_ASSERT(ETH_STM32_RX_BUF_SIZE % 4 == 0, "Rx buffer size must be a multiple of 4"); |
| |
| struct eth_stm32_rx_buffer_header { |
| struct eth_stm32_rx_buffer_header *next; |
| uint16_t size; |
| bool used; |
| }; |
| |
| struct eth_stm32_tx_buffer_header { |
| ETH_BufferTypeDef tx_buff; |
| bool used; |
| }; |
| |
| struct eth_stm32_tx_context { |
| struct net_pkt *pkt; |
| uint16_t first_tx_buffer_index; |
| }; |
| |
| static struct eth_stm32_rx_buffer_header dma_rx_buffer_header[ETH_RXBUFNB]; |
| static struct eth_stm32_tx_buffer_header dma_tx_buffer_header[ETH_TXBUFNB]; |
| |
| void HAL_ETH_RxAllocateCallback(uint8_t **buf) |
| { |
| for (size_t i = 0; i < ETH_RXBUFNB; ++i) { |
| if (!dma_rx_buffer_header[i].used) { |
| dma_rx_buffer_header[i].next = NULL; |
| dma_rx_buffer_header[i].size = 0; |
| dma_rx_buffer_header[i].used = true; |
| *buf = dma_rx_buffer[i]; |
| return; |
| } |
| } |
| *buf = NULL; |
| } |
| |
| /* Pointer to an array of ETH_STM32_RX_BUF_SIZE uint8_t's */ |
| typedef uint8_t (*RxBufferPtr)[ETH_STM32_RX_BUF_SIZE]; |
| |
| /* called by HAL_ETH_ReadData() */ |
| void HAL_ETH_RxLinkCallback(void **pStart, void **pEnd, uint8_t *buff, uint16_t Length) |
| { |
| /* buff points to the begin on one of the rx buffers, |
| * so we can compute the index of the given buffer |
| */ |
| size_t index = (RxBufferPtr)buff - &dma_rx_buffer[0]; |
| struct eth_stm32_rx_buffer_header *header = &dma_rx_buffer_header[index]; |
| |
| __ASSERT_NO_MSG(index < ETH_RXBUFNB); |
| |
| header->size = Length; |
| |
| if (!*pStart) { |
| /* first packet, set head pointer of linked list */ |
| *pStart = header; |
| *pEnd = header; |
| } else { |
| __ASSERT_NO_MSG(*pEnd != NULL); |
| /* not the first packet, add to list and adjust tail pointer */ |
| ((struct eth_stm32_rx_buffer_header *)*pEnd)->next = header; |
| *pEnd = header; |
| } |
| } |
| |
| /* Called by HAL_ETH_ReleaseTxPacket */ |
| void HAL_ETH_TxFreeCallback(uint32_t *buff) |
| { |
| __ASSERT_NO_MSG(buff != NULL); |
| |
| /* buff is the user context in tx_config.pData */ |
| struct eth_stm32_tx_context *ctx = (struct eth_stm32_tx_context *)buff; |
| struct eth_stm32_tx_buffer_header *buffer_header = |
| &dma_tx_buffer_header[ctx->first_tx_buffer_index]; |
| |
| while (buffer_header != NULL) { |
| buffer_header->used = false; |
| if (buffer_header->tx_buff.next != NULL) { |
| buffer_header = CONTAINER_OF(buffer_header->tx_buff.next, |
| struct eth_stm32_tx_buffer_header, tx_buff); |
| } else { |
| buffer_header = NULL; |
| } |
| } |
| } |
| |
| /* allocate a tx buffer and mark it as used */ |
| static inline uint16_t allocate_tx_buffer(void) |
| { |
| for (;;) { |
| for (uint16_t index = 0; index < ETH_TXBUFNB; index++) { |
| if (!dma_tx_buffer_header[index].used) { |
| dma_tx_buffer_header[index].used = true; |
| return index; |
| } |
| } |
| k_yield(); |
| } |
| } |
| #endif /* CONFIG_ETH_STM32_HAL_API_V2 */ |
| |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) || defined(CONFIG_ETH_STM32_HAL_API_V2) |
| static ETH_TxPacketConfig tx_config; |
| #endif |
| |
| static HAL_StatusTypeDef read_eth_phy_register(ETH_HandleTypeDef *heth, |
| uint32_t PHYAddr, |
| uint32_t PHYReg, |
| uint32_t *RegVal) |
| { |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) || defined(CONFIG_ETH_STM32_HAL_API_V2) |
| return HAL_ETH_ReadPHYRegister(heth, PHYAddr, PHYReg, RegVal); |
| #else |
| ARG_UNUSED(PHYAddr); |
| return HAL_ETH_ReadPHYRegister(heth, PHYReg, RegVal); |
| #endif /* CONFIG_SOC_SERIES_STM32H7X || CONFIG_ETH_STM32_HAL_API_V2 */ |
| } |
| |
| static inline void setup_mac_filter(ETH_HandleTypeDef *heth) |
| { |
| __ASSERT_NO_MSG(heth != NULL); |
| |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) |
| ETH_MACFilterConfigTypeDef MACFilterConf; |
| |
| HAL_ETH_GetMACFilterConfig(heth, &MACFilterConf); |
| #if defined(CONFIG_ETH_STM32_MULTICAST_FILTER) |
| MACFilterConf.HashMulticast = ENABLE; |
| MACFilterConf.PassAllMulticast = DISABLE; |
| #else |
| MACFilterConf.HashMulticast = DISABLE; |
| MACFilterConf.PassAllMulticast = ENABLE; |
| #endif /* CONFIG_ETH_STM32_MULTICAST_FILTER */ |
| MACFilterConf.HachOrPerfectFilter = DISABLE; |
| |
| HAL_ETH_SetMACFilterConfig(heth, &MACFilterConf); |
| |
| k_sleep(K_MSEC(1)); |
| #else |
| uint32_t tmp = heth->Instance->MACFFR; |
| |
| /* clear all multicast filter bits, resulting in perfect filtering */ |
| tmp &= ~(ETH_MULTICASTFRAMESFILTER_PERFECTHASHTABLE | |
| ETH_MULTICASTFRAMESFILTER_HASHTABLE | |
| ETH_MULTICASTFRAMESFILTER_PERFECT | |
| ETH_MULTICASTFRAMESFILTER_NONE); |
| |
| if (IS_ENABLED(CONFIG_ETH_STM32_MULTICAST_FILTER)) { |
| /* enable multicast hash receive filter */ |
| tmp |= ETH_MULTICASTFRAMESFILTER_HASHTABLE; |
| } else { |
| /* enable receiving all multicast frames */ |
| tmp |= ETH_MULTICASTFRAMESFILTER_NONE; |
| } |
| |
| heth->Instance->MACFFR = tmp; |
| |
| /* Wait until the write operation will be taken into account: |
| * at least four TX_CLK/RX_CLK clock cycles |
| */ |
| tmp = heth->Instance->MACFFR; |
| k_sleep(K_MSEC(1)); |
| heth->Instance->MACFFR = tmp; |
| #endif /* CONFIG_SOC_SERIES_STM32H7X) */ |
| } |
| |
| #if defined(CONFIG_PTP_CLOCK_STM32_HAL) |
| static bool eth_is_ptp_pkt(struct net_if *iface, struct net_pkt *pkt) |
| { |
| #if defined(CONFIG_NET_VLAN) |
| struct net_eth_vlan_hdr *hdr_vlan; |
| struct ethernet_context *eth_ctx; |
| |
| eth_ctx = net_if_l2_data(iface); |
| if (net_eth_is_vlan_enabled(eth_ctx, iface)) { |
| hdr_vlan = (struct net_eth_vlan_hdr *)NET_ETH_HDR(pkt); |
| |
| if (ntohs(hdr_vlan->type) != NET_ETH_PTYPE_PTP) { |
| return false; |
| } |
| } else |
| #endif |
| { |
| if (ntohs(NET_ETH_HDR(pkt)->type) != NET_ETH_PTYPE_PTP) { |
| return false; |
| } |
| } |
| |
| net_pkt_set_priority(pkt, NET_PRIORITY_CA); |
| |
| return true; |
| } |
| #if defined(CONFIG_ETH_STM32_HAL_API_V2) |
| void HAL_ETH_TxPtpCallback(uint32_t *buff, ETH_TimeStampTypeDef *timestamp) |
| { |
| struct eth_stm32_tx_context *ctx = (struct eth_stm32_tx_context *)buff; |
| |
| ctx->pkt->timestamp.second = timestamp->TimeStampHigh; |
| ctx->pkt->timestamp.nanosecond = timestamp->TimeStampLow; |
| |
| net_if_add_tx_timestamp(ctx->pkt); |
| } |
| #endif /* CONFIG_ETH_STM32_HAL_API_V2 */ |
| #endif /* CONFIG_PTP_CLOCK_STM32_HAL */ |
| |
| static int eth_tx(const struct device *dev, struct net_pkt *pkt) |
| { |
| struct eth_stm32_hal_dev_data *dev_data = dev->data; |
| ETH_HandleTypeDef *heth; |
| int res; |
| size_t total_len; |
| #if defined(CONFIG_ETH_STM32_HAL_API_V2) |
| size_t remaining_read; |
| struct eth_stm32_tx_context ctx = {.pkt = pkt, .first_tx_buffer_index = 0}; |
| struct eth_stm32_tx_buffer_header *buf_header = NULL; |
| #else |
| uint8_t *dma_buffer; |
| __IO ETH_DMADescTypeDef *dma_tx_desc; |
| #endif /* CONFIG_ETH_STM32_HAL_API_V2 */ |
| HAL_StatusTypeDef hal_ret = HAL_OK; |
| #if defined(CONFIG_PTP_CLOCK_STM32_HAL) |
| bool timestamped_frame; |
| #endif /* CONFIG_PTP_CLOCK_STM32_HAL */ |
| |
| __ASSERT_NO_MSG(pkt != NULL); |
| __ASSERT_NO_MSG(pkt->frags != NULL); |
| __ASSERT_NO_MSG(dev != NULL); |
| __ASSERT_NO_MSG(dev_data != NULL); |
| |
| heth = &dev_data->heth; |
| |
| total_len = net_pkt_get_len(pkt); |
| if (total_len > (ETH_STM32_TX_BUF_SIZE * ETH_TXBUFNB)) { |
| LOG_ERR("PKT too big"); |
| return -EIO; |
| } |
| |
| k_mutex_lock(&dev_data->tx_mutex, K_FOREVER); |
| |
| #if defined(CONFIG_ETH_STM32_HAL_API_V2) |
| ctx.first_tx_buffer_index = allocate_tx_buffer(); |
| buf_header = &dma_tx_buffer_header[ctx.first_tx_buffer_index]; |
| #else /* CONFIG_ETH_STM32_HAL_API_V2 */ |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) |
| uint32_t cur_tx_desc_idx; |
| |
| cur_tx_desc_idx = heth->TxDescList.CurTxDesc; |
| dma_tx_desc = (ETH_DMADescTypeDef *)heth->TxDescList.TxDesc[cur_tx_desc_idx]; |
| #else |
| dma_tx_desc = heth->TxDesc; |
| #endif /* CONFIG_SOC_SERIES_STM32H7X */ |
| |
| while (IS_ETH_DMATXDESC_OWN(dma_tx_desc) != (uint32_t)RESET) { |
| k_yield(); |
| } |
| #endif /* CONFIG_ETH_STM32_HAL_API_V2 */ |
| |
| #if defined(CONFIG_PTP_CLOCK_STM32_HAL) |
| timestamped_frame = eth_is_ptp_pkt(net_pkt_iface(pkt), pkt); |
| if (timestamped_frame) { |
| /* Enable transmit timestamp */ |
| #if defined(CONFIG_ETH_STM32_HAL_API_V2) |
| HAL_ETH_PTP_InsertTxTimestamp(heth); |
| #elif defined(CONFIG_SOC_SERIES_STM32H7X) |
| dma_tx_desc->DESC2 |= ETH_DMATXNDESCRF_TTSE; |
| #else |
| dma_tx_desc->Status |= ETH_DMATXDESC_TTSE; |
| #endif /* CONFIG_ETH_STM32_HAL_API_V2 */ |
| } |
| #endif /* CONFIG_PTP_CLOCK_STM32_HAL */ |
| |
| #if defined(CONFIG_ETH_STM32_HAL_API_V2) |
| remaining_read = total_len; |
| /* fill and allocate buffer until remaining data fits in one buffer */ |
| while (remaining_read > ETH_STM32_TX_BUF_SIZE) { |
| if (net_pkt_read(pkt, buf_header->tx_buff.buffer, ETH_STM32_TX_BUF_SIZE)) { |
| res = -ENOBUFS; |
| goto error; |
| } |
| const uint16_t next_buffer_id = allocate_tx_buffer(); |
| |
| buf_header->tx_buff.len = ETH_STM32_TX_BUF_SIZE; |
| /* append new buffer to the linked list */ |
| buf_header->tx_buff.next = &dma_tx_buffer_header[next_buffer_id].tx_buff; |
| /* and adjust tail pointer */ |
| buf_header = &dma_tx_buffer_header[next_buffer_id]; |
| remaining_read -= ETH_STM32_TX_BUF_SIZE; |
| } |
| if (net_pkt_read(pkt, buf_header->tx_buff.buffer, remaining_read)) { |
| res = -ENOBUFS; |
| goto error; |
| } |
| buf_header->tx_buff.len = remaining_read; |
| buf_header->tx_buff.next = NULL; |
| |
| #else /* CONFIG_ETH_STM32_HAL_API_V2 */ |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) |
| dma_buffer = dma_tx_buffer[cur_tx_desc_idx]; |
| #else |
| dma_buffer = (uint8_t *)(dma_tx_desc->Buffer1Addr); |
| #endif /* CONFIG_SOC_SERIES_STM32H7X */ |
| |
| if (net_pkt_read(pkt, dma_buffer, total_len)) { |
| res = -ENOBUFS; |
| goto error; |
| } |
| |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) |
| ETH_BufferTypeDef tx_buffer_def; |
| |
| tx_buffer_def.buffer = dma_buffer; |
| tx_buffer_def.len = total_len; |
| tx_buffer_def.next = NULL; |
| #endif /* CONFIG_SOC_SERIES_STM32H7X */ |
| #endif /* CONFIG_ETH_STM32_HAL_API_V2 */ |
| |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) || defined(CONFIG_ETH_STM32_HAL_API_V2) |
| |
| tx_config.Length = total_len; |
| #if defined(CONFIG_ETH_STM32_HAL_API_V2) |
| tx_config.pData = &ctx; |
| tx_config.TxBuffer = &dma_tx_buffer_header[ctx.first_tx_buffer_index].tx_buff; |
| #else |
| tx_config.TxBuffer = &tx_buffer_def; |
| #endif /* CONFIG_ETH_STM32_HAL_API_V2 */ |
| |
| /* Reset TX complete interrupt semaphore before TX request*/ |
| k_sem_reset(&dev_data->tx_int_sem); |
| |
| /* tx_buffer is allocated on function stack, we need */ |
| /* to wait for the transfer to complete */ |
| /* So it is not freed before the interrupt happens */ |
| hal_ret = HAL_ETH_Transmit_IT(heth, &tx_config); |
| |
| if (hal_ret != HAL_OK) { |
| LOG_ERR("HAL_ETH_Transmit: failed!"); |
| res = -EIO; |
| goto error; |
| } |
| |
| /* Wait for end of TX buffer transmission */ |
| /* If the semaphore timeout breaks, it means */ |
| /* an error occurred or IT was not fired */ |
| if (k_sem_take(&dev_data->tx_int_sem, |
| K_MSEC(ETH_DMA_TX_TIMEOUT_MS)) != 0) { |
| |
| LOG_ERR("HAL_ETH_TransmitIT tx_int_sem take timeout"); |
| res = -EIO; |
| |
| #ifndef CONFIG_ETH_STM32_HAL_API_V2 |
| /* Content of the packet could be the reason for timeout */ |
| LOG_HEXDUMP_ERR(dma_buffer, total_len, "eth packet timeout"); |
| #endif |
| |
| /* Check for errors */ |
| /* Ethernet device was put in error state */ |
| /* Error state is unrecoverable ? */ |
| if (HAL_ETH_GetState(heth) == HAL_ETH_STATE_ERROR) { |
| LOG_ERR("%s: ETH in error state: errorcode:%x", |
| __func__, |
| HAL_ETH_GetError(heth)); |
| /* TODO recover from error state by restarting eth */ |
| } |
| |
| /* Check for DMA errors */ |
| if (HAL_ETH_GetDMAError(heth)) { |
| LOG_ERR("%s: ETH DMA error: dmaerror:%x", |
| __func__, |
| HAL_ETH_GetDMAError(heth)); |
| /* DMA fatal bus errors are putting in error state*/ |
| /* TODO recover from this */ |
| } |
| |
| /* Check for MAC errors */ |
| if (HAL_ETH_GetDMAError(heth)) { |
| LOG_ERR("%s: ETH DMA error: macerror:%x", |
| __func__, |
| HAL_ETH_GetDMAError(heth)); |
| /* MAC errors are putting in error state*/ |
| /* TODO recover from this */ |
| } |
| |
| goto error; |
| } |
| |
| #else |
| hal_ret = HAL_ETH_TransmitFrame(heth, total_len); |
| |
| if (hal_ret != HAL_OK) { |
| LOG_ERR("HAL_ETH_Transmit: failed!"); |
| res = -EIO; |
| goto error; |
| } |
| |
| /* When Transmit Underflow flag is set, clear it and issue a |
| * Transmit Poll Demand to resume transmission. |
| */ |
| if ((heth->Instance->DMASR & ETH_DMASR_TUS) != (uint32_t)RESET) { |
| /* Clear TUS ETHERNET DMA flag */ |
| heth->Instance->DMASR = ETH_DMASR_TUS; |
| /* Resume DMA transmission*/ |
| heth->Instance->DMATPDR = 0; |
| res = -EIO; |
| goto error; |
| } |
| #endif /* CONFIG_SOC_SERIES_STM32H7X || CONFIG_ETH_STM32_HAL_API_V2 */ |
| |
| #if defined(CONFIG_PTP_CLOCK_STM32_HAL) && !defined(CONFIG_ETH_STM32_HAL_API_V2) |
| if (timestamped_frame) { |
| /* Retrieve transmission timestamp from last DMA TX descriptor */ |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) |
| ETH_TxDescListTypeDef * dma_tx_desc_list; |
| |
| __IO ETH_DMADescTypeDef *last_dma_tx_desc; |
| |
| dma_tx_desc_list = &heth->TxDescList; |
| for (uint32_t i = 0; i < ETH_TX_DESC_CNT; i++) { |
| const uint32_t last_desc_idx = (cur_tx_desc_idx + i) % ETH_TX_DESC_CNT; |
| |
| last_dma_tx_desc = |
| (ETH_DMADescTypeDef *)dma_tx_desc_list->TxDesc[last_desc_idx]; |
| if (last_dma_tx_desc->DESC3 & ETH_DMATXNDESCWBF_LD) { |
| break; |
| } |
| } |
| |
| while (IS_ETH_DMATXDESC_OWN(last_dma_tx_desc) != (uint32_t)RESET) { |
| /* Wait for transmission */ |
| k_yield(); |
| } |
| |
| if ((last_dma_tx_desc->DESC3 & ETH_DMATXNDESCWBF_LD) && |
| (last_dma_tx_desc->DESC3 & ETH_DMATXNDESCWBF_TTSS)) { |
| pkt->timestamp.second = last_dma_tx_desc->DESC1; |
| pkt->timestamp.nanosecond = last_dma_tx_desc->DESC0; |
| } else { |
| /* Invalid value */ |
| pkt->timestamp.second = UINT64_MAX; |
| pkt->timestamp.nanosecond = UINT32_MAX; |
| } |
| #else |
| __IO ETH_DMADescTypeDef *last_dma_tx_desc = dma_tx_desc; |
| |
| while (!(last_dma_tx_desc->Status & ETH_DMATXDESC_LS) && |
| last_dma_tx_desc->Buffer2NextDescAddr) { |
| last_dma_tx_desc = |
| (ETH_DMADescTypeDef *)last_dma_tx_desc->Buffer2NextDescAddr; |
| } |
| |
| while (IS_ETH_DMATXDESC_OWN(last_dma_tx_desc) != (uint32_t)RESET) { |
| /* Wait for transmission */ |
| k_yield(); |
| } |
| |
| if (last_dma_tx_desc->Status & ETH_DMATXDESC_LS && |
| last_dma_tx_desc->Status & ETH_DMATXDESC_TTSS) { |
| pkt->timestamp.second = last_dma_tx_desc->TimeStampHigh; |
| pkt->timestamp.nanosecond = last_dma_tx_desc->TimeStampLow; |
| } else { |
| /* Invalid value */ |
| pkt->timestamp.second = UINT64_MAX; |
| pkt->timestamp.nanosecond = UINT32_MAX; |
| } |
| #endif /* CONFIG_SOC_SERIES_STM32H7X */ |
| |
| net_if_add_tx_timestamp(pkt); |
| } |
| #endif /* CONFIG_PTP_CLOCK_STM32_HAL && !CONFIG_ETH_STM32_HAL_API_V2 */ |
| |
| res = 0; |
| error: |
| |
| #if defined(CONFIG_ETH_STM32_HAL_API_V2) |
| /* free package tx buffer */ |
| if (res != 0) { |
| HAL_ETH_TxFreeCallback((uint32_t *)&ctx); |
| } else if (HAL_ETH_ReleaseTxPacket(heth) != HAL_OK) { |
| LOG_ERR("HAL_ETH_ReleaseTxPacket failed"); |
| res = -EIO; |
| } |
| #endif |
| |
| k_mutex_unlock(&dev_data->tx_mutex); |
| |
| return res; |
| } |
| |
| static struct net_if *get_iface(struct eth_stm32_hal_dev_data *ctx, |
| uint16_t vlan_tag) |
| { |
| #if defined(CONFIG_NET_VLAN) |
| struct net_if *iface; |
| |
| iface = net_eth_get_vlan_iface(ctx->iface, vlan_tag); |
| if (!iface) { |
| return ctx->iface; |
| } |
| |
| return iface; |
| #else |
| ARG_UNUSED(vlan_tag); |
| |
| return ctx->iface; |
| #endif |
| } |
| |
| static struct net_pkt *eth_rx(const struct device *dev, uint16_t *vlan_tag) |
| { |
| struct eth_stm32_hal_dev_data *dev_data; |
| ETH_HandleTypeDef *heth; |
| struct net_pkt *pkt; |
| size_t total_len = 0; |
| #if defined(CONFIG_ETH_STM32_HAL_API_V2) |
| void *appbuf = NULL; |
| struct eth_stm32_rx_buffer_header *rx_header; |
| #else |
| #if !defined(CONFIG_SOC_SERIES_STM32H7X) |
| __IO ETH_DMADescTypeDef *dma_rx_desc; |
| #endif /* !CONFIG_SOC_SERIES_STM32H7X */ |
| uint8_t *dma_buffer; |
| HAL_StatusTypeDef hal_ret = HAL_OK; |
| #endif /* CONFIG_ETH_STM32_HAL_API_V2 */ |
| #if defined(CONFIG_PTP_CLOCK_STM32_HAL) |
| struct net_ptp_time timestamp; |
| #if defined(CONFIG_ETH_STM32_HAL_API_V2) |
| ETH_TimeStampTypeDef ts_registers; |
| #endif /* CONFIG_ETH_STM32_HAL_API_V2 */ |
| /* Default to invalid value. */ |
| timestamp.second = UINT64_MAX; |
| timestamp.nanosecond = UINT32_MAX; |
| #endif /* CONFIG_PTP_CLOCK_STM32_HAL */ |
| |
| __ASSERT_NO_MSG(dev != NULL); |
| |
| dev_data = dev->data; |
| |
| __ASSERT_NO_MSG(dev_data != NULL); |
| |
| heth = &dev_data->heth; |
| |
| #if defined(CONFIG_ETH_STM32_HAL_API_V2) |
| if (HAL_ETH_ReadData(heth, &appbuf) != HAL_OK) { |
| /* no frame available */ |
| return NULL; |
| } |
| |
| /* computing total length */ |
| for (rx_header = (struct eth_stm32_rx_buffer_header *)appbuf; |
| rx_header; rx_header = rx_header->next) { |
| total_len += rx_header->size; |
| } |
| #elif defined(CONFIG_SOC_SERIES_STM32H7X) |
| if (HAL_ETH_IsRxDataAvailable(heth) != true) { |
| /* no frame available */ |
| return NULL; |
| } |
| |
| ETH_BufferTypeDef rx_buffer_def; |
| uint32_t frame_length = 0; |
| |
| hal_ret = HAL_ETH_GetRxDataBuffer(heth, &rx_buffer_def); |
| if (hal_ret != HAL_OK) { |
| LOG_ERR("HAL_ETH_GetRxDataBuffer: failed with state: %d", |
| hal_ret); |
| return NULL; |
| } |
| |
| hal_ret = HAL_ETH_GetRxDataLength(heth, &frame_length); |
| if (hal_ret != HAL_OK) { |
| LOG_ERR("HAL_ETH_GetRxDataLength: failed with state: %d", |
| hal_ret); |
| return NULL; |
| } |
| |
| total_len = frame_length; |
| dma_buffer = rx_buffer_def.buffer; |
| #else |
| hal_ret = HAL_ETH_GetReceivedFrame_IT(heth); |
| if (hal_ret != HAL_OK) { |
| /* no frame available */ |
| return NULL; |
| } |
| |
| total_len = heth->RxFrameInfos.length; |
| dma_buffer = (uint8_t *)heth->RxFrameInfos.buffer; |
| #endif /* CONFIG_SOC_SERIES_STM32H7X */ |
| |
| #if defined(CONFIG_PTP_CLOCK_STM32_HAL) |
| #if defined(CONFIG_ETH_STM32_HAL_API_V2) |
| |
| if (HAL_ETH_PTP_GetRxTimestamp(heth, &ts_registers) == HAL_OK) { |
| timestamp.second = ts_registers.TimeStampHigh; |
| timestamp.nanosecond = ts_registers.TimeStampLow; |
| } |
| |
| #elif defined(CONFIG_SOC_SERIES_STM32H7X) |
| ETH_RxDescListTypeDef * dma_rx_desc_list; |
| |
| dma_rx_desc_list = &heth->RxDescList; |
| if (dma_rx_desc_list->AppDescNbr) { |
| __IO ETH_DMADescTypeDef *last_dma_rx_desc; |
| |
| const uint32_t last_desc_idx = |
| (dma_rx_desc_list->FirstAppDesc + dma_rx_desc_list->AppDescNbr - 1U) |
| % ETH_RX_DESC_CNT; |
| |
| last_dma_rx_desc = |
| (ETH_DMADescTypeDef *)dma_rx_desc_list->RxDesc[last_desc_idx]; |
| |
| if (dma_rx_desc_list->AppContextDesc && |
| last_dma_rx_desc->DESC1 & ETH_DMARXNDESCWBF_TSA) { |
| /* Retrieve timestamp from context DMA descriptor */ |
| __IO ETH_DMADescTypeDef *context_dma_rx_desc; |
| |
| const uint32_t context_desc_idx = (last_desc_idx + 1U) % ETH_RX_DESC_CNT; |
| |
| context_dma_rx_desc = |
| (ETH_DMADescTypeDef *)dma_rx_desc_list->RxDesc[context_desc_idx]; |
| if (context_dma_rx_desc->DESC1 != UINT32_MAX || |
| context_dma_rx_desc->DESC0 != UINT32_MAX) { |
| timestamp.second = context_dma_rx_desc->DESC1; |
| timestamp.nanosecond = context_dma_rx_desc->DESC0; |
| } |
| } |
| } |
| #else |
| __IO ETH_DMADescTypeDef *last_dma_rx_desc; |
| |
| last_dma_rx_desc = heth->RxFrameInfos.LSRxDesc; |
| if (last_dma_rx_desc->TimeStampHigh != UINT32_MAX || |
| last_dma_rx_desc->TimeStampLow != UINT32_MAX) { |
| timestamp.second = last_dma_rx_desc->TimeStampHigh; |
| timestamp.nanosecond = last_dma_rx_desc->TimeStampLow; |
| } |
| #endif /* CONFIG_SOC_SERIES_STM32H7X */ |
| #endif /* CONFIG_PTP_CLOCK_STM32_HAL */ |
| |
| pkt = net_pkt_rx_alloc_with_buffer(get_iface(dev_data, *vlan_tag), |
| total_len, AF_UNSPEC, 0, K_MSEC(100)); |
| if (!pkt) { |
| LOG_ERR("Failed to obtain RX buffer"); |
| goto release_desc; |
| } |
| |
| #if defined(CONFIG_ETH_STM32_HAL_API_V2) |
| for (rx_header = (struct eth_stm32_rx_buffer_header *)appbuf; |
| rx_header; rx_header = rx_header->next) { |
| const size_t index = rx_header - &dma_rx_buffer_header[0]; |
| |
| __ASSERT_NO_MSG(index < ETH_RXBUFNB); |
| if (net_pkt_write(pkt, dma_rx_buffer[index], rx_header->size)) { |
| LOG_ERR("Failed to append RX buffer to context buffer"); |
| net_pkt_unref(pkt); |
| pkt = NULL; |
| goto release_desc; |
| } |
| } |
| #else |
| if (net_pkt_write(pkt, dma_buffer, total_len)) { |
| LOG_ERR("Failed to append RX buffer to context buffer"); |
| net_pkt_unref(pkt); |
| pkt = NULL; |
| goto release_desc; |
| } |
| #endif /* CONFIG_ETH_STM32_HAL_API_V2 */ |
| |
| release_desc: |
| #if defined(CONFIG_ETH_STM32_HAL_API_V2) |
| for (rx_header = (struct eth_stm32_rx_buffer_header *)appbuf; |
| rx_header; rx_header = rx_header->next) { |
| rx_header->used = false; |
| } |
| |
| #elif defined(CONFIG_SOC_SERIES_STM32H7X) |
| hal_ret = HAL_ETH_BuildRxDescriptors(heth); |
| if (hal_ret != HAL_OK) { |
| LOG_ERR("HAL_ETH_BuildRxDescriptors: failed: %d", hal_ret); |
| } |
| #else |
| /* Release descriptors to DMA */ |
| /* Point to first descriptor */ |
| dma_rx_desc = heth->RxFrameInfos.FSRxDesc; |
| /* Set Own bit in Rx descriptors: gives the buffers back to DMA */ |
| for (int i = 0; i < heth->RxFrameInfos.SegCount; i++) { |
| dma_rx_desc->Status |= ETH_DMARXDESC_OWN; |
| dma_rx_desc = (ETH_DMADescTypeDef *) |
| (dma_rx_desc->Buffer2NextDescAddr); |
| } |
| |
| /* Clear Segment_Count */ |
| heth->RxFrameInfos.SegCount = 0; |
| |
| /* When Rx Buffer unavailable flag is set: clear it |
| * and resume reception. |
| */ |
| if ((heth->Instance->DMASR & ETH_DMASR_RBUS) != (uint32_t)RESET) { |
| /* Clear RBUS ETHERNET DMA flag */ |
| heth->Instance->DMASR = ETH_DMASR_RBUS; |
| /* Resume DMA reception */ |
| heth->Instance->DMARPDR = 0; |
| } |
| #endif /* CONFIG_ETH_STM32_HAL_API_V2 */ |
| |
| if (!pkt) { |
| goto out; |
| } |
| |
| #if defined(CONFIG_NET_VLAN) |
| struct net_eth_hdr *hdr = NET_ETH_HDR(pkt); |
| |
| if (ntohs(hdr->type) == NET_ETH_PTYPE_VLAN) { |
| struct net_eth_vlan_hdr *hdr_vlan = |
| (struct net_eth_vlan_hdr *)NET_ETH_HDR(pkt); |
| |
| net_pkt_set_vlan_tci(pkt, ntohs(hdr_vlan->vlan.tci)); |
| *vlan_tag = net_pkt_vlan_tag(pkt); |
| |
| #if CONFIG_NET_TC_RX_COUNT > 1 |
| enum net_priority prio; |
| |
| prio = net_vlan2priority(net_pkt_vlan_priority(pkt)); |
| net_pkt_set_priority(pkt, prio); |
| #endif |
| } else { |
| net_pkt_set_iface(pkt, dev_data->iface); |
| } |
| #endif /* CONFIG_NET_VLAN */ |
| |
| #if defined(CONFIG_PTP_CLOCK_STM32_HAL) |
| if (eth_is_ptp_pkt(get_iface(dev_data, *vlan_tag), pkt)) { |
| pkt->timestamp.second = timestamp.second; |
| pkt->timestamp.nanosecond = timestamp.nanosecond; |
| } else { |
| /* Invalid value */ |
| pkt->timestamp.second = UINT64_MAX; |
| pkt->timestamp.nanosecond = UINT32_MAX; |
| } |
| #endif /* CONFIG_PTP_CLOCK_STM32_HAL */ |
| |
| out: |
| if (!pkt) { |
| eth_stats_update_errors_rx(get_iface(dev_data, *vlan_tag)); |
| } |
| |
| return pkt; |
| } |
| |
| static void rx_thread(void *arg1, void *unused1, void *unused2) |
| { |
| uint16_t vlan_tag = NET_VLAN_TAG_UNSPEC; |
| const struct device *dev; |
| struct eth_stm32_hal_dev_data *dev_data; |
| struct net_pkt *pkt; |
| int res; |
| uint32_t status; |
| HAL_StatusTypeDef hal_ret = HAL_OK; |
| |
| __ASSERT_NO_MSG(arg1 != NULL); |
| ARG_UNUSED(unused1); |
| ARG_UNUSED(unused2); |
| |
| dev = (const struct device *)arg1; |
| dev_data = dev->data; |
| |
| __ASSERT_NO_MSG(dev_data != NULL); |
| |
| while (1) { |
| res = k_sem_take(&dev_data->rx_int_sem, |
| K_MSEC(CONFIG_ETH_STM32_CARRIER_CHECK_RX_IDLE_TIMEOUT_MS)); |
| if (res == 0) { |
| /* semaphore taken, update link status and receive packets */ |
| if (dev_data->link_up != true) { |
| dev_data->link_up = true; |
| net_eth_carrier_on(get_iface(dev_data, |
| vlan_tag)); |
| } |
| while ((pkt = eth_rx(dev, &vlan_tag)) != NULL) { |
| res = net_recv_data(net_pkt_iface(pkt), pkt); |
| if (res < 0) { |
| eth_stats_update_errors_rx( |
| net_pkt_iface(pkt)); |
| LOG_ERR("Failed to enqueue frame " |
| "into RX queue: %d", res); |
| net_pkt_unref(pkt); |
| } |
| } |
| } else if (res == -EAGAIN) { |
| /* semaphore timeout period expired, check link status */ |
| hal_ret = read_eth_phy_register(&dev_data->heth, |
| PHY_ADDR, PHY_BSR, (uint32_t *) &status); |
| if (hal_ret == HAL_OK) { |
| if ((status & PHY_LINKED_STATUS) == PHY_LINKED_STATUS) { |
| if (dev_data->link_up != true) { |
| dev_data->link_up = true; |
| net_eth_carrier_on( |
| get_iface(dev_data, |
| vlan_tag)); |
| } |
| } else { |
| if (dev_data->link_up != false) { |
| dev_data->link_up = false; |
| net_eth_carrier_off( |
| get_iface(dev_data, |
| vlan_tag)); |
| } |
| } |
| } |
| } |
| } |
| } |
| |
| static void eth_isr(const struct device *dev) |
| { |
| struct eth_stm32_hal_dev_data *dev_data; |
| ETH_HandleTypeDef *heth; |
| |
| __ASSERT_NO_MSG(dev != NULL); |
| |
| dev_data = dev->data; |
| |
| __ASSERT_NO_MSG(dev_data != NULL); |
| |
| heth = &dev_data->heth; |
| |
| __ASSERT_NO_MSG(heth != NULL); |
| |
| HAL_ETH_IRQHandler(heth); |
| } |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) || defined(CONFIG_ETH_STM32_HAL_API_V2) |
| void HAL_ETH_TxCpltCallback(ETH_HandleTypeDef *heth_handle) |
| { |
| __ASSERT_NO_MSG(heth_handle != NULL); |
| |
| struct eth_stm32_hal_dev_data *dev_data = |
| CONTAINER_OF(heth_handle, struct eth_stm32_hal_dev_data, heth); |
| |
| __ASSERT_NO_MSG(dev_data != NULL); |
| |
| k_sem_give(&dev_data->tx_int_sem); |
| |
| } |
| #endif /* CONFIG_SOC_SERIES_STM32H7X || CONFIG_ETH_STM32_HAL_API_V2 */ |
| |
| #if defined(CONFIG_ETH_STM32_HAL_API_V2) |
| void HAL_ETH_ErrorCallback(ETH_HandleTypeDef *heth) |
| { |
| /* Do not log errors. If errors are reported due to high traffic, |
| * logging errors will only increase traffic issues |
| */ |
| #if defined(CONFIG_NET_STATISTICS_ETHERNET) |
| __ASSERT_NO_MSG(heth != NULL); |
| |
| uint32_t dma_error; |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) |
| uint32_t mac_error; |
| #endif /* CONFIG_SOC_SERIES_STM32H7X */ |
| const uint32_t error_code = HAL_ETH_GetError(heth); |
| |
| struct eth_stm32_hal_dev_data *dev_data = |
| CONTAINER_OF(heth, struct eth_stm32_hal_dev_data, heth); |
| |
| switch (error_code) { |
| case HAL_ETH_ERROR_DMA: |
| dma_error = HAL_ETH_GetDMAError(heth); |
| |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) |
| if ((dma_error & ETH_DMA_RX_WATCHDOG_TIMEOUT_FLAG) || |
| (dma_error & ETH_DMA_RX_PROCESS_STOPPED_FLAG) || |
| (dma_error & ETH_DMA_RX_BUFFER_UNAVAILABLE_FLAG)) { |
| eth_stats_update_errors_rx(dev_data->iface); |
| } |
| if ((dma_error & ETH_DMA_EARLY_TX_IT_FLAG) || |
| (dma_error & ETH_DMA_TX_PROCESS_STOPPED_FLAG)) { |
| eth_stats_update_errors_tx(dev_data->iface); |
| } |
| #else |
| if ((dma_error & ETH_DMASR_RWTS) || |
| (dma_error & ETH_DMASR_RPSS) || |
| (dma_error & ETH_DMASR_RBUS)) { |
| eth_stats_update_errors_rx(dev_data->iface); |
| } |
| if ((dma_error & ETH_DMASR_ETS) || |
| (dma_error & ETH_DMASR_TPSS) || |
| (dma_error & ETH_DMASR_TJTS)) { |
| eth_stats_update_errors_tx(dev_data->iface); |
| } |
| #endif /* CONFIG_SOC_SERIES_STM32H7X */ |
| break; |
| |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) |
| case HAL_ETH_ERROR_MAC: |
| mac_error = HAL_ETH_GetMACError(heth); |
| |
| if (mac_error & ETH_RECEIVE_WATCHDOG_TIMEOUT) { |
| eth_stats_update_errors_rx(dev_data->iface); |
| } |
| |
| if ((mac_error & ETH_EXECESSIVE_COLLISIONS) || |
| (mac_error & ETH_LATE_COLLISIONS) || |
| (mac_error & ETH_EXECESSIVE_DEFERRAL) || |
| (mac_error & ETH_TRANSMIT_JABBR_TIMEOUT) || |
| (mac_error & ETH_LOSS_OF_CARRIER) || |
| (mac_error & ETH_NO_CARRIER)) { |
| eth_stats_update_errors_tx(dev_data->iface); |
| } |
| break; |
| #endif /* CONFIG_SOC_SERIES_STM32H7X */ |
| } |
| |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) |
| dev_data->stats.error_details.rx_crc_errors = heth->Instance->MMCRCRCEPR; |
| dev_data->stats.error_details.rx_align_errors = heth->Instance->MMCRAEPR; |
| #else |
| dev_data->stats.error_details.rx_crc_errors = heth->Instance->MMCRFCECR; |
| dev_data->stats.error_details.rx_align_errors = heth->Instance->MMCRFAECR; |
| #endif /* CONFIG_SOC_SERIES_STM32H7X */ |
| |
| #endif /* CONFIG_NET_STATISTICS_ETHERNET */ |
| } |
| #elif defined(CONFIG_SOC_SERIES_STM32H7X) |
| /* DMA and MAC errors callback only appear in H7 series */ |
| void HAL_ETH_DMAErrorCallback(ETH_HandleTypeDef *heth_handle) |
| { |
| __ASSERT_NO_MSG(heth_handle != NULL); |
| |
| LOG_ERR("%s errorcode:%x dmaerror:%x", |
| __func__, |
| HAL_ETH_GetError(heth_handle), |
| HAL_ETH_GetDMAError(heth_handle)); |
| |
| /* State of eth handle is ERROR in case of unrecoverable error */ |
| /* unrecoverable (ETH_DMACSR_FBE | ETH_DMACSR_TPS | ETH_DMACSR_RPS) */ |
| if (HAL_ETH_GetState(heth_handle) == HAL_ETH_STATE_ERROR) { |
| LOG_ERR("%s ethernet in error state", __func__); |
| /* TODO restart the ETH peripheral to recover */ |
| return; |
| } |
| |
| /* Recoverable errors don't put ETH in error state */ |
| /* ETH_DMACSR_CDE | ETH_DMACSR_ETI | ETH_DMACSR_RWT */ |
| /* | ETH_DMACSR_RBU | ETH_DMACSR_AIS) */ |
| |
| /* TODO Check if we were TX transmitting and the unlock semaphore */ |
| /* To return the error as soon as possible else we'll just wait */ |
| /* for the timeout */ |
| |
| |
| } |
| void HAL_ETH_MACErrorCallback(ETH_HandleTypeDef *heth_handle) |
| { |
| __ASSERT_NO_MSG(heth_handle != NULL); |
| |
| /* MAC errors dumping */ |
| LOG_ERR("%s errorcode:%x macerror:%x", |
| __func__, |
| HAL_ETH_GetError(heth_handle), |
| HAL_ETH_GetMACError(heth_handle)); |
| |
| /* State of eth handle is ERROR in case of unrecoverable error */ |
| if (HAL_ETH_GetState(heth_handle) == HAL_ETH_STATE_ERROR) { |
| LOG_ERR("%s ethernet in error state", __func__); |
| /* TODO restart or reconfig ETH peripheral to recover */ |
| |
| return; |
| } |
| } |
| #endif /* CONFIG_ETH_STM32_HAL_API_V2 */ |
| |
| void HAL_ETH_RxCpltCallback(ETH_HandleTypeDef *heth_handle) |
| { |
| __ASSERT_NO_MSG(heth_handle != NULL); |
| |
| struct eth_stm32_hal_dev_data *dev_data = |
| CONTAINER_OF(heth_handle, struct eth_stm32_hal_dev_data, heth); |
| |
| __ASSERT_NO_MSG(dev_data != NULL); |
| |
| k_sem_give(&dev_data->rx_int_sem); |
| } |
| |
| static void generate_mac(uint8_t *mac_addr) |
| { |
| #if defined(ETH_STM32_RANDOM_MAC) |
| /* Either CONFIG_ETH_STM32_HAL_RANDOM_MAC or device tree property */ |
| /* "zephyr,random-mac-address" is set, generate a random mac address */ |
| gen_random_mac(mac_addr, ST_OUI_B0, ST_OUI_B1, ST_OUI_B2); |
| #else /* Use user defined mac address */ |
| mac_addr[0] = ST_OUI_B0; |
| mac_addr[1] = ST_OUI_B1; |
| mac_addr[2] = ST_OUI_B2; |
| #if NODE_HAS_VALID_MAC_ADDR(DT_DRV_INST(0)) |
| mac_addr[3] = NODE_MAC_ADDR_OCTET(DT_DRV_INST(0), 3); |
| mac_addr[4] = NODE_MAC_ADDR_OCTET(DT_DRV_INST(0), 4); |
| mac_addr[5] = NODE_MAC_ADDR_OCTET(DT_DRV_INST(0), 5); |
| #elif defined(CONFIG_ETH_STM32_HAL_USER_STATIC_MAC) |
| mac_addr[3] = CONFIG_ETH_STM32_HAL_MAC3; |
| mac_addr[4] = CONFIG_ETH_STM32_HAL_MAC4; |
| mac_addr[5] = CONFIG_ETH_STM32_HAL_MAC5; |
| #else |
| /* Nothing defined by the user, use device id */ |
| hwinfo_get_device_id(&mac_addr[3], 3); |
| #endif /* NODE_HAS_VALID_MAC_ADDR(DT_DRV_INST(0))) */ |
| #endif |
| } |
| |
| static int eth_initialize(const struct device *dev) |
| { |
| struct eth_stm32_hal_dev_data *dev_data; |
| const struct eth_stm32_hal_dev_cfg *cfg; |
| ETH_HandleTypeDef *heth; |
| HAL_StatusTypeDef hal_ret = HAL_OK; |
| int ret = 0; |
| |
| __ASSERT_NO_MSG(dev != NULL); |
| |
| dev_data = dev->data; |
| cfg = dev->config; |
| |
| __ASSERT_NO_MSG(dev_data != NULL); |
| __ASSERT_NO_MSG(cfg != NULL); |
| |
| dev_data->clock = DEVICE_DT_GET(STM32_CLOCK_CONTROL_NODE); |
| |
| if (!device_is_ready(dev_data->clock)) { |
| LOG_ERR("clock control device not ready"); |
| return -ENODEV; |
| } |
| |
| /* enable clock */ |
| ret = clock_control_on(dev_data->clock, |
| (clock_control_subsys_t)&cfg->pclken); |
| ret |= clock_control_on(dev_data->clock, |
| (clock_control_subsys_t)&cfg->pclken_tx); |
| ret |= clock_control_on(dev_data->clock, |
| (clock_control_subsys_t)&cfg->pclken_rx); |
| #if DT_INST_CLOCKS_HAS_NAME(0, mac_clk_ptp) |
| ret |= clock_control_on(dev_data->clock, |
| (clock_control_subsys_t)&cfg->pclken_ptp); |
| #endif |
| |
| if (ret) { |
| LOG_ERR("Failed to enable ethernet clock"); |
| return -EIO; |
| } |
| |
| /* configure pinmux */ |
| ret = pinctrl_apply_state(cfg->pcfg, PINCTRL_STATE_DEFAULT); |
| if (ret < 0) { |
| LOG_ERR("Could not configure ethernet pins"); |
| return ret; |
| } |
| |
| heth = &dev_data->heth; |
| |
| generate_mac(dev_data->mac_addr); |
| |
| heth->Init.MACAddr = dev_data->mac_addr; |
| |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) || defined(CONFIG_ETH_STM32_HAL_API_V2) |
| heth->Init.TxDesc = dma_tx_desc_tab; |
| heth->Init.RxDesc = dma_rx_desc_tab; |
| heth->Init.RxBuffLen = ETH_STM32_RX_BUF_SIZE; |
| #endif /* CONFIG_SOC_SERIES_STM32H7X || CONFIG_ETH_STM32_HAL_API_V2 */ |
| |
| hal_ret = HAL_ETH_Init(heth); |
| if (hal_ret == HAL_TIMEOUT) { |
| /* HAL Init time out. This could be linked to */ |
| /* a recoverable error. Log the issue and continue */ |
| /* driver initialisation */ |
| LOG_ERR("HAL_ETH_Init Timed out"); |
| } else if (hal_ret != HAL_OK) { |
| LOG_ERR("HAL_ETH_Init failed: %d", hal_ret); |
| return -EINVAL; |
| } |
| |
| #if defined(CONFIG_PTP_CLOCK_STM32_HAL) |
| /* Enable timestamping of RX packets. We enable all packets to be |
| * timestamped to cover both IEEE 1588 and gPTP. |
| */ |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) |
| heth->Instance->MACTSCR |= ETH_MACTSCR_TSENALL; |
| #else |
| heth->Instance->PTPTSCR |= ETH_PTPTSCR_TSSARFE; |
| #endif /* CONFIG_SOC_SERIES_STM32H7X */ |
| #endif /* CONFIG_PTP_CLOCK_STM32_HAL */ |
| |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) || defined(CONFIG_ETH_STM32_HAL_API_V2) |
| /* Tx config init: */ |
| memset(&tx_config, 0, sizeof(ETH_TxPacketConfig)); |
| tx_config.Attributes = ETH_TX_PACKETS_FEATURES_CSUM | |
| ETH_TX_PACKETS_FEATURES_CRCPAD; |
| tx_config.ChecksumCtrl = IS_ENABLED(CONFIG_ETH_STM32_HW_CHECKSUM) ? |
| ETH_CHECKSUM_IPHDR_PAYLOAD_INSERT_PHDR_CALC : ETH_CHECKSUM_DISABLE; |
| tx_config.CRCPadCtrl = ETH_CRC_PAD_INSERT; |
| #endif /* CONFIG_SOC_SERIES_STM32H7X || CONFIG_ETH_STM32_HAL_API_V2 */ |
| |
| dev_data->link_up = false; |
| |
| /* Initialize semaphores */ |
| k_mutex_init(&dev_data->tx_mutex); |
| k_sem_init(&dev_data->rx_int_sem, 0, K_SEM_MAX_LIMIT); |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) || defined(CONFIG_ETH_STM32_HAL_API_V2) |
| k_sem_init(&dev_data->tx_int_sem, 0, K_SEM_MAX_LIMIT); |
| #endif /* CONFIG_SOC_SERIES_STM32H7X || CONFIG_ETH_STM32_HAL_API_V2 */ |
| |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) || defined(CONFIG_ETH_STM32_HAL_API_V2) |
| /* Adjust MDC clock range depending on HCLK frequency: */ |
| HAL_ETH_SetMDIOClockRange(heth); |
| |
| /* @TODO: read duplex mode and speed from PHY and set it to ETH */ |
| |
| ETH_MACConfigTypeDef mac_config; |
| |
| HAL_ETH_GetMACConfig(heth, &mac_config); |
| mac_config.DuplexMode = IS_ENABLED(CONFIG_ETH_STM32_MODE_HALFDUPLEX) ? |
| ETH_HALFDUPLEX_MODE : ETH_FULLDUPLEX_MODE; |
| mac_config.Speed = IS_ENABLED(CONFIG_ETH_STM32_SPEED_10M) ? |
| ETH_SPEED_10M : ETH_SPEED_100M; |
| hal_ret = HAL_ETH_SetMACConfig(heth, &mac_config); |
| if (hal_ret != HAL_OK) { |
| LOG_ERR("HAL_ETH_SetMACConfig: failed: %d", hal_ret); |
| } |
| #endif /* CONFIG_SOC_SERIES_STM32H7X || CONFIG_ETH_STM32_HAL_API_V2 */ |
| |
| #if defined(CONFIG_ETH_STM32_HAL_API_V2) |
| |
| /* prepare tx buffer header */ |
| for (uint16_t i = 0; i < ETH_TXBUFNB; ++i) { |
| dma_tx_buffer_header[i].tx_buff.buffer = dma_tx_buffer[i]; |
| } |
| |
| hal_ret = HAL_ETH_Start_IT(heth); |
| #elif defined(CONFIG_SOC_SERIES_STM32H7X) |
| for (uint32_t i = 0; i < ETH_RX_DESC_CNT; i++) { |
| hal_ret = HAL_ETH_DescAssignMemory(heth, i, dma_rx_buffer[i], |
| NULL); |
| if (hal_ret != HAL_OK) { |
| LOG_ERR("HAL_ETH_DescAssignMemory: failed: %d, i: %d", |
| hal_ret, i); |
| return -EINVAL; |
| } |
| } |
| |
| hal_ret = HAL_ETH_Start_IT(heth); |
| #else |
| HAL_ETH_DMATxDescListInit(heth, dma_tx_desc_tab, |
| &dma_tx_buffer[0][0], ETH_TXBUFNB); |
| HAL_ETH_DMARxDescListInit(heth, dma_rx_desc_tab, |
| &dma_rx_buffer[0][0], ETH_RXBUFNB); |
| |
| hal_ret = HAL_ETH_Start(heth); |
| #endif /* CONFIG_ETH_STM32_HAL_API_V2 */ |
| |
| if (hal_ret != HAL_OK) { |
| LOG_ERR("HAL_ETH_Start{_IT} failed"); |
| } |
| |
| setup_mac_filter(heth); |
| |
| |
| |
| LOG_DBG("MAC %02x:%02x:%02x:%02x:%02x:%02x", |
| dev_data->mac_addr[0], dev_data->mac_addr[1], |
| dev_data->mac_addr[2], dev_data->mac_addr[3], |
| dev_data->mac_addr[4], dev_data->mac_addr[5]); |
| |
| return 0; |
| } |
| |
| #if defined(CONFIG_ETH_STM32_MULTICAST_FILTER) |
| |
| #if defined(CONFIG_NET_NATIVE_IPV6) |
| static void add_ipv6_multicast_addr(const struct in6_addr *addr) |
| { |
| uint32_t i; |
| |
| for (i = 0; i < NET_IF_MAX_IPV6_MADDR; i++) { |
| if (net_ipv6_is_addr_unspecified(&multicast_ipv6_joined_addrs[i])) { |
| net_ipv6_addr_copy_raw((uint8_t *)&multicast_ipv6_joined_addrs[i], |
| (uint8_t *)addr); |
| break; |
| } |
| } |
| } |
| |
| static void remove_ipv6_multicast_addr(const struct in6_addr *addr) |
| { |
| uint32_t i; |
| |
| for (i = 0; i < NET_IF_MAX_IPV6_MADDR; i++) { |
| if (net_ipv6_addr_cmp(&multicast_ipv6_joined_addrs[i], addr)) { |
| net_ipv6_addr_copy_raw((uint8_t *)&multicast_ipv6_joined_addrs[i], |
| (uint8_t *)net_ipv6_unspecified_address); |
| break; |
| } |
| } |
| } |
| #endif /* CONFIG_NET_NATIVE_IPV6 */ |
| |
| #if defined(CONFIG_NET_NATIVE_IPV4) |
| static void add_ipv4_multicast_addr(const struct in_addr *addr) |
| { |
| uint32_t i; |
| |
| for (i = 0; i < NET_IF_MAX_IPV4_MADDR; i++) { |
| if (net_ipv4_is_addr_unspecified(&multicast_ipv4_joined_addrs[i])) { |
| net_ipv4_addr_copy_raw((uint8_t *)&multicast_ipv4_joined_addrs[i], |
| (uint8_t *)addr); |
| break; |
| } |
| } |
| } |
| |
| static void remove_ipv4_multicast_addr(const struct in_addr *addr) |
| { |
| uint32_t i; |
| |
| for (i = 0; i < NET_IF_MAX_IPV4_MADDR; i++) { |
| if (net_ipv4_addr_cmp(&multicast_ipv4_joined_addrs[i], addr)) { |
| multicast_ipv4_joined_addrs[i].s_addr = 0; |
| break; |
| } |
| } |
| } |
| #endif /* CONFIG_NET_NATIVE_IPV4 */ |
| |
| static uint32_t reverse(uint32_t val) |
| { |
| uint32_t res = 0; |
| int i; |
| |
| for (i = 0; i < 32; i++) { |
| if (val & BIT(i)) { |
| res |= BIT(31 - i); |
| } |
| } |
| |
| return res; |
| } |
| |
| static void net_if_stm32_mcast_cb(struct net_if *iface, |
| const struct net_addr *addr, |
| bool is_joined) |
| { |
| ARG_UNUSED(addr); |
| |
| const struct device *dev; |
| struct eth_stm32_hal_dev_data *dev_data; |
| ETH_HandleTypeDef *heth; |
| struct net_eth_addr mac_addr; |
| uint32_t crc; |
| uint32_t hash_table[2]; |
| uint32_t hash_index; |
| int i; |
| |
| dev = net_if_get_device(iface); |
| |
| dev_data = (struct eth_stm32_hal_dev_data *)dev->data; |
| |
| heth = &dev_data->heth; |
| |
| hash_table[0] = 0; |
| hash_table[1] = 0; |
| |
| if (is_joined) { |
| /* Save a copy of the hash table which we update with |
| * the hash for a single multicast address for join |
| */ |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) |
| hash_table[0] = heth->Instance->MACHT0R; |
| hash_table[1] = heth->Instance->MACHT1R; |
| #else |
| hash_table[0] = heth->Instance->MACHTLR; |
| hash_table[1] = heth->Instance->MACHTHR; |
| #endif /* CONFIG_SOC_SERIES_STM32H7X */ |
| } |
| |
| k_mutex_lock(&multicast_addr_lock, K_FOREVER); |
| |
| #if defined(CONFIG_NET_NATIVE_IPV6) |
| if (is_joined) { |
| /* When joining only update the hash filter with the joining |
| * multicast address. |
| */ |
| add_ipv6_multicast_addr(&addr->in6_addr); |
| |
| net_eth_ipv6_mcast_to_mac_addr(&addr->in6_addr, &mac_addr); |
| crc = reverse(crc32_ieee(mac_addr.addr, |
| sizeof(struct net_eth_addr))); |
| hash_index = (crc >> 26) & 0x3f; |
| hash_table[hash_index / 32] |= (1 << (hash_index % 32)); |
| } else { |
| /* When leaving its better to compute the full hash table |
| * for all the multicast addresses that we're aware of. |
| */ |
| remove_ipv6_multicast_addr(&addr->in6_addr); |
| |
| for (i = 0; i < NET_IF_MAX_IPV6_MADDR; i++) { |
| if (net_ipv6_is_addr_unspecified(&multicast_ipv6_joined_addrs[i])) { |
| continue; |
| } |
| |
| net_eth_ipv6_mcast_to_mac_addr(&multicast_ipv6_joined_addrs[i], |
| &mac_addr); |
| crc = reverse(crc32_ieee(mac_addr.addr, |
| sizeof(struct net_eth_addr))); |
| hash_index = (crc >> 26) & 0x3f; |
| hash_table[hash_index / 32] |= (1 << (hash_index % 32)); |
| } |
| } |
| #endif /* CONFIG_NET_IPV6 */ |
| |
| #if defined(CONFIG_NET_NATIVE_IPV4) |
| if (is_joined) { |
| /* When joining only update the hash filter with the joining |
| * multicast address. |
| */ |
| add_ipv4_multicast_addr(&addr->in_addr); |
| |
| net_eth_ipv4_mcast_to_mac_addr(&addr->in_addr, &mac_addr); |
| crc = reverse(crc32_ieee(mac_addr.addr, |
| sizeof(struct net_eth_addr))); |
| hash_index = (crc >> 26) & 0x3f; |
| hash_table[hash_index / 32] |= (1 << (hash_index % 32)); |
| } else { |
| /* When leaving its better to compute the full hash table |
| * for all the multicast addresses that we're aware of. |
| */ |
| remove_ipv4_multicast_addr(&addr->in_addr); |
| |
| for (i = 0; i < NET_IF_MAX_IPV4_MADDR; i++) { |
| if (net_ipv4_is_addr_unspecified(&multicast_ipv4_joined_addrs[i])) { |
| continue; |
| } |
| |
| net_eth_ipv4_mcast_to_mac_addr(&multicast_ipv4_joined_addrs[i], |
| &mac_addr); |
| crc = reverse(crc32_ieee(mac_addr.addr, |
| sizeof(struct net_eth_addr))); |
| hash_index = (crc >> 26) & 0x3f; |
| hash_table[hash_index / 32] |= (1 << (hash_index % 32)); |
| } |
| } |
| #endif /* CONFIG_NET_IPV4 */ |
| |
| k_mutex_unlock(&multicast_addr_lock); |
| |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) |
| heth->Instance->MACHT0R = hash_table[0]; |
| heth->Instance->MACHT1R = hash_table[1]; |
| #else |
| heth->Instance->MACHTLR = hash_table[0]; |
| heth->Instance->MACHTHR = hash_table[1]; |
| #endif /* CONFIG_SOC_SERIES_STM32H7X */ |
| } |
| |
| #endif /* CONFIG_ETH_STM32_MULTICAST_FILTER */ |
| |
| static void eth_iface_init(struct net_if *iface) |
| { |
| const struct device *dev; |
| struct eth_stm32_hal_dev_data *dev_data; |
| bool is_first_init = false; |
| |
| __ASSERT_NO_MSG(iface != NULL); |
| |
| dev = net_if_get_device(iface); |
| __ASSERT_NO_MSG(dev != NULL); |
| |
| dev_data = dev->data; |
| __ASSERT_NO_MSG(dev_data != NULL); |
| |
| /* For VLAN, this value is only used to get the correct L2 driver. |
| * The iface pointer in context should contain the main interface |
| * if the VLANs are enabled. |
| */ |
| if (dev_data->iface == NULL) { |
| dev_data->iface = iface; |
| is_first_init = true; |
| } |
| |
| #if defined(CONFIG_ETH_STM32_MULTICAST_FILTER) |
| net_if_mcast_mon_register(&mcast_monitor, iface, net_if_stm32_mcast_cb); |
| #endif /* CONFIG_ETH_STM32_MULTICAST_FILTER */ |
| |
| /* Register Ethernet MAC Address with the upper layer */ |
| net_if_set_link_addr(iface, dev_data->mac_addr, |
| sizeof(dev_data->mac_addr), |
| NET_LINK_ETHERNET); |
| |
| ethernet_init(iface); |
| |
| net_if_carrier_off(iface); |
| |
| net_lldp_set_lldpdu(iface); |
| |
| if (is_first_init) { |
| const struct eth_stm32_hal_dev_cfg *cfg = dev->config; |
| /* Now that the iface is setup, we are safe to enable IRQs. */ |
| __ASSERT_NO_MSG(cfg->config_func != NULL); |
| cfg->config_func(); |
| |
| /* Start interruption-poll thread */ |
| k_thread_create(&dev_data->rx_thread, dev_data->rx_thread_stack, |
| K_KERNEL_STACK_SIZEOF(dev_data->rx_thread_stack), |
| rx_thread, (void *) dev, NULL, NULL, |
| K_PRIO_COOP(CONFIG_ETH_STM32_HAL_RX_THREAD_PRIO), |
| 0, K_NO_WAIT); |
| |
| k_thread_name_set(&dev_data->rx_thread, "stm_eth"); |
| } |
| } |
| |
| static enum ethernet_hw_caps eth_stm32_hal_get_capabilities(const struct device *dev) |
| { |
| ARG_UNUSED(dev); |
| |
| return ETHERNET_LINK_10BASE_T | ETHERNET_LINK_100BASE_T |
| #if defined(CONFIG_NET_VLAN) |
| | ETHERNET_HW_VLAN |
| #endif |
| #if defined(CONFIG_NET_PROMISCUOUS_MODE) |
| | ETHERNET_PROMISC_MODE |
| #endif |
| #if defined(CONFIG_PTP_CLOCK_STM32_HAL) |
| | ETHERNET_PTP |
| #endif |
| #if defined(CONFIG_NET_LLDP) |
| | ETHERNET_LLDP |
| #endif |
| #if defined(CONFIG_ETH_STM32_HW_CHECKSUM) |
| | ETHERNET_HW_RX_CHKSUM_OFFLOAD |
| | ETHERNET_HW_TX_CHKSUM_OFFLOAD |
| #endif |
| ; |
| } |
| |
| static int eth_stm32_hal_set_config(const struct device *dev, |
| enum ethernet_config_type type, |
| const struct ethernet_config *config) |
| { |
| int ret = -ENOTSUP; |
| struct eth_stm32_hal_dev_data *dev_data; |
| ETH_HandleTypeDef *heth; |
| |
| dev_data = dev->data; |
| heth = &dev_data->heth; |
| |
| switch (type) { |
| case ETHERNET_CONFIG_TYPE_MAC_ADDRESS: |
| memcpy(dev_data->mac_addr, config->mac_address.addr, 6); |
| heth->Instance->MACA0HR = (dev_data->mac_addr[5] << 8) | |
| dev_data->mac_addr[4]; |
| heth->Instance->MACA0LR = (dev_data->mac_addr[3] << 24) | |
| (dev_data->mac_addr[2] << 16) | |
| (dev_data->mac_addr[1] << 8) | |
| dev_data->mac_addr[0]; |
| net_if_set_link_addr(dev_data->iface, dev_data->mac_addr, |
| sizeof(dev_data->mac_addr), |
| NET_LINK_ETHERNET); |
| ret = 0; |
| break; |
| case ETHERNET_CONFIG_TYPE_PROMISC_MODE: |
| #if defined(CONFIG_NET_PROMISCUOUS_MODE) |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) |
| if (config->promisc_mode) { |
| heth->Instance->MACPFR |= ETH_MACPFR_PR; |
| } else { |
| heth->Instance->MACPFR &= ~ETH_MACPFR_PR; |
| } |
| #else |
| if (config->promisc_mode) { |
| heth->Instance->MACFFR |= ETH_MACFFR_PM; |
| } else { |
| heth->Instance->MACFFR &= ~ETH_MACFFR_PM; |
| } |
| #endif /* CONFIG_SOC_SERIES_STM32H7X */ |
| ret = 0; |
| #endif /* CONFIG_NET_PROMISCUOUS_MODE */ |
| break; |
| default: |
| break; |
| } |
| |
| return ret; |
| } |
| |
| #if defined(CONFIG_PTP_CLOCK_STM32_HAL) |
| static const struct device *eth_stm32_get_ptp_clock(const struct device *dev) |
| { |
| struct eth_stm32_hal_dev_data *dev_data = dev->data; |
| |
| return dev_data->ptp_clock; |
| } |
| #endif /* CONFIG_PTP_CLOCK_STM32_HAL */ |
| |
| #if defined(CONFIG_NET_STATISTICS_ETHERNET) |
| static struct net_stats_eth *eth_stm32_hal_get_stats(const struct device *dev) |
| { |
| struct eth_stm32_hal_dev_data *dev_data = dev->data; |
| |
| return &dev_data->stats; |
| } |
| #endif /* CONFIG_NET_STATISTICS_ETHERNET */ |
| |
| static const struct ethernet_api eth_api = { |
| .iface_api.init = eth_iface_init, |
| #if defined(CONFIG_PTP_CLOCK_STM32_HAL) |
| .get_ptp_clock = eth_stm32_get_ptp_clock, |
| #endif /* CONFIG_PTP_CLOCK_STM32_HAL */ |
| .get_capabilities = eth_stm32_hal_get_capabilities, |
| .set_config = eth_stm32_hal_set_config, |
| .send = eth_tx, |
| #if defined(CONFIG_NET_STATISTICS_ETHERNET) |
| .get_stats = eth_stm32_hal_get_stats, |
| #endif /* CONFIG_NET_STATISTICS_ETHERNET */ |
| }; |
| |
| static void eth0_irq_config(void) |
| { |
| IRQ_CONNECT(DT_INST_IRQN(0), DT_INST_IRQ(0, priority), eth_isr, |
| DEVICE_DT_INST_GET(0), 0); |
| irq_enable(DT_INST_IRQN(0)); |
| } |
| |
| PINCTRL_DT_INST_DEFINE(0); |
| |
| static const struct eth_stm32_hal_dev_cfg eth0_config = { |
| .config_func = eth0_irq_config, |
| .pclken = {.bus = DT_INST_CLOCKS_CELL_BY_NAME(0, stmmaceth, bus), |
| .enr = DT_INST_CLOCKS_CELL_BY_NAME(0, stmmaceth, bits)}, |
| .pclken_tx = {.bus = DT_INST_CLOCKS_CELL_BY_NAME(0, mac_clk_tx, bus), |
| .enr = DT_INST_CLOCKS_CELL_BY_NAME(0, mac_clk_tx, bits)}, |
| .pclken_rx = {.bus = DT_INST_CLOCKS_CELL_BY_NAME(0, mac_clk_rx, bus), |
| .enr = DT_INST_CLOCKS_CELL_BY_NAME(0, mac_clk_rx, bits)}, |
| #if DT_INST_CLOCKS_HAS_NAME(0, mac_clk_ptp) |
| .pclken_ptp = {.bus = DT_INST_CLOCKS_CELL_BY_NAME(0, mac_clk_ptp, bus), |
| .enr = DT_INST_CLOCKS_CELL_BY_NAME(0, mac_clk_ptp, bits)}, |
| #endif |
| .pcfg = PINCTRL_DT_INST_DEV_CONFIG_GET(0), |
| }; |
| |
| static struct eth_stm32_hal_dev_data eth0_data = { |
| .heth = { |
| .Instance = (ETH_TypeDef *)DT_INST_REG_ADDR(0), |
| .Init = { |
| #if !defined(CONFIG_SOC_SERIES_STM32H7X) && !defined(CONFIG_ETH_STM32_HAL_API_V2) |
| #if defined(CONFIG_ETH_STM32_AUTO_NEGOTIATION_ENABLE) |
| .AutoNegotiation = ETH_AUTONEGOTIATION_ENABLE, |
| #else |
| .AutoNegotiation = ETH_AUTONEGOTIATION_DISABLE, |
| .Speed = IS_ENABLED(CONFIG_ETH_STM32_SPEED_10M) ? |
| ETH_SPEED_10M : ETH_SPEED_100M, |
| .DuplexMode = IS_ENABLED(CONFIG_ETH_STM32_MODE_HALFDUPLEX) ? |
| ETH_MODE_HALFDUPLEX : ETH_MODE_FULLDUPLEX, |
| #endif /* !CONFIG_ETH_STM32_AUTO_NEGOTIATION_ENABLE */ |
| .PhyAddress = PHY_ADDR, |
| .RxMode = ETH_RXINTERRUPT_MODE, |
| .ChecksumMode = IS_ENABLED(CONFIG_ETH_STM32_HW_CHECKSUM) ? |
| ETH_CHECKSUM_BY_HARDWARE : ETH_CHECKSUM_BY_SOFTWARE, |
| #endif /* !CONFIG_SOC_SERIES_STM32H7X */ |
| .MediaInterface = IS_ENABLED(CONFIG_ETH_STM32_HAL_MII) ? |
| ETH_MEDIA_INTERFACE_MII : ETH_MEDIA_INTERFACE_RMII, |
| }, |
| }, |
| }; |
| |
| ETH_NET_DEVICE_DT_INST_DEFINE(0, eth_initialize, |
| NULL, ð0_data, ð0_config, |
| CONFIG_ETH_INIT_PRIORITY, ð_api, ETH_STM32_HAL_MTU); |
| |
| #if defined(CONFIG_PTP_CLOCK_STM32_HAL) |
| |
| struct ptp_context { |
| struct eth_stm32_hal_dev_data *eth_dev_data; |
| }; |
| |
| static struct ptp_context ptp_stm32_0_context; |
| |
| static int ptp_clock_stm32_set(const struct device *dev, |
| struct net_ptp_time *tm) |
| { |
| struct ptp_context *ptp_context = dev->data; |
| struct eth_stm32_hal_dev_data *eth_dev_data = ptp_context->eth_dev_data; |
| ETH_HandleTypeDef *heth = ð_dev_data->heth; |
| unsigned int key; |
| |
| key = irq_lock(); |
| |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) |
| heth->Instance->MACSTSUR = tm->second; |
| heth->Instance->MACSTNUR = tm->nanosecond; |
| heth->Instance->MACTSCR |= ETH_MACTSCR_TSINIT; |
| while (heth->Instance->MACTSCR & ETH_MACTSCR_TSINIT_Msk) { |
| /* spin lock */ |
| } |
| #else |
| heth->Instance->PTPTSHUR = tm->second; |
| heth->Instance->PTPTSLUR = tm->nanosecond; |
| heth->Instance->PTPTSCR |= ETH_PTPTSCR_TSSTI; |
| while (heth->Instance->PTPTSCR & ETH_PTPTSCR_TSSTI_Msk) { |
| /* spin lock */ |
| } |
| #endif /* CONFIG_SOC_SERIES_STM32H7X */ |
| |
| irq_unlock(key); |
| |
| return 0; |
| } |
| |
| static int ptp_clock_stm32_get(const struct device *dev, |
| struct net_ptp_time *tm) |
| { |
| struct ptp_context *ptp_context = dev->data; |
| struct eth_stm32_hal_dev_data *eth_dev_data = ptp_context->eth_dev_data; |
| ETH_HandleTypeDef *heth = ð_dev_data->heth; |
| unsigned int key; |
| uint32_t second_2; |
| |
| key = irq_lock(); |
| |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) |
| tm->second = heth->Instance->MACSTSR; |
| tm->nanosecond = heth->Instance->MACSTNR; |
| second_2 = heth->Instance->MACSTSR; |
| #else |
| tm->second = heth->Instance->PTPTSHR; |
| tm->nanosecond = heth->Instance->PTPTSLR; |
| second_2 = heth->Instance->PTPTSHR; |
| #endif /* CONFIG_SOC_SERIES_STM32H7X */ |
| |
| irq_unlock(key); |
| |
| if (tm->second != second_2 && tm->nanosecond < NSEC_PER_SEC / 2) { |
| /* Second rollover has happened during first measurement: second register |
| * was read before second boundary and nanosecond register was read after. |
| * We will use second_2 as a new second value. |
| */ |
| tm->second = second_2; |
| } |
| |
| return 0; |
| } |
| |
| static int ptp_clock_stm32_adjust(const struct device *dev, int increment) |
| { |
| struct ptp_context *ptp_context = dev->data; |
| struct eth_stm32_hal_dev_data *eth_dev_data = ptp_context->eth_dev_data; |
| ETH_HandleTypeDef *heth = ð_dev_data->heth; |
| int key, ret; |
| |
| if ((increment <= (int32_t)(-NSEC_PER_SEC)) || |
| (increment >= (int32_t)NSEC_PER_SEC)) { |
| ret = -EINVAL; |
| } else { |
| key = irq_lock(); |
| |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) |
| heth->Instance->MACSTSUR = 0; |
| if (increment >= 0) { |
| heth->Instance->MACSTNUR = increment; |
| } else { |
| heth->Instance->MACSTNUR = ETH_MACSTNUR_ADDSUB | (NSEC_PER_SEC + increment); |
| } |
| heth->Instance->MACTSCR |= ETH_MACTSCR_TSUPDT; |
| while (heth->Instance->MACTSCR & ETH_MACTSCR_TSUPDT_Msk) { |
| /* spin lock */ |
| } |
| #else |
| heth->Instance->PTPTSHUR = 0; |
| if (increment >= 0) { |
| heth->Instance->PTPTSLUR = increment; |
| } else { |
| heth->Instance->PTPTSLUR = ETH_PTPTSLUR_TSUPNS | (-increment); |
| } |
| heth->Instance->PTPTSCR |= ETH_PTPTSCR_TSSTU; |
| while (heth->Instance->PTPTSCR & ETH_PTPTSCR_TSSTU_Msk) { |
| /* spin lock */ |
| } |
| #endif /* CONFIG_SOC_SERIES_STM32H7X */ |
| |
| ret = 0; |
| irq_unlock(key); |
| } |
| |
| return ret; |
| } |
| |
| static int ptp_clock_stm32_rate_adjust(const struct device *dev, double ratio) |
| { |
| struct ptp_context *ptp_context = dev->data; |
| struct eth_stm32_hal_dev_data *eth_dev_data = ptp_context->eth_dev_data; |
| ETH_HandleTypeDef *heth = ð_dev_data->heth; |
| int key, ret; |
| uint32_t addend_val; |
| |
| /* No change needed */ |
| if (ratio == 1.0f) { |
| return 0; |
| } |
| |
| key = irq_lock(); |
| |
| ratio *= eth_dev_data->clk_ratio_adj; |
| |
| /* Limit possible ratio */ |
| if (ratio * 100 < CONFIG_ETH_STM32_HAL_PTP_CLOCK_ADJ_MIN_PCT || |
| ratio * 100 > CONFIG_ETH_STM32_HAL_PTP_CLOCK_ADJ_MAX_PCT) { |
| ret = -EINVAL; |
| goto error; |
| } |
| |
| /* Save new ratio */ |
| eth_dev_data->clk_ratio_adj = ratio; |
| |
| /* Update addend register */ |
| addend_val = UINT32_MAX * eth_dev_data->clk_ratio * ratio; |
| |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) |
| heth->Instance->MACTSAR = addend_val; |
| heth->Instance->MACTSCR |= ETH_MACTSCR_TSADDREG; |
| while (heth->Instance->MACTSCR & ETH_MACTSCR_TSADDREG_Msk) { |
| /* spin lock */ |
| } |
| #else |
| heth->Instance->PTPTSAR = addend_val; |
| heth->Instance->PTPTSCR |= ETH_PTPTSCR_TSARU; |
| while (heth->Instance->PTPTSCR & ETH_PTPTSCR_TSARU_Msk) { |
| /* spin lock */ |
| } |
| #endif /* CONFIG_SOC_SERIES_STM32H7X */ |
| |
| ret = 0; |
| |
| error: |
| irq_unlock(key); |
| |
| return ret; |
| } |
| |
| static const struct ptp_clock_driver_api api = { |
| .set = ptp_clock_stm32_set, |
| .get = ptp_clock_stm32_get, |
| .adjust = ptp_clock_stm32_adjust, |
| .rate_adjust = ptp_clock_stm32_rate_adjust, |
| }; |
| |
| static int ptp_stm32_init(const struct device *port) |
| { |
| const struct device *const dev = DEVICE_DT_GET(DT_NODELABEL(mac)); |
| struct eth_stm32_hal_dev_data *eth_dev_data = dev->data; |
| const struct eth_stm32_hal_dev_cfg *eth_cfg = dev->config; |
| struct ptp_context *ptp_context = port->data; |
| ETH_HandleTypeDef *heth = ð_dev_data->heth; |
| int ret; |
| uint32_t ptp_hclk_rate; |
| uint32_t ss_incr_ns; |
| uint32_t addend_val; |
| |
| eth_dev_data->ptp_clock = port; |
| ptp_context->eth_dev_data = eth_dev_data; |
| |
| /* Mask the Timestamp Trigger interrupt */ |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) |
| heth->Instance->MACIER &= ~(ETH_MACIER_TSIE); |
| #else |
| heth->Instance->MACIMR &= ~(ETH_MACIMR_TSTIM); |
| #endif /* CONFIG_SOC_SERIES_STM32H7X */ |
| |
| /* Enable timestamping */ |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) |
| heth->Instance->MACTSCR |= ETH_MACTSCR_TSENA; |
| #else |
| heth->Instance->PTPTSCR |= ETH_PTPTSCR_TSE; |
| #endif /* CONFIG_SOC_SERIES_STM32H7X */ |
| |
| /* Query ethernet clock rate */ |
| ret = clock_control_get_rate(eth_dev_data->clock, |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) |
| (clock_control_subsys_t)ð_cfg->pclken, |
| #else |
| (clock_control_subsys_t)ð_cfg->pclken_ptp, |
| #endif /* CONFIG_SOC_SERIES_STM32H7X */ |
| &ptp_hclk_rate); |
| if (ret) { |
| LOG_ERR("Failed to query ethernet clock"); |
| return -EIO; |
| } |
| |
| /* Program the subsecond increment register based on the PTP clock freq */ |
| if (NSEC_PER_SEC % CONFIG_ETH_STM32_HAL_PTP_CLOCK_SRC_HZ != 0) { |
| LOG_ERR("PTP clock period must be an integer nanosecond value"); |
| return -EINVAL; |
| } |
| ss_incr_ns = NSEC_PER_SEC / CONFIG_ETH_STM32_HAL_PTP_CLOCK_SRC_HZ; |
| if (ss_incr_ns > UINT8_MAX) { |
| LOG_ERR("PTP clock period is more than %d nanoseconds", UINT8_MAX); |
| return -EINVAL; |
| } |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) |
| heth->Instance->MACSSIR = ss_incr_ns << ETH_MACMACSSIR_SSINC_Pos; |
| #else |
| heth->Instance->PTPSSIR = ss_incr_ns; |
| #endif /* CONFIG_SOC_SERIES_STM32H7X */ |
| |
| /* Program timestamp addend register */ |
| eth_dev_data->clk_ratio = |
| ((double)CONFIG_ETH_STM32_HAL_PTP_CLOCK_SRC_HZ) / ((double)ptp_hclk_rate); |
| /* |
| * clk_ratio is a ratio between desired PTP clock frequency and HCLK rate. |
| * Because HCLK is defined by a physical oscillator, it might drift due |
| * to manufacturing tolerances and environmental effects (e.g. temperature). |
| * clk_ratio_adj compensates for such inaccuracies. It starts off as 1.0 |
| * and gets adjusted by calling ptp_clock_stm32_rate_adjust(). |
| */ |
| eth_dev_data->clk_ratio_adj = 1.0f; |
| addend_val = |
| UINT32_MAX * eth_dev_data->clk_ratio * eth_dev_data->clk_ratio_adj; |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) |
| heth->Instance->MACTSAR = addend_val; |
| heth->Instance->MACTSCR |= ETH_MACTSCR_TSADDREG; |
| while (heth->Instance->MACTSCR & ETH_MACTSCR_TSADDREG_Msk) { |
| k_yield(); |
| } |
| #else |
| heth->Instance->PTPTSAR = addend_val; |
| heth->Instance->PTPTSCR |= ETH_PTPTSCR_TSARU; |
| while (heth->Instance->PTPTSCR & ETH_PTPTSCR_TSARU_Msk) { |
| k_yield(); |
| } |
| #endif /* CONFIG_SOC_SERIES_STM32H7X */ |
| |
| /* Enable fine timestamp correction method */ |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) |
| heth->Instance->MACTSCR |= ETH_MACTSCR_TSCFUPDT; |
| #else |
| heth->Instance->PTPTSCR |= ETH_PTPTSCR_TSFCU; |
| #endif /* CONFIG_SOC_SERIES_STM32H7X */ |
| |
| /* Enable nanosecond rollover into a new second */ |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) |
| heth->Instance->MACTSCR |= ETH_MACTSCR_TSCTRLSSR; |
| #else |
| heth->Instance->PTPTSCR |= ETH_PTPTSCR_TSSSR; |
| #endif /* CONFIG_SOC_SERIES_STM32H7X */ |
| |
| /* Initialize timestamp */ |
| #if defined(CONFIG_SOC_SERIES_STM32H7X) |
| heth->Instance->MACSTSUR = 0; |
| heth->Instance->MACSTNUR = 0; |
| heth->Instance->MACTSCR |= ETH_MACTSCR_TSINIT; |
| while (heth->Instance->MACTSCR & ETH_MACTSCR_TSINIT_Msk) { |
| k_yield(); |
| } |
| #else |
| heth->Instance->PTPTSHUR = 0; |
| heth->Instance->PTPTSLUR = 0; |
| heth->Instance->PTPTSCR |= ETH_PTPTSCR_TSSTI; |
| while (heth->Instance->PTPTSCR & ETH_PTPTSCR_TSSTI_Msk) { |
| k_yield(); |
| } |
| #endif /* CONFIG_SOC_SERIES_STM32H7X */ |
| |
| #if defined(CONFIG_ETH_STM32_HAL_API_V2) |
| /* Set PTP Configuration done */ |
| heth->IsPtpConfigured = HAL_ETH_PTP_CONFIGURATED; |
| #endif |
| |
| return 0; |
| } |
| |
| DEVICE_DEFINE(stm32_ptp_clock_0, PTP_CLOCK_NAME, ptp_stm32_init, |
| NULL, &ptp_stm32_0_context, NULL, POST_KERNEL, |
| CONFIG_ETH_STM32_HAL_PTP_CLOCK_INIT_PRIO, &api); |
| |
| #endif /* CONFIG_PTP_CLOCK_STM32_HAL */ |