| /* |
| * Copyright (c) 2017 Intel Corporation. |
| * |
| * SPDX-License-Identifier: Apache-2.0 |
| */ |
| #define LOG_MODULE_NAME eth_dw |
| #define LOG_LEVEL CONFIG_ETHERNET_LOG_LEVEL |
| |
| #include <logging/log.h> |
| LOG_MODULE_REGISTER(LOG_MODULE_NAME); |
| |
| #include <soc.h> |
| #include <device.h> |
| #include <errno.h> |
| #include <init.h> |
| #include <kernel.h> |
| #include <misc/__assert.h> |
| #include <net/net_core.h> |
| #include <net/net_pkt.h> |
| #include <stdbool.h> |
| #include <stdio.h> |
| #include <string.h> |
| #include <sys_io.h> |
| #include <net/ethernet.h> |
| #include <ethernet/eth_stats.h> |
| |
| #include "eth_dw_priv.h" |
| |
| #ifdef CONFIG_SHARED_IRQ |
| #include <shared_irq.h> |
| #endif |
| |
| #define TX_BUSY_LOOP_SPINS 20 |
| |
| static inline u32_t eth_read(u32_t base_addr, u32_t offset) |
| { |
| return sys_read32(base_addr + offset); |
| } |
| |
| static inline void eth_write(u32_t base_addr, u32_t offset, |
| u32_t val) |
| { |
| sys_write32(val, base_addr + offset); |
| } |
| |
| static void eth_rx(struct device *dev) |
| { |
| struct eth_runtime *context = dev->driver_data; |
| struct net_pkt *pkt; |
| u32_t frm_len; |
| int r; |
| |
| /* Check whether the RX descriptor is still owned by the device. If not, |
| * process the received frame or an error that may have occurred. |
| */ |
| if (context->rx_desc.own) { |
| LOG_ERR("Spurious receive interrupt from Ethernet MAC"); |
| return; |
| } |
| |
| if (context->rx_desc.err_summary) { |
| LOG_ERR("Error receiving frame: RDES0 = %08x, RDES1 = %08x", |
| context->rx_desc.rdes0, context->rx_desc.rdes1); |
| goto release_desc; |
| } |
| |
| frm_len = context->rx_desc.frm_len; |
| if (frm_len > sizeof(context->rx_buf)) { |
| LOG_ERR("Frame too large: %u", frm_len); |
| goto release_desc; |
| } |
| |
| /* Throw away the last 4 bytes (CRC). See IntelĀ® Quark TM SoC X1000 |
| * datasheet, Table 95 (Receive Descriptor Fields (RDES0)), "frame |
| * length": |
| * These bits indicate the byte length of the received frame that |
| * was transferred to host memory (including CRC). |
| * If the CRC is not removed here, packet processing in upper layers |
| * will fail since the packet length will be different from the |
| * received frame length by exactly 4 bytes. |
| */ |
| if (frm_len < sizeof(u32_t)) { |
| LOG_ERR("Frame too small: %u", frm_len); |
| goto error; |
| } else { |
| frm_len -= sizeof(u32_t); |
| } |
| |
| pkt = net_pkt_rx_alloc_with_buffer(context->iface, frm_len, |
| AF_UNSPEC, 0, K_NO_WAIT); |
| if (!pkt) { |
| LOG_ERR("Failed to obtain RX buffer"); |
| goto error; |
| } |
| |
| if (net_pkt_write(pkt, (void *)context->rx_buf, frm_len)) { |
| LOG_ERR("Failed to append RX buffer to context buffer"); |
| net_pkt_unref(pkt); |
| goto error; |
| } |
| |
| r = net_recv_data(context->iface, pkt); |
| if (r < 0) { |
| LOG_ERR("Failed to enqueue frame into RX queue: %d", r); |
| net_pkt_unref(pkt); |
| goto error; |
| } |
| |
| goto release_desc; |
| error: |
| eth_stats_update_errors_rx(context->iface); |
| release_desc: |
| /* Return ownership of the RX descriptor to the device. */ |
| context->rx_desc.own = 1U; |
| |
| /* Request that the device check for an available RX descriptor, since |
| * ownership of the descriptor was just transferred to the device. |
| */ |
| eth_write(context->base_addr, REG_ADDR_RX_POLL_DEMAND, 1); |
| } |
| |
| static void eth_tx_spin_wait(struct eth_runtime *context) |
| { |
| int spins; |
| |
| for (spins = 0; spins < TX_BUSY_LOOP_SPINS; spins++) { |
| if (!context->tx_desc.own) { |
| return; |
| } |
| } |
| |
| while (context->tx_desc.own) { |
| k_yield(); |
| } |
| } |
| |
| static void eth_tx_data(struct eth_runtime *context, u8_t *data, u16_t len) |
| { |
| #if CONFIG_ETHERNET_LOG_LEVEL >= LOG_LEVEL_DBG |
| /* Check whether an error occurred transmitting the previous frame. */ |
| if (context->tx_desc.err_summary) { |
| LOG_ERR("Error transmitting frame: TDES0 = %08x," |
| "TDES1 = %08x", context->tx_desc.tdes0, |
| context->tx_desc.tdes1); |
| } |
| #endif |
| |
| /* Update transmit descriptor. */ |
| context->tx_desc.buf1_ptr = data; |
| context->tx_desc.tx_buf1_sz = len; |
| eth_write(context->base_addr, REG_ADDR_TX_DESC_LIST, |
| (u32_t)&context->tx_desc); |
| |
| context->tx_desc.own = 1U; |
| |
| /* Request that the device check for an available TX descriptor, since |
| * ownership of the descriptor was just transferred to the device. |
| */ |
| eth_write(context->base_addr, REG_ADDR_TX_POLL_DEMAND, 1); |
| |
| /* Ensure DMA transfer has been completed. */ |
| eth_tx_spin_wait(context); |
| } |
| |
| /* @brief Transmit the current Ethernet frame. |
| * |
| * This procedure will block indefinitely until all fragments from a |
| * net_buf have been transmitted. Data is copied using DMA directly |
| * from each fragment's data pointer. This procedure might yield to |
| * other threads while waiting for the DMA transfer to finish. |
| */ |
| static int eth_tx(struct device *dev, struct net_pkt *pkt) |
| { |
| struct eth_runtime *context = dev->driver_data; |
| struct net_buf *frag; |
| |
| /* Ensure we're clear to transmit. */ |
| eth_tx_spin_wait(context); |
| |
| for (frag = pkt->frags; frag; frag = frag->frags) { |
| eth_tx_data(context, frag->data, frag->len); |
| } |
| |
| return 0; |
| } |
| |
| static void eth_dw_isr(struct device *dev) |
| { |
| struct eth_runtime *context = dev->driver_data; |
| #ifdef CONFIG_SHARED_IRQ |
| u32_t int_status; |
| |
| int_status = eth_read(context->base_addr, REG_ADDR_STATUS); |
| |
| /* If using with shared IRQ, this function will be called |
| * by the shared IRQ driver. So check here if the interrupt |
| * is coming from the GPIO controller (or somewhere else). |
| */ |
| if ((int_status & STATUS_RX_INT) == 0U) { |
| return; |
| } |
| #endif |
| eth_rx(dev); |
| |
| /* Acknowledge the interrupt. */ |
| eth_write(context->base_addr, REG_ADDR_STATUS, |
| STATUS_NORMAL_INT | STATUS_RX_INT); |
| } |
| |
| #ifdef CONFIG_PCI |
| static inline int eth_setup(struct device *dev) |
| { |
| struct eth_runtime *context = dev->driver_data; |
| |
| pci_bus_scan_init(); |
| |
| if (!pci_bus_scan(&context->pci_dev)) |
| return 0; |
| |
| #ifdef CONFIG_PCI_ENUMERATION |
| context->base_addr = context->pci_dev.addr; |
| #endif |
| pci_enable_regs(&context->pci_dev); |
| pci_enable_bus_master(&context->pci_dev); |
| |
| pci_show(&context->pci_dev); |
| |
| return 1; |
| } |
| #else |
| #define eth_setup(_unused_) (1) |
| #endif /* CONFIG_PCI */ |
| |
| static int eth_initialize_internal(struct net_if *iface) |
| { |
| struct device *dev = net_if_get_device(iface); |
| struct eth_runtime *context = dev->driver_data; |
| const struct eth_config *config = dev->config->config_info; |
| u32_t base_addr; |
| |
| context->iface = iface; |
| base_addr = context->base_addr; |
| |
| /* Read the MAC address from the device. */ |
| context->mac_addr.words[1] = eth_read(base_addr, REG_ADDR_MACADDR_HI); |
| context->mac_addr.words[0] = eth_read(base_addr, REG_ADDR_MACADDR_LO); |
| |
| net_if_set_link_addr(context->iface, context->mac_addr.bytes, |
| sizeof(context->mac_addr.bytes), |
| NET_LINK_ETHERNET); |
| |
| /* Initialize the frame filter enabling unicast messages */ |
| eth_write(base_addr, REG_ADDR_MAC_FRAME_FILTER, MAC_FILTER_4_PM); |
| |
| |
| /* Initialize receive descriptor. */ |
| context->rx_desc.rdes0 = 0U; |
| context->rx_desc.rdes1 = 0U; |
| |
| context->rx_desc.buf1_ptr = (u8_t *)context->rx_buf; |
| context->rx_desc.first_desc = 1U; |
| context->rx_desc.last_desc = 1U; |
| context->rx_desc.own = 1U; |
| context->rx_desc.rx_buf1_sz = sizeof(context->rx_buf); |
| context->rx_desc.rx_end_of_ring = 1U; |
| |
| /* Install receive descriptor. */ |
| eth_write(base_addr, REG_ADDR_RX_DESC_LIST, (u32_t)&context->rx_desc); |
| |
| /* Initialize transmit descriptor. */ |
| context->tx_desc.tdes0 = 0U; |
| context->tx_desc.tdes1 = 0U; |
| context->tx_desc.buf1_ptr = NULL; |
| context->tx_desc.tx_buf1_sz = 0U; |
| context->tx_desc.first_seg_in_frm = 1U; |
| context->tx_desc.last_seg_in_frm = 1U; |
| context->tx_desc.tx_end_of_ring = 1U; |
| |
| /* Install transmit descriptor. */ |
| eth_write(context->base_addr, REG_ADDR_TX_DESC_LIST, |
| (u32_t)&context->tx_desc); |
| |
| eth_write(base_addr, REG_ADDR_MAC_CONF, |
| /* Set the RMII speed to 100Mbps */ |
| MAC_CONF_14_RMII_100M | |
| /* Enable full-duplex mode */ |
| MAC_CONF_11_DUPLEX | |
| /* Enable transmitter */ |
| MAC_CONF_3_TX_EN | |
| /* Enable receiver */ |
| MAC_CONF_2_RX_EN); |
| |
| eth_write(base_addr, REG_ADDR_INT_ENABLE, |
| INT_ENABLE_NORMAL | |
| /* Enable receive interrupts */ |
| INT_ENABLE_RX); |
| |
| /* Mask all the MMC interrupts */ |
| eth_write(base_addr, REG_MMC_RX_INTR_MASK, MMC_DEFAULT_MASK); |
| eth_write(base_addr, REG_MMC_TX_INTR_MASK, MMC_DEFAULT_MASK); |
| eth_write(base_addr, REG_MMC_RX_IPC_INTR_MASK, MMC_DEFAULT_MASK); |
| |
| eth_write(base_addr, REG_ADDR_DMA_OPERATION, |
| /* Enable receive store-and-forward mode for simplicity. */ |
| OP_MODE_25_RX_STORE_N_FORWARD | |
| /* Enable transmit store-and-forward mode for simplicity. */ |
| OP_MODE_21_TX_STORE_N_FORWARD | |
| /* Place the transmitter state machine in the Running state. */ |
| OP_MODE_13_START_TX | |
| /* Place the receiver state machine in the Running state. */ |
| OP_MODE_1_START_RX); |
| |
| LOG_INF("Enabled 100M full-duplex mode"); |
| |
| config->config_func(dev); |
| |
| return 0; |
| } |
| |
| static void eth_initialize(struct net_if *iface) |
| { |
| int r = eth_initialize_internal(iface); |
| |
| if (r < 0) { |
| LOG_ERR("Could not initialize ethernet device: %d", r); |
| } |
| } |
| |
| static enum ethernet_hw_caps eth_dw_get_capabilities(struct device *dev) |
| { |
| ARG_UNUSED(dev); |
| |
| return ETHERNET_LINK_10BASE_T | ETHERNET_LINK_100BASE_T; |
| } |
| |
| static const struct ethernet_api api_funcs = { |
| .iface_api.init = eth_initialize, |
| |
| .get_capabilities = eth_dw_get_capabilities, |
| .send = eth_tx, |
| }; |
| |
| /* Bindings to the plaform */ |
| #if CONFIG_ETH_DW_0 |
| static void eth_config_0_irq(struct device *dev) |
| { |
| const struct eth_config *config = dev->config->config_info; |
| struct device *shared_irq_dev; |
| |
| #ifdef CONFIG_ETH_DW_0_IRQ_DIRECT |
| ARG_UNUSED(shared_irq_dev); |
| IRQ_CONNECT(ETH_DW_0_IRQ, CONFIG_ETH_DW_0_IRQ_PRI, eth_dw_isr, |
| DEVICE_GET(eth_dw_0), 0); |
| irq_enable(ETH_DW_0_IRQ); |
| #elif defined(CONFIG_ETH_DW_0_IRQ_SHARED) |
| shared_irq_dev = device_get_binding(config->shared_irq_dev_name); |
| __ASSERT(shared_irq_dev != NULL, "Failed to get eth_dw device binding"); |
| shared_irq_isr_register(shared_irq_dev, (isr_t)eth_dw_isr, dev); |
| shared_irq_enable(shared_irq_dev, dev); |
| #endif |
| } |
| |
| static const struct eth_config eth_config_0 = { |
| #ifdef CONFIG_ETH_DW_0_IRQ_DIRECT |
| .irq_num = ETH_DW_0_IRQ, |
| #endif |
| .config_func = eth_config_0_irq, |
| |
| #ifdef CONFIG_ETH_DW_0_IRQ_SHARED |
| .shared_irq_dev_name = DT_ETH_DW_0_IRQ_SHARED_NAME, |
| #endif |
| }; |
| |
| static struct eth_runtime eth_0_runtime = { |
| .base_addr = ETH_DW_0_BASE_ADDR, |
| #if CONFIG_PCI |
| .pci_dev.class_type = ETH_DW_PCI_CLASS, |
| .pci_dev.bus = ETH_DW_0_PCI_BUS, |
| .pci_dev.dev = ETH_DW_0_PCI_DEV, |
| .pci_dev.vendor_id = ETH_DW_PCI_VENDOR_ID, |
| .pci_dev.device_id = ETH_DW_PCI_DEVICE_ID, |
| .pci_dev.function = ETH_DW_0_PCI_FUNCTION, |
| .pci_dev.bar = ETH_DW_0_PCI_BAR, |
| #endif |
| }; |
| |
| NET_DEVICE_INIT(eth_dw_0, CONFIG_ETH_DW_0_NAME, |
| eth_setup, ð_0_runtime, |
| ð_config_0, CONFIG_ETH_INIT_PRIORITY, &api_funcs, |
| ETHERNET_L2, NET_L2_GET_CTX_TYPE(ETHERNET_L2), |
| NET_ETH_MTU); |
| #endif /* CONFIG_ETH_DW_0 */ |