@@ -169,6 +169,7 @@ source "drivers/net/ethernet/sis/Kconfig"
source "drivers/net/ethernet/sfc/Kconfig"
source "drivers/net/ethernet/sgi/Kconfig"
source "drivers/net/ethernet/smsc/Kconfig"
+source "drivers/net/ethernet/socionext/Kconfig"
source "drivers/net/ethernet/stmicro/Kconfig"
source "drivers/net/ethernet/sun/Kconfig"
source "drivers/net/ethernet/tehuti/Kconfig"
@@ -80,6 +80,7 @@ obj-$(CONFIG_SFC) += sfc/
obj-$(CONFIG_SFC_FALCON) += sfc/falcon/
obj-$(CONFIG_NET_VENDOR_SGI) += sgi/
obj-$(CONFIG_NET_VENDOR_SMSC) += smsc/
+obj-$(CONFIG_NET_VENDOR_SNI) += socionext/
obj-$(CONFIG_NET_VENDOR_STMICRO) += stmicro/
obj-$(CONFIG_NET_VENDOR_SUN) += sun/
obj-$(CONFIG_NET_VENDOR_TEHUTI) += tehuti/
new file mode 100644
@@ -0,0 +1,29 @@
+#
+# Socionext Network device configuration
+#
+
+config NET_VENDOR_SNI
+ bool "Socionext devices"
+ default y
+ ---help---
+ If you have a network (Ethernet) card belonging to this class, say Y.
+
+ Note that the answer to this question doesn't directly affect the
+ the questions about Socionext cards. If you say Y, you will be asked for
+ your specific card in the following questions.
+
+if NET_VENDOR_SNI
+
+config SNI_NETSEC
+ tristate "NETSEC Driver Support"
+ depends on OF
+ select PHYLIB
+ select MII
+help
+ Enable for NETSEC support of Socionext FGAMC4 IP
+ Provides Gigabit ethernet support
+
+ To compile this driver as a module, choose M here: the module will be
+ called netsec. If unsure, say N.
+
+endif # NET_VENDOR_SNI
new file mode 100644
@@ -0,0 +1 @@
+obj-$(CONFIG_SNI_NETSEC) += netsec/
new file mode 100644
@@ -0,0 +1,6 @@
+obj-m := netsec.o
+netsec-objs := netsec_desc_ring_access.o \
+ netsec_netdev.o \
+ netsec_ethtool.o \
+ netsec_platform.o \
+ netsec_gmac_access.o
new file mode 100644
@@ -0,0 +1,386 @@
+/**
+ * netsec.h
+ *
+ * Copyright (C) 2011 - 2014 Fujitsu Semiconductor Limited.
+ * Copyright (C) 2014 Linaro Ltd Andy Green <andy.green@linaro.org>
+ * All rights reserved.
+ *
+ * This program is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU General Public License
+ * as published by the Free Software Foundation; either version 2
+ * of the License, or (at your option) any later version.
+ */
+#ifndef NETSEC_INTERNAL_H
+#define NETSEC_INTERNAL_H
+
+#include <linux/netdevice.h>
+#include <linux/types.h>
+#include <linux/device.h>
+#include <linux/phy.h>
+#include <linux/ethtool.h>
+#include <linux/of_address.h>
+#include <linux/of_mdio.h>
+#include <linux/etherdevice.h>
+#include <net/sock.h>
+
+#define NETSEC_FLOW_CONTROL_START_THRESHOLD 36
+#define NETSEC_FLOW_CONTROL_STOP_THRESHOLD 48
+
+#define NETSEC_CLK_MHZ 1000000
+
+#define NETSEC_RX_PKT_BUF_LEN 1522
+#define NETSEC_RX_JUMBO_PKT_BUF_LEN 9022
+
+#define NETSEC_NETDEV_TX_PKT_SCAT_NUM_MAX 19
+
+#define DESC_NUM 128
+
+#define NETSEC_TX_SHIFT_OWN_FIELD 31
+#define NETSEC_TX_SHIFT_LD_FIELD 30
+#define NETSEC_TX_SHIFT_DRID_FIELD 24
+#define NETSEC_TX_SHIFT_PT_FIELD 21
+#define NETSEC_TX_SHIFT_TDRID_FIELD 16
+#define NETSEC_TX_SHIFT_CC_FIELD 15
+#define NETSEC_TX_SHIFT_FS_FIELD 9
+#define NETSEC_TX_LAST 8
+#define NETSEC_TX_SHIFT_CO 7
+#define NETSEC_TX_SHIFT_SO 6
+#define NETSEC_TX_SHIFT_TRS_FIELD 4
+
+#define NETSEC_RX_PKT_OWN_FIELD 31
+#define NETSEC_RX_PKT_LD_FIELD 30
+#define NETSEC_RX_PKT_SDRID_FIELD 24
+#define NETSEC_RX_PKT_FR_FIELD 23
+#define NETSEC_RX_PKT_ER_FIELD 21
+#define NETSEC_RX_PKT_ERR_FIELD 16
+#define NETSEC_RX_PKT_TDRID_FIELD 12
+#define NETSEC_RX_PKT_FS_FIELD 9
+#define NETSEC_RX_PKT_LS_FIELD 8
+#define NETSEC_RX_PKT_CO_FIELD 6
+
+#define NETSEC_RX_PKT_ERR_MASK 3
+
+#define NETSEC_MAX_TX_PKT_LEN 1518
+#define NETSEC_MAX_TX_JUMBO_PKT_LEN 9018
+
+enum netsec_rings {
+ NETSEC_RING_TX,
+ NETSEC_RING_RX
+};
+
+#define NETSEC_RING_GMAC 15
+#define NETSEC_RING_MAX 1
+
+#define NETSEC_TCP_SEG_LEN_MAX 1460
+#define NETSEC_TCP_JUMBO_SEG_LEN_MAX 8960
+
+#define NETSEC_RX_CKSUM_NOTAVAIL 0
+#define NETSEC_RX_CKSUM_OK 1
+#define NETSEC_RX_CKSUM_NG 2
+
+#define NETSEC_TOP_IRQ_REG_CODE_LOAD_END BIT(20)
+#define NETSEC_IRQ_TRANSITION_COMPLETE BIT(4)
+#define NETSEC_IRQ_RX BIT(1)
+#define NETSEC_IRQ_TX BIT(0)
+
+#define NETSEC_IRQ_EMPTY BIT(17)
+#define NETSEC_IRQ_ERR BIT(16)
+#define NETSEC_IRQ_PKT_CNT BIT(15)
+#define NETSEC_IRQ_TIMEUP BIT(14)
+#define NETSEC_IRQ_RCV (NETSEC_IRQ_PKT_CNT | NETSEC_IRQ_TIMEUP)
+
+#define NETSEC_IRQ_TX_DONE BIT(15)
+#define NETSEC_IRQ_SND (NETSEC_IRQ_TX_DONE | NETSEC_IRQ_TIMEUP)
+
+#define NETSEC_MODE_TRANS_COMP_IRQ_N2T BIT(20)
+#define NETSEC_MODE_TRANS_COMP_IRQ_T2N BIT(19)
+
+#define NETSEC_DESC_MIN 2
+#define NETSEC_DESC_MAX 2047
+#define NETSEC_INT_PKTCNT_MAX 2047
+
+#define NETSEC_FLOW_START_TH_MAX 95
+#define NETSEC_FLOW_STOP_TH_MAX 95
+#define NETSEC_FLOW_PAUSE_TIME_MIN 5
+
+#define NETSEC_CLK_EN_REG_DOM_ALL 0x3f
+
+#define NETSEC_REG_TOP_STATUS 0x80
+#define NETSEC_REG_TOP_INTEN 0x81
+#define NETSEC_REG_INTEN_SET 0x8d
+#define NETSEC_REG_INTEN_CLR 0x8e
+#define NETSEC_REG_NRM_TX_STATUS 0x100
+#define NETSEC_REG_NRM_TX_INTEN 0x101
+#define NETSEC_REG_NRM_TX_INTEN_SET 0x10a
+#define NETSEC_REG_NRM_TX_INTEN_CLR 0x10b
+#define NETSEC_REG_NRM_RX_STATUS 0x110
+#define NETSEC_REG_NRM_RX_INTEN 0x111
+#define NETSEC_REG_NRM_RX_INTEN_SET 0x11a
+#define NETSEC_REG_NRM_RX_INTEN_CLR 0x11b
+#define NETSEC_REG_RESERVED_RX_DESC_START 0x122
+#define NETSEC_REG_RESERVED_TX_DESC_START 0x132
+#define NETSEC_REG_CLK_EN 0x40
+#define NETSEC_REG_SOFT_RST 0x41
+#define NETSEC_REG_PKT_CTRL 0x50
+#define NETSEC_REG_COM_INIT 0x48
+#define NETSEC_REG_DMA_TMR_CTRL 0x83
+#define NETSEC_REG_F_TAIKI_MC_VER 0x8b
+#define NETSEC_REG_F_TAIKI_VER 0x8c
+#define NETSEC_REG_DMA_HM_CTRL 0x85
+#define NETSEC_REG_DMA_MH_CTRL 0x88
+#define NETSEC_REG_NRM_TX_PKTCNT 0x104
+#define NETSEC_REG_NRM_TX_DONE_TXINT_PKTCNT 0x106
+#define NETSEC_REG_NRM_RX_RXINT_PKTCNT 0x116
+#define NETSEC_REG_NRM_TX_TXINT_TMR 0x108
+#define NETSEC_REG_NRM_RX_RXINT_TMR 0x118
+#define NETSEC_REG_NRM_TX_DONE_PKTCNT 0x105
+#define NETSEC_REG_NRM_RX_PKTCNT 0x115
+#define NETSEC_REG_NRM_TX_TMR 0x107
+#define NETSEC_REG_NRM_RX_TMR 0x117
+#define NETSEC_REG_NRM_TX_DESC_START_UP 0x10d
+#define NETSEC_REG_NRM_TX_DESC_START_LW 0x102
+#define NETSEC_REG_NRM_RX_DESC_START_UP 0x11d
+#define NETSEC_REG_NRM_RX_DESC_START_LW 0x112
+#define NETSEC_REG_NRM_TX_CONFIG 0x10c
+#define NETSEC_REG_NRM_RX_CONFIG 0x11c
+#define MAC_REG_DATA 0x470
+#define MAC_REG_CMD 0x471
+#define MAC_REG_FLOW_TH 0x473
+#define MAC_REG_INTF_SEL 0x475
+#define MAC_REG_DESC_INIT 0x47f
+#define MAC_REG_DESC_SOFT_RST 0x481
+#define NETSEC_REG_MODE_TRANS_COMP_STATUS 0x140
+#define GMAC_REG_MCR 0x0000
+#define GMAC_REG_MFFR 0x0004
+#define GMAC_REG_GAR 0x0010
+#define GMAC_REG_GDR 0x0014
+#define GMAC_REG_FCR 0x0018
+#define GMAC_REG_BMR 0x1000
+#define GMAC_REG_RDLAR 0x100c
+#define GMAC_REG_TDLAR 0x1010
+#define GMAC_REG_OMR 0x1018
+
+#define NETSEC_PKT_CTRL_REG_MODE_NRM BIT(28)
+#define NETSEC_PKT_CTRL_REG_EN_JUMBO BIT(27)
+#define NETSEC_PKT_CTRL_REG_LOG_CHKSUM_ER BIT(3)
+#define NETSEC_PKT_CTRL_REG_LOG_HD_INCOMPLETE BIT(2)
+#define NETSEC_PKT_CTRL_REG_LOG_HD_ER BIT(1)
+#define NETSEC_PKT_CTRL_REG_DRP_NO_MATCH BIT(0)
+
+#define NETSEC_CLK_EN_REG_DOM_G BIT(5)
+#define NETSEC_CLK_EN_REG_DOM_C BIT(1)
+#define NETSEC_CLK_EN_REG_DOM_D BIT(0)
+
+#define NETSEC_COM_INIT_REG_PKT BIT(1)
+#define NETSEC_COM_INIT_REG_CORE BIT(0)
+
+#define NETSEC_SOFT_RST_REG_RESET 0
+#define NETSEC_SOFT_RST_REG_RUN BIT(31)
+
+#define NETSEC_DMA_CTRL_REG_STOP 1
+#define MH_CTRL__MODE_TRANS BIT(20)
+
+#define NETSEC_GMAC_CMD_ST_READ 0
+#define NETSEC_GMAC_CMD_ST_WRITE BIT(28)
+#define NETSEC_GMAC_CMD_ST_BUSY BIT(31)
+
+#define NETSEC_GMAC_BMR_REG_COMMON 0x00412080
+#define NETSEC_GMAC_BMR_REG_RESET 0x00020181
+#define NETSEC_GMAC_BMR_REG_SWR 0x00000001
+
+#define NETSEC_GMAC_OMR_REG_ST BIT(13)
+#define NETSEC_GMAC_OMR_REG_SR BIT(1)
+
+#define NETSEC_GMAC_MCR_REG_IBN BIT(30)
+#define NETSEC_GMAC_MCR_REG_CST BIT(25)
+#define NETSEC_GMAC_MCR_REG_JE BIT(20)
+#define NETSEC_MCR_PS BIT(15)
+#define NETSEC_GMAC_MCR_REG_FES BIT(14)
+#define NETSEC_GMAC_MCR_REG_FULL_DUPLEX_COMMON 0x0000280c
+#define NETSEC_GMAC_MCR_REG_HALF_DUPLEX_COMMON 0x0001a00c
+
+#define NETSEC_FCR_RFE BIT(2)
+#define NETSEC_FCR_TFE BIT(1)
+
+#define NETSEC_GMAC_GAR_REG_GW BIT(1)
+#define NETSEC_GMAC_GAR_REG_GB BIT(0)
+
+#define NETSEC_GMAC_GAR_REG_SHIFT_PA 11
+#define NETSEC_GMAC_GAR_REG_SHIFT_GR 6
+#define GMAC_REG_SHIFT_CR_GAR 2
+
+#define NETSEC_GMAC_GAR_REG_CR_25_35_MHZ 2
+#define NETSEC_GMAC_GAR_REG_CR_35_60_MHZ 3
+#define NETSEC_GMAC_GAR_REG_CR_60_100_MHZ 0
+#define NETSEC_GMAC_GAR_REG_CR_100_150_MHZ 1
+#define NETSEC_GMAC_GAR_REG_CR_150_250_MHZ 4
+#define NETSEC_GMAC_GAR_REG_CR_250_300_MHZ 5
+
+#define NETSEC_REG_NETSEC_VER_F_TAIKI 0x50000
+
+#define NETSEC_REG_DESC_RING_CONFIG_CFG_UP BIT(31)
+#define NETSEC_REG_DESC_RING_CONFIG_CH_RST BIT(30)
+#define NETSEC_REG_DESC_TMR_MODE 4
+#define NETSEC_REG_DESC_ENDIAN 0
+
+#define NETSEC_MAC_DESC_SOFT_RST_SOFT_RST 1
+#define NETSEC_MAC_DESC_INIT_REG_INIT 1
+
+/* this is used to interpret a register layout */
+struct netsec_pkt_ctrlaram {
+ u8 log_chksum_er_flag:1;
+ u8 log_hd_imcomplete_flag:1;
+ u8 log_hd_er_flag:1;
+};
+
+struct netsec_param {
+ struct netsec_pkt_ctrlaram pkt_ctrlaram;
+ bool use_jumbo_pkt_flag;
+};
+
+struct netsec_mac_mode {
+ u16 flow_start_th;
+ u16 flow_stop_th;
+ u16 pause_time;
+ bool flow_ctrl_enable_flag;
+};
+
+struct netsec_desc_ring {
+ spinlock_t spinlock_desc; /* protect descriptor access */
+ phys_addr_t desc_phys;
+ struct netsec_frag_info *frag;
+ struct sk_buff **priv;
+ void *ring_vaddr;
+ enum netsec_rings id;
+ int len;
+ u16 tx_done_num;
+ u16 rx_num;
+ u16 head;
+ u16 tail;
+ bool running;
+ bool full;
+};
+
+struct netsec_frag_info {
+ dma_addr_t dma_addr;
+ void *addr;
+ u16 len;
+};
+
+struct netsec_priv {
+ struct netsec_desc_ring desc_ring[NETSEC_RING_MAX + 1];
+ struct ethtool_coalesce et_coalesce;
+ struct netsec_mac_mode mac_mode;
+ struct netsec_param param;
+ struct napi_struct napi;
+ phys_addr_t rdlar_pa, tdlar_pa;
+ phy_interface_t phy_interface;
+ spinlock_t tx_queue_lock; /* protect transmit queue */
+ struct netsec_frag_info tx_info[MAX_SKB_FRAGS];
+ struct net_device *ndev;
+ struct device_node *phy_np;
+ struct mii_bus *mii_bus;
+ void __iomem *ioaddr;
+ struct device *dev;
+ struct clk *clk[3];
+ phys_addr_t scb_set_normal_tx_paddr;
+ u32 scb_pkt_ctrl_reg;
+ u32 rx_pkt_buf_len;
+ u32 msg_enable;
+ u32 freq;
+ int actual_link_speed;
+ int clock_count;
+ bool rx_cksum_offload_flag;
+ bool actual_duplex;
+ bool irq_registered;
+};
+
+struct netsec_tx_de {
+ u32 attr;
+ u32 data_buf_addr_up;
+ u32 data_buf_addr_lw;
+ u32 buf_len_info;
+};
+
+struct netsec_rx_de {
+ u32 attr;
+ u32 data_buf_addr_up;
+ u32 data_buf_addr_lw;
+ u32 buf_len_info;
+};
+
+struct netsec_tx_pkt_ctrl {
+ u16 tcp_seg_len;
+ bool tcp_seg_offload_flag;
+ bool cksum_offload_flag;
+};
+
+struct netsec_rx_pkt_info {
+ int rx_cksum_result;
+ int err_code;
+ bool is_fragmented;
+ bool err_flag;
+};
+
+struct netsec_skb_cb {
+ bool is_rx;
+};
+
+static inline void netsec_writel(struct netsec_priv *priv,
+ u32 reg_addr, u32 val)
+{
+ writel_relaxed(val, priv->ioaddr + (reg_addr << 2));
+}
+
+static inline u32 netsec_readl(struct netsec_priv *priv, u32 reg_addr)
+{
+ return readl_relaxed(priv->ioaddr + (reg_addr << 2));
+}
+
+static inline void netsec_mark_skb_type(struct sk_buff *skb, bool is_rx)
+{
+ struct netsec_skb_cb *cb = (struct netsec_skb_cb *)skb->cb;
+
+ cb->is_rx = is_rx;
+}
+
+static inline bool skb_is_rx(struct sk_buff *skb)
+{
+ struct netsec_skb_cb *cb = (struct netsec_skb_cb *)skb->cb;
+
+ return cb->is_rx;
+}
+
+extern const struct net_device_ops netsec_netdev_ops;
+extern const struct ethtool_ops netsec_ethtool_ops;
+
+int netsec_start_gmac(struct netsec_priv *priv);
+int netsec_stop_gmac(struct netsec_priv *priv);
+int netsec_mii_register(struct netsec_priv *priv);
+void netsec_mii_unregister(struct netsec_priv *priv);
+int netsec_start_desc_ring(struct netsec_priv *priv, enum netsec_rings id);
+void netsec_stop_desc_ring(struct netsec_priv *priv, enum netsec_rings id);
+u16 netsec_get_rx_num(struct netsec_priv *priv);
+u16 netsec_get_tx_avail_num(struct netsec_priv *priv);
+int netsec_clean_tx_desc_ring(struct netsec_priv *priv);
+int netsec_clean_rx_desc_ring(struct netsec_priv *priv);
+int netsec_set_tx_pkt_data(struct netsec_priv *priv,
+ const struct netsec_tx_pkt_ctrl *tx_ctrl,
+ u8 count_frags, const struct netsec_frag_info *info,
+ struct sk_buff *skb);
+int netsec_get_rx_pkt_data(struct netsec_priv *priv,
+ struct netsec_rx_pkt_info *rxpi,
+ struct netsec_frag_info *frag, u16 *len,
+ struct sk_buff **skb);
+void netsec_ring_irq_enable(struct netsec_priv *priv,
+ enum netsec_rings id, u32 i);
+void netsec_ring_irq_disable(struct netsec_priv *priv,
+ enum netsec_rings id, u32 i);
+int netsec_alloc_desc_ring(struct netsec_priv *priv, enum netsec_rings id);
+void netsec_free_desc_ring(struct netsec_priv *priv,
+ struct netsec_desc_ring *desc);
+int netsec_setup_rx_desc(struct netsec_priv *priv,
+ struct netsec_desc_ring *desc);
+int netsec_netdev_napi_poll(struct napi_struct *napi_p, int budget);
+
+#endif /* NETSEC_INTERNAL_H */
new file mode 100644
@@ -0,0 +1,618 @@
+/**
+ * drivers/net/ethernet/socionext/netsec/netsec_desc_ring_access.c
+ *
+ * Copyright (C) 2011-2014 Fujitsu Semiconductor Limited.
+ * Copyright (C) 2014 Linaro Ltd Andy Green <andy.green@linaro.org>
+ * All rights reserved.
+ *
+ * This program is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU General Public License
+ * as published by the Free Software Foundation; either version 2
+ * of the License, or (at your option) any later version.
+ */
+
+#include <linux/spinlock.h>
+#include <linux/dma-mapping.h>
+
+#include "netsec.h"
+
+static const u32 ads_irq_set[] = {
+ NETSEC_REG_NRM_TX_INTEN_SET,
+ NETSEC_REG_NRM_RX_INTEN_SET,
+};
+
+static const u32 desc_ring_irq_inten_clr_reg_addr[] = {
+ NETSEC_REG_NRM_TX_INTEN_CLR,
+ NETSEC_REG_NRM_RX_INTEN_CLR,
+};
+
+static const u32 int_tmr_reg_addr[] = {
+ NETSEC_REG_NRM_TX_TXINT_TMR,
+ NETSEC_REG_NRM_RX_RXINT_TMR,
+};
+
+static const u32 rx_pkt_cnt_reg_addr[] = {
+ 0,
+ NETSEC_REG_NRM_RX_PKTCNT,
+};
+
+static const u32 tx_pkt_cnt_reg_addr[] = {
+ NETSEC_REG_NRM_TX_PKTCNT,
+ 0,
+};
+
+static const u32 int_pkt_cnt_reg_addr[] = {
+ NETSEC_REG_NRM_TX_DONE_TXINT_PKTCNT,
+ NETSEC_REG_NRM_RX_RXINT_PKTCNT,
+};
+
+static const u32 tx_done_pkt_addr[] = {
+ NETSEC_REG_NRM_TX_DONE_PKTCNT,
+ 0,
+};
+
+static const u32 netsec_desc_mask[] = {
+ [NETSEC_RING_TX] = NETSEC_GMAC_OMR_REG_ST,
+ [NETSEC_RING_RX] = NETSEC_GMAC_OMR_REG_SR
+};
+
+void netsec_ring_irq_enable(struct netsec_priv *priv,
+ enum netsec_rings id, u32 irqf)
+{
+ netsec_writel(priv, ads_irq_set[id], irqf);
+}
+
+void netsec_ring_irq_disable(struct netsec_priv *priv,
+ enum netsec_rings id, u32 irqf)
+{
+ netsec_writel(priv, desc_ring_irq_inten_clr_reg_addr[id], irqf);
+}
+
+static struct sk_buff *alloc_rx_pkt_buf(struct netsec_priv *priv,
+ struct netsec_frag_info *info)
+{
+ struct sk_buff *skb;
+
+ skb = netdev_alloc_skb_ip_align(priv->ndev, info->len);
+ if (!skb)
+ return NULL;
+
+ netsec_mark_skb_type(skb, NETSEC_RING_RX);
+ info->addr = skb->data;
+ info->dma_addr = dma_map_single(priv->dev, info->addr, info->len,
+ DMA_FROM_DEVICE);
+ if (dma_mapping_error(priv->dev, info->dma_addr)) {
+ dev_kfree_skb(skb);
+ return NULL;
+ }
+
+ return skb;
+}
+
+int netsec_alloc_desc_ring(struct netsec_priv *priv, enum netsec_rings id)
+{
+ struct netsec_desc_ring *desc = &priv->desc_ring[id];
+ int ret = 0;
+
+ desc->id = id;
+ desc->len = sizeof(struct netsec_tx_de); /* rx and tx desc same size */
+
+ spin_lock_init(&desc->spinlock_desc);
+
+ desc->ring_vaddr = dma_zalloc_coherent(priv->dev, desc->len * DESC_NUM,
+ &desc->desc_phys, GFP_KERNEL);
+ if (!desc->ring_vaddr) {
+ ret = -ENOMEM;
+ goto err;
+ }
+
+ desc->frag = kcalloc(DESC_NUM, sizeof(*desc->frag), GFP_KERNEL);
+ if (!desc->frag) {
+ ret = -ENOMEM;
+ goto err;
+ }
+
+ desc->priv = kcalloc(DESC_NUM, sizeof(struct sk_buff *), GFP_KERNEL);
+ if (!desc->priv) {
+ ret = -ENOMEM;
+ goto err;
+ }
+
+ return 0;
+
+err:
+ netsec_free_desc_ring(priv, desc);
+
+ return ret;
+}
+
+static void netsec_uninit_pkt_desc_ring(struct netsec_priv *priv,
+ struct netsec_desc_ring *desc)
+{
+ struct netsec_frag_info *frag;
+ u32 status;
+ u16 idx;
+
+ for (idx = 0; idx < DESC_NUM; idx++) {
+ frag = &desc->frag[idx];
+ if (!frag->addr)
+ continue;
+
+ status = *(u32 *)(desc->ring_vaddr + desc->len * idx);
+
+ dma_unmap_single(priv->dev, frag->dma_addr, frag->len,
+ skb_is_rx(desc->priv[idx]) ? DMA_FROM_DEVICE :
+ DMA_TO_DEVICE);
+ if ((status >> NETSEC_TX_LAST) & 1)
+ dev_kfree_skb(desc->priv[idx]);
+ }
+
+ memset(desc->frag, 0, sizeof(struct netsec_frag_info) * DESC_NUM);
+ memset(desc->priv, 0, sizeof(struct sk_buff *) * DESC_NUM);
+ memset(desc->ring_vaddr, 0, desc->len * DESC_NUM);
+}
+
+void netsec_free_desc_ring(struct netsec_priv *priv,
+ struct netsec_desc_ring *desc)
+{
+ if (desc->ring_vaddr && desc->frag && desc->priv)
+ netsec_uninit_pkt_desc_ring(priv, desc);
+
+ if (desc->ring_vaddr) {
+ dma_free_coherent(priv->dev, desc->len * DESC_NUM,
+ desc->ring_vaddr, desc->desc_phys);
+ desc->ring_vaddr = NULL;
+ }
+ kfree(desc->frag);
+ desc->frag = NULL;
+ kfree(desc->priv);
+ desc->priv = NULL;
+}
+
+static void netsec_set_rx_de(struct netsec_priv *priv,
+ struct netsec_desc_ring *desc, u16 idx,
+ const struct netsec_frag_info *info,
+ struct sk_buff *skb)
+{
+ struct netsec_rx_de *de = desc->ring_vaddr + desc->len * idx;
+ u32 attr = 1 << NETSEC_RX_PKT_OWN_FIELD | 1 << NETSEC_RX_PKT_FS_FIELD |
+ 1 << NETSEC_RX_PKT_LS_FIELD;
+
+ if (idx == DESC_NUM - 1)
+ attr |= 1 << NETSEC_RX_PKT_LD_FIELD;
+
+ de->data_buf_addr_up = info->dma_addr >> 32;
+ de->data_buf_addr_lw = info->dma_addr & 0xffffffff;
+ de->buf_len_info = info->len;
+ /* desc->attr makes the descriptor live, so it must be physically
+ * written last after the rest of the descriptor body is already there
+ */
+ wmb();
+ de->attr = attr;
+
+ desc->frag[idx].dma_addr = info->dma_addr;
+ desc->frag[idx].addr = info->addr;
+ desc->frag[idx].len = info->len;
+
+ desc->priv[idx] = skb;
+}
+
+int netsec_setup_rx_desc(struct netsec_priv *priv,
+ struct netsec_desc_ring *desc)
+{
+ struct netsec_frag_info info;
+ struct sk_buff *skb;
+ int n;
+
+ info.len = priv->rx_pkt_buf_len;
+
+ for (n = 0; n < DESC_NUM; n++) {
+ skb = alloc_rx_pkt_buf(priv, &info);
+ if (!skb) {
+ netsec_uninit_pkt_desc_ring(priv, desc);
+ return -ENOMEM;
+ }
+ netsec_set_rx_de(priv, desc, n, &info, skb);
+ }
+
+ return 0;
+}
+
+static void netsec_set_tx_desc_entry(struct netsec_priv *priv,
+ struct netsec_desc_ring *desc,
+ const struct netsec_tx_pkt_ctrl *tx_ctrl,
+ bool first_flag, bool last_flag,
+ const struct netsec_frag_info *frag,
+ struct sk_buff *skb)
+{
+ struct netsec_tx_de tx_desc_entry;
+ int idx = desc->head;
+
+ memset(&tx_desc_entry, 0, sizeof(struct netsec_tx_de));
+
+ tx_desc_entry.attr = (1 << NETSEC_TX_SHIFT_OWN_FIELD) |
+ (desc->id << NETSEC_TX_SHIFT_DRID_FIELD) |
+ (1 << NETSEC_TX_SHIFT_PT_FIELD) |
+ (NETSEC_RING_GMAC << NETSEC_TX_SHIFT_TDRID_FIELD) |
+ (first_flag << NETSEC_TX_SHIFT_FS_FIELD) |
+ (last_flag << NETSEC_TX_LAST) |
+ (tx_ctrl->cksum_offload_flag <<
+ NETSEC_TX_SHIFT_CO) |
+ (tx_ctrl->tcp_seg_offload_flag <<
+ NETSEC_TX_SHIFT_SO) |
+ (1 << NETSEC_TX_SHIFT_TRS_FIELD);
+ if (idx == DESC_NUM - 1)
+ tx_desc_entry.attr |= (1 << NETSEC_TX_SHIFT_LD_FIELD);
+
+ tx_desc_entry.data_buf_addr_up = frag->dma_addr >> 32;
+ tx_desc_entry.data_buf_addr_lw = frag->dma_addr & 0xffffffff;
+ tx_desc_entry.buf_len_info = (tx_ctrl->tcp_seg_len << 16) | frag->len;
+
+ memcpy(desc->ring_vaddr + (desc->len * idx), &tx_desc_entry, desc->len);
+
+ desc->frag[idx].dma_addr = frag->dma_addr;
+ desc->frag[idx].addr = frag->addr;
+ desc->frag[idx].len = frag->len;
+
+ desc->priv[idx] = skb;
+}
+
+static void netsec_get_rx_de(struct netsec_priv *priv,
+ struct netsec_desc_ring *desc, u16 idx,
+ struct netsec_rx_pkt_info *rxpi,
+ struct netsec_frag_info *frag, u16 *len,
+ struct sk_buff **skb)
+{
+ struct netsec_rx_de de;
+
+ memset(&de, 0, sizeof(struct netsec_rx_de));
+ memset(rxpi, 0, sizeof(struct netsec_rx_pkt_info));
+ memcpy(&de, ((void *)desc->ring_vaddr + desc->len * idx), desc->len);
+
+ dev_dbg(priv->dev, "%08x\n", *(u32 *)&de);
+ *len = de.buf_len_info >> 16;
+
+ rxpi->is_fragmented = (de.attr >> NETSEC_RX_PKT_FR_FIELD) & 1;
+ rxpi->err_flag = (de.attr >> NETSEC_RX_PKT_ER_FIELD) & 1;
+ rxpi->rx_cksum_result = (de.attr >> NETSEC_RX_PKT_CO_FIELD) & 3;
+ rxpi->err_code = (de.attr >> NETSEC_RX_PKT_ERR_FIELD) &
+ NETSEC_RX_PKT_ERR_MASK;
+ memcpy(frag, &desc->frag[idx], sizeof(*frag));
+ *skb = desc->priv[idx];
+}
+
+static void netsec_inc_desc_head_idx(struct netsec_priv *priv,
+ struct netsec_desc_ring *desc, u16 inc)
+{
+ u32 sum;
+
+ sum = desc->head + inc;
+
+ if (sum >= DESC_NUM)
+ sum -= DESC_NUM;
+
+ desc->head = sum;
+ desc->full = desc->head == desc->tail;
+}
+
+static void netsec_inc_desc_tail_idx(struct netsec_priv *priv,
+ struct netsec_desc_ring *desc)
+{
+ u32 sum;
+
+ sum = desc->tail + 1;
+
+ if (sum >= DESC_NUM)
+ sum -= DESC_NUM;
+
+ desc->tail = sum;
+ desc->full = false;
+}
+
+static u16 netsec_get_tx_avail_num_sub(struct netsec_priv *priv,
+ const struct netsec_desc_ring *desc)
+{
+ if (desc->full)
+ return 0;
+
+ if (desc->tail > desc->head)
+ return desc->tail - desc->head;
+
+ return DESC_NUM + desc->tail - desc->head;
+}
+
+static u16 netsec_get_tx_done_num_sub(struct netsec_priv *priv,
+ struct netsec_desc_ring *desc)
+{
+ desc->tx_done_num += netsec_readl(priv, tx_done_pkt_addr[desc->id]);
+
+ return desc->tx_done_num;
+}
+
+static int netsec_set_irq_coalesce_param(struct netsec_priv *priv,
+ enum netsec_rings id)
+{
+ int max_frames, tmr;
+
+ switch (id) {
+ case NETSEC_RING_TX:
+ max_frames = priv->et_coalesce.tx_max_coalesced_frames;
+ tmr = priv->et_coalesce.tx_coalesce_usecs;
+ break;
+ case NETSEC_RING_RX:
+ max_frames = priv->et_coalesce.rx_max_coalesced_frames;
+ tmr = priv->et_coalesce.rx_coalesce_usecs;
+ break;
+ default:
+ return -EINVAL;
+ }
+
+ netsec_writel(priv, int_pkt_cnt_reg_addr[id], max_frames);
+ netsec_writel(priv, int_tmr_reg_addr[id], ((tmr != 0) << 31) | tmr);
+
+ return 0;
+}
+
+int netsec_start_desc_ring(struct netsec_priv *priv, enum netsec_rings id)
+{
+ struct netsec_desc_ring *desc = &priv->desc_ring[id];
+ int ret = 0;
+
+ spin_lock_bh(&desc->spinlock_desc);
+
+ if (desc->running) {
+ ret = -EBUSY;
+ goto err;
+ }
+
+ switch (desc->id) {
+ case NETSEC_RING_RX:
+ netsec_writel(priv, ads_irq_set[id], NETSEC_IRQ_RCV);
+ break;
+ case NETSEC_RING_TX:
+ netsec_writel(priv, ads_irq_set[id], NETSEC_IRQ_EMPTY);
+ break;
+ }
+
+ netsec_set_irq_coalesce_param(priv, desc->id);
+ desc->running = true;
+
+err:
+ spin_unlock_bh(&desc->spinlock_desc);
+
+ return ret;
+}
+
+void netsec_stop_desc_ring(struct netsec_priv *priv, enum netsec_rings id)
+{
+ struct netsec_desc_ring *desc = &priv->desc_ring[id];
+
+ spin_lock_bh(&desc->spinlock_desc);
+ if (desc->running)
+ netsec_writel(priv, desc_ring_irq_inten_clr_reg_addr[id],
+ NETSEC_IRQ_RCV | NETSEC_IRQ_EMPTY |
+ NETSEC_IRQ_SND);
+
+ desc->running = false;
+ spin_unlock_bh(&desc->spinlock_desc);
+}
+
+u16 netsec_get_rx_num(struct netsec_priv *priv)
+{
+ struct netsec_desc_ring *desc = &priv->desc_ring[NETSEC_RING_RX];
+ u32 result;
+
+ spin_lock(&desc->spinlock_desc);
+ if (desc->running) {
+ result = netsec_readl(priv,
+ rx_pkt_cnt_reg_addr[NETSEC_RING_RX]);
+ desc->rx_num += result;
+ if (result)
+ netsec_inc_desc_head_idx(priv, desc, result);
+ }
+ spin_unlock(&desc->spinlock_desc);
+
+ return desc->rx_num;
+}
+
+u16 netsec_get_tx_avail_num(struct netsec_priv *priv)
+{
+ struct netsec_desc_ring *desc = &priv->desc_ring[NETSEC_RING_TX];
+ u16 result;
+
+ spin_lock(&desc->spinlock_desc);
+
+ if (!desc->running) {
+ netif_err(priv, drv, priv->ndev,
+ "%s: not running tx desc\n", __func__);
+ result = 0;
+ goto err;
+ }
+
+ result = netsec_get_tx_avail_num_sub(priv, desc);
+
+err:
+ spin_unlock(&desc->spinlock_desc);
+
+ return result;
+}
+
+int netsec_clean_tx_desc_ring(struct netsec_priv *priv)
+{
+ struct netsec_desc_ring *desc = &priv->desc_ring[NETSEC_RING_TX];
+ unsigned int pkts = 0, bytes = 0;
+ struct netsec_frag_info *frag;
+ struct netsec_tx_de *entry;
+ bool is_last;
+
+ spin_lock(&desc->spinlock_desc);
+
+ netsec_get_tx_done_num_sub(priv, desc);
+
+ while ((desc->tail != desc->head || desc->full) && desc->tx_done_num) {
+ frag = &desc->frag[desc->tail];
+ entry = desc->ring_vaddr + desc->len * desc->tail;
+ is_last = (entry->attr >> NETSEC_TX_LAST) & 1;
+
+ dma_unmap_single(priv->dev, frag->dma_addr, frag->len,
+ DMA_TO_DEVICE);
+ if (is_last) {
+ pkts++;
+ bytes += desc->priv[desc->tail]->len;
+ dev_kfree_skb(desc->priv[desc->tail]);
+ }
+ memset(frag, 0, sizeof(*frag));
+ netsec_inc_desc_tail_idx(priv, desc);
+
+ if (is_last)
+ desc->tx_done_num--;
+ }
+
+ spin_unlock(&desc->spinlock_desc);
+
+ priv->ndev->stats.tx_packets += pkts;
+ priv->ndev->stats.tx_bytes += bytes;
+
+ netdev_completed_queue(priv->ndev, pkts, bytes);
+
+ return 0;
+}
+
+int netsec_clean_rx_desc_ring(struct netsec_priv *priv)
+{
+ struct netsec_desc_ring *desc = &priv->desc_ring[NETSEC_RING_RX];
+
+ spin_lock(&desc->spinlock_desc);
+
+ while (desc->full || (desc->tail != desc->head)) {
+ netsec_set_rx_de(priv, desc, desc->tail,
+ &desc->frag[desc->tail],
+ desc->priv[desc->tail]);
+ desc->rx_num--;
+ netsec_inc_desc_tail_idx(priv, desc);
+ }
+
+ spin_unlock(&desc->spinlock_desc);
+
+ return 0;
+}
+
+int netsec_set_tx_pkt_data(struct netsec_priv *priv,
+ const struct netsec_tx_pkt_ctrl *tx_ctrl,
+ u8 count_frags, const struct netsec_frag_info *info,
+ struct sk_buff *skb)
+{
+ struct netsec_desc_ring *desc;
+ u32 sum_len = 0;
+ unsigned int i;
+ int ret = 0;
+
+ if (tx_ctrl->tcp_seg_offload_flag && !tx_ctrl->cksum_offload_flag)
+ return -EINVAL;
+
+ if (tx_ctrl->tcp_seg_offload_flag) {
+ if (tx_ctrl->tcp_seg_len == 0)
+ return -EINVAL;
+
+ if (priv->param.use_jumbo_pkt_flag) {
+ if (tx_ctrl->tcp_seg_len > NETSEC_TCP_JUMBO_SEG_LEN_MAX)
+ return -EINVAL;
+ } else {
+ if (tx_ctrl->tcp_seg_len > NETSEC_TCP_SEG_LEN_MAX)
+ return -EINVAL;
+ }
+ } else {
+ if (tx_ctrl->tcp_seg_len)
+ return -EINVAL;
+ }
+
+ if (!count_frags)
+ return -ERANGE;
+
+ for (i = 0; i < count_frags; i++) {
+ if ((info[i].len == 0) || (info[i].len > 0xffff)) {
+ netif_err(priv, drv, priv->ndev,
+ "%s: bad info len\n", __func__);
+ return -EINVAL;
+ }
+ sum_len += info[i].len;
+ }
+
+ if (!tx_ctrl->tcp_seg_offload_flag) {
+ if (priv->param.use_jumbo_pkt_flag) {
+ if (sum_len > NETSEC_MAX_TX_JUMBO_PKT_LEN)
+ return -EINVAL;
+ } else {
+ if (sum_len > NETSEC_MAX_TX_PKT_LEN)
+ return -EINVAL;
+ }
+ }
+
+ desc = &priv->desc_ring[NETSEC_RING_TX];
+ spin_lock(&desc->spinlock_desc);
+
+ if (!desc->running) {
+ ret = -ENODEV;
+ goto end;
+ }
+
+ smp_rmb(); /* we need to see a consistent view of pending tx count */
+ if (count_frags > netsec_get_tx_avail_num_sub(priv, desc)) {
+ ret = -EBUSY;
+ goto end;
+ }
+
+ for (i = 0; i < count_frags; i++) {
+ netsec_set_tx_desc_entry(priv, desc, tx_ctrl, i == 0,
+ i == count_frags - 1, &info[i], skb);
+ netsec_inc_desc_head_idx(priv, desc, 1);
+ }
+
+ wmb(); /* ensure the descriptor is flushed */
+ netsec_writel(priv, tx_pkt_cnt_reg_addr[NETSEC_RING_TX], 1);
+
+end:
+ spin_unlock(&desc->spinlock_desc);
+
+ return ret;
+}
+
+int netsec_get_rx_pkt_data(struct netsec_priv *priv,
+ struct netsec_rx_pkt_info *rxpi,
+ struct netsec_frag_info *frag, u16 *len,
+ struct sk_buff **skb)
+{
+ struct netsec_desc_ring *desc = &priv->desc_ring[NETSEC_RING_RX];
+ struct netsec_frag_info info;
+ struct sk_buff *tmp_skb;
+ int ret = 0;
+
+ spin_lock(&desc->spinlock_desc);
+
+ if (desc->rx_num == 0) {
+ dev_err(priv->dev, "%s 0 len rx\n", __func__);
+ ret = -EINVAL;
+ goto err;
+ }
+
+ info.len = priv->rx_pkt_buf_len;
+ rmb(); /* we need to ensure we only see current data in descriptor */
+ tmp_skb = alloc_rx_pkt_buf(priv, &info);
+ if (!tmp_skb) {
+ netsec_set_rx_de(priv, desc, desc->tail,
+ &desc->frag[desc->tail],
+ desc->priv[desc->tail]);
+ ret = -ENOMEM;
+ } else {
+ netsec_get_rx_de(priv, desc, desc->tail, rxpi, frag, len, skb);
+ netsec_set_rx_de(priv, desc, desc->tail, &info, tmp_skb);
+ }
+
+ netsec_inc_desc_tail_idx(priv, desc);
+ desc->rx_num--;
+
+err:
+ spin_unlock(&desc->spinlock_desc);
+
+ return ret;
+}
new file mode 100644
@@ -0,0 +1,76 @@
+/**
+ * drivers/net/ethernet/socionext/netsec/netsec_ethtool.c
+ *
+ * Copyright (C) 2013-2014 Fujitsu Semiconductor Limited.
+ * Copyright (C) 2014 Linaro Ltd Andy Green <andy.green@linaro.org>
+ * All rights reserved.
+ *
+ * This program is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU General Public License
+ * as published by the Free Software Foundation; either version 2
+ * of the License, or (at your option) any later version.
+ */
+
+#include "netsec.h"
+
+static void netsec_et_get_drvinfo(struct net_device *net_device,
+ struct ethtool_drvinfo *info)
+{
+ strlcpy(info->driver, "netsec", sizeof(info->driver));
+ strlcpy(info->bus_info, dev_name(net_device->dev.parent),
+ sizeof(info->bus_info));
+}
+
+static int netsec_et_get_coalesce(struct net_device *net_device,
+ struct ethtool_coalesce *et_coalesce)
+{
+ struct netsec_priv *priv = netdev_priv(net_device);
+
+ *et_coalesce = priv->et_coalesce;
+
+ return 0;
+}
+
+static int netsec_et_set_coalesce(struct net_device *net_device,
+ struct ethtool_coalesce *et_coalesce)
+{
+ struct netsec_priv *priv = netdev_priv(net_device);
+
+ if (et_coalesce->rx_max_coalesced_frames > NETSEC_INT_PKTCNT_MAX)
+ return -EINVAL;
+ if (et_coalesce->tx_max_coalesced_frames > NETSEC_INT_PKTCNT_MAX)
+ return -EINVAL;
+ if (!et_coalesce->rx_max_coalesced_frames)
+ return -EINVAL;
+ if (!et_coalesce->tx_max_coalesced_frames)
+ return -EINVAL;
+
+ priv->et_coalesce = *et_coalesce;
+
+ return 0;
+}
+
+static u32 netsec_et_get_msglevel(struct net_device *dev)
+{
+ struct netsec_priv *priv = netdev_priv(dev);
+
+ return priv->msg_enable;
+}
+
+static void netsec_et_set_msglevel(struct net_device *dev, u32 datum)
+{
+ struct netsec_priv *priv = netdev_priv(dev);
+
+ priv->msg_enable = datum;
+}
+
+const struct ethtool_ops netsec_ethtool_ops = {
+ .get_drvinfo = netsec_et_get_drvinfo,
+ .get_link_ksettings = phy_ethtool_get_link_ksettings,
+ .set_link_ksettings = phy_ethtool_set_link_ksettings,
+ .get_link = ethtool_op_get_link,
+ .get_coalesce = netsec_et_get_coalesce,
+ .set_coalesce = netsec_et_set_coalesce,
+ .get_msglevel = netsec_et_get_msglevel,
+ .set_msglevel = netsec_et_set_msglevel,
+};
new file mode 100644
@@ -0,0 +1,329 @@
+/**
+ * drivers/net/ethernet/socionext/netsec/netsec_gmac_access.c
+ *
+ * Copyright (C) 2011-2014 Fujitsu Semiconductor Limited.
+ * Copyright (C) 2014 Linaro Ltd Andy Green <andy.green@linaro.org>
+ * All rights reserved.
+ *
+ * This program is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU General Public License
+ * as published by the Free Software Foundation; either version 2
+ * of the License, or (at your option) any later version.
+ */
+#include "netsec.h"
+
+#define TIMEOUT_SPINS_MAC 1000
+#define TIMEOUT_SECONDARY_MS_MAC 100
+
+static u32 netsec_clk_type(u32 freq)
+{
+ if (freq < 35 * NETSEC_CLK_MHZ)
+ return NETSEC_GMAC_GAR_REG_CR_25_35_MHZ;
+ if (freq < 60 * NETSEC_CLK_MHZ)
+ return NETSEC_GMAC_GAR_REG_CR_35_60_MHZ;
+ if (freq < 100 * NETSEC_CLK_MHZ)
+ return NETSEC_GMAC_GAR_REG_CR_60_100_MHZ;
+ if (freq < 150 * NETSEC_CLK_MHZ)
+ return NETSEC_GMAC_GAR_REG_CR_100_150_MHZ;
+ if (freq < 250 * NETSEC_CLK_MHZ)
+ return NETSEC_GMAC_GAR_REG_CR_150_250_MHZ;
+
+ return NETSEC_GMAC_GAR_REG_CR_250_300_MHZ;
+}
+
+static int netsec_wait_while_busy(struct netsec_priv *priv, u32 addr, u32 mask)
+{
+ u32 timeout = TIMEOUT_SPINS_MAC;
+
+ while (--timeout && netsec_readl(priv, addr) & mask)
+ ;
+ if (timeout)
+ return 0;
+
+ timeout = TIMEOUT_SECONDARY_MS_MAC;
+ while (--timeout && netsec_readl(priv, addr) & mask)
+ usleep_range(1000, 2000);
+
+ if (timeout)
+ return 0;
+
+ netdev_WARN(priv->ndev, "%s: timeout\n", __func__);
+
+ return -ETIMEDOUT;
+}
+
+static int netsec_mac_write(struct netsec_priv *priv, u32 addr, u32 value)
+{
+ netsec_writel(priv, MAC_REG_DATA, value);
+ netsec_writel(priv, MAC_REG_CMD, addr | NETSEC_GMAC_CMD_ST_WRITE);
+ return netsec_wait_while_busy(priv,
+ MAC_REG_CMD, NETSEC_GMAC_CMD_ST_BUSY);
+}
+
+static int netsec_mac_read(struct netsec_priv *priv, u32 addr, u32 *read)
+{
+ int ret;
+
+ netsec_writel(priv, MAC_REG_CMD, addr | NETSEC_GMAC_CMD_ST_READ);
+ ret = netsec_wait_while_busy(priv,
+ MAC_REG_CMD, NETSEC_GMAC_CMD_ST_BUSY);
+ if (ret)
+ return ret;
+
+ *read = netsec_readl(priv, MAC_REG_DATA);
+
+ return 0;
+}
+
+static int netsec_mac_wait_while_busy(struct netsec_priv *priv,
+ u32 addr, u32 mask)
+{
+ u32 timeout = TIMEOUT_SPINS_MAC;
+ int ret, data;
+
+ do {
+ ret = netsec_mac_read(priv, addr, &data);
+ if (ret)
+ break;
+ } while (--timeout && (data & mask));
+
+ if (timeout)
+ return 0;
+
+ timeout = TIMEOUT_SECONDARY_MS_MAC;
+ do {
+ usleep_range(1000, 2000);
+
+ ret = netsec_mac_read(priv, addr, &data);
+ if (ret)
+ break;
+ } while (--timeout && (data & mask));
+
+ if (timeout && !ret)
+ return 0;
+
+ netdev_WARN(priv->ndev, "%s: timeout\n", __func__);
+
+ return -ETIMEDOUT;
+}
+
+static int netsec_mac_update_to_phy_state(struct netsec_priv *priv)
+{
+ struct phy_device *phydev = priv->ndev->phydev;
+ u32 value = 0;
+
+ value = phydev->duplex ? NETSEC_GMAC_MCR_REG_FULL_DUPLEX_COMMON :
+ NETSEC_GMAC_MCR_REG_HALF_DUPLEX_COMMON;
+
+ if (phydev->speed != SPEED_1000)
+ value |= NETSEC_MCR_PS;
+
+ if ((priv->phy_interface != PHY_INTERFACE_MODE_GMII) &&
+ (phydev->speed == SPEED_100))
+ value |= NETSEC_GMAC_MCR_REG_FES;
+
+ value |= NETSEC_GMAC_MCR_REG_CST | NETSEC_GMAC_MCR_REG_JE;
+
+ if (priv->phy_interface == PHY_INTERFACE_MODE_RGMII)
+ value |= NETSEC_GMAC_MCR_REG_IBN;
+
+ if (netsec_mac_write(priv, GMAC_REG_MCR, value))
+ return -ETIMEDOUT;
+
+ priv->actual_link_speed = phydev->speed;
+ priv->actual_duplex = phydev->duplex;
+ netif_info(priv, drv, priv->ndev, "%s: %uMbps, duplex:%d\n",
+ __func__, phydev->speed, phydev->duplex);
+
+ return 0;
+}
+
+/* NB netsec_start_gmac() only called from adjust_link */
+
+int netsec_start_gmac(struct netsec_priv *priv)
+{
+ struct phy_device *phydev = priv->ndev->phydev;
+ u32 value = 0;
+ int ret;
+
+ if (priv->desc_ring[NETSEC_RING_TX].running &&
+ priv->desc_ring[NETSEC_RING_RX].running)
+ return 0;
+
+ if (!priv->desc_ring[NETSEC_RING_RX].running &&
+ !priv->desc_ring[NETSEC_RING_TX].running) {
+ if (phydev->speed != SPEED_1000)
+ value = (NETSEC_GMAC_MCR_REG_CST |
+ NETSEC_GMAC_MCR_REG_HALF_DUPLEX_COMMON);
+
+ if (netsec_mac_write(priv, GMAC_REG_MCR, value))
+ return -ETIMEDOUT;
+ if (netsec_mac_write(priv, GMAC_REG_BMR,
+ NETSEC_GMAC_BMR_REG_RESET))
+ return -ETIMEDOUT;
+
+ /* Wait soft reset */
+ usleep_range(1000, 5000);
+
+ ret = netsec_mac_read(priv, GMAC_REG_BMR, &value);
+ if (ret)
+ return ret;
+ if (value & NETSEC_GMAC_BMR_REG_SWR)
+ return -EAGAIN;
+
+ netsec_writel(priv, MAC_REG_DESC_SOFT_RST, 1);
+ if (netsec_wait_while_busy(priv, MAC_REG_DESC_SOFT_RST, 1))
+ return -ETIMEDOUT;
+
+ netsec_writel(priv, MAC_REG_DESC_INIT, 1);
+ if (netsec_wait_while_busy(priv, MAC_REG_DESC_INIT, 1))
+ return -ETIMEDOUT;
+
+ if (netsec_mac_write(priv, GMAC_REG_BMR,
+ NETSEC_GMAC_BMR_REG_COMMON))
+ return -ETIMEDOUT;
+ if (netsec_mac_write(priv, GMAC_REG_RDLAR, priv->rdlar_pa))
+ return -ETIMEDOUT;
+ if (netsec_mac_write(priv, GMAC_REG_TDLAR, priv->tdlar_pa))
+ return -ETIMEDOUT;
+ if (netsec_mac_write(priv, GMAC_REG_MFFR, 0x80000001))
+ return -ETIMEDOUT;
+
+ ret = netsec_mac_update_to_phy_state(priv);
+ if (ret)
+ return ret;
+
+ if (priv->mac_mode.flow_ctrl_enable_flag) {
+ netsec_writel(priv, MAC_REG_FLOW_TH,
+ (priv->mac_mode.flow_stop_th << 16) |
+ priv->mac_mode.flow_start_th);
+ if (netsec_mac_write(priv, GMAC_REG_FCR,
+ (priv->mac_mode.pause_time << 16) |
+ NETSEC_FCR_RFE | NETSEC_FCR_TFE))
+ return -ETIMEDOUT;
+ }
+ }
+
+ ret = netsec_mac_read(priv, GMAC_REG_OMR, &value);
+ if (ret)
+ return ret;
+
+ if (!priv->desc_ring[NETSEC_RING_RX].running) {
+ value |= NETSEC_GMAC_OMR_REG_SR;
+ netsec_start_desc_ring(priv, NETSEC_RING_RX);
+ }
+ if (!priv->desc_ring[NETSEC_RING_TX].running) {
+ value |= NETSEC_GMAC_OMR_REG_ST;
+ netsec_start_desc_ring(priv, NETSEC_RING_TX);
+ }
+
+ if (netsec_mac_write(priv, GMAC_REG_OMR, value))
+ return -ETIMEDOUT;
+
+ netsec_writel(priv, NETSEC_REG_INTEN_SET,
+ NETSEC_IRQ_TX | NETSEC_IRQ_RX);
+
+ return 0;
+}
+
+int netsec_stop_gmac(struct netsec_priv *priv)
+{
+ u32 value;
+ int ret;
+
+ ret = netsec_mac_read(priv, GMAC_REG_OMR, &value);
+ if (ret)
+ return ret;
+
+ if (priv->desc_ring[NETSEC_RING_RX].running) {
+ value &= ~NETSEC_GMAC_OMR_REG_SR;
+ netsec_stop_desc_ring(priv, NETSEC_RING_RX);
+ }
+ if (priv->desc_ring[NETSEC_RING_TX].running) {
+ value &= ~NETSEC_GMAC_OMR_REG_ST;
+ netsec_stop_desc_ring(priv, NETSEC_RING_TX);
+ }
+
+ priv->actual_link_speed = 0;
+ priv->actual_duplex = false;
+
+ return netsec_mac_write(priv, GMAC_REG_OMR, value);
+}
+
+static int netsec_phy_write(struct mii_bus *bus,
+ int phy_addr, int reg, u16 val)
+{
+ struct netsec_priv *priv = bus->priv;
+
+ if (netsec_mac_write(priv, GMAC_REG_GDR, val))
+ return -ETIMEDOUT;
+ if (netsec_mac_write(priv, GMAC_REG_GAR,
+ phy_addr << NETSEC_GMAC_GAR_REG_SHIFT_PA |
+ reg << NETSEC_GMAC_GAR_REG_SHIFT_GR |
+ NETSEC_GMAC_GAR_REG_GW | NETSEC_GMAC_GAR_REG_GB) |
+ (netsec_clk_type(priv->freq) <<
+ GMAC_REG_SHIFT_CR_GAR))
+ return -ETIMEDOUT;
+
+ return netsec_mac_wait_while_busy(priv, GMAC_REG_GAR,
+ NETSEC_GMAC_GAR_REG_GB);
+}
+
+static int netsec_phy_read(struct mii_bus *bus, int phy_addr, int reg_addr)
+{
+ struct netsec_priv *priv = bus->priv;
+ u32 data;
+ int ret;
+
+ if (netsec_mac_write(priv, GMAC_REG_GAR, NETSEC_GMAC_GAR_REG_GB |
+ phy_addr << NETSEC_GMAC_GAR_REG_SHIFT_PA |
+ reg_addr << NETSEC_GMAC_GAR_REG_SHIFT_GR |
+ (netsec_clk_type(priv->freq) <<
+ GMAC_REG_SHIFT_CR_GAR)))
+ return -ETIMEDOUT;
+
+ ret = netsec_mac_wait_while_busy(priv, GMAC_REG_GAR,
+ NETSEC_GMAC_GAR_REG_GB);
+ if (ret)
+ return 0;
+
+ ret = netsec_mac_read(priv, GMAC_REG_GDR, &data);
+ if (ret)
+ return ret;
+
+ return data;
+}
+
+int netsec_mii_register(struct netsec_priv *priv)
+{
+ struct mii_bus *bus = mdiobus_alloc();
+ struct resource res;
+ int ret;
+
+ if (!bus)
+ return -ENOMEM;
+
+ of_address_to_resource(priv->dev->of_node, 0, &res);
+ snprintf(bus->id, MII_BUS_ID_SIZE, "%s", priv->dev->of_node->full_name);
+ bus->priv = priv;
+ bus->name = "SNI NETSEC MDIO";
+ bus->read = netsec_phy_read;
+ bus->write = netsec_phy_write;
+ bus->parent = priv->dev;
+ priv->mii_bus = bus;
+
+ ret = of_mdiobus_register(bus, priv->dev->of_node);
+ if (ret) {
+ mdiobus_free(bus);
+ return ret;
+ }
+
+ return 0;
+}
+
+void netsec_mii_unregister(struct netsec_priv *priv)
+{
+ mdiobus_unregister(priv->mii_bus);
+ mdiobus_free(priv->mii_bus);
+ priv->mii_bus = NULL;
+}
new file mode 100644
@@ -0,0 +1,558 @@
+/**
+ * drivers/net/ethernet/socionext/netsec/netsec_netdev.c
+ *
+ * Copyright (C) 2013-2014 Fujitsu Semiconductor Limited.
+ * Copyright (C) 2014 Linaro Ltd Andy Green <andy.green@linaro.org>
+ * All rights reserved.
+ *
+ * This program is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU General Public License
+ * as published by the Free Software Foundation; either version 2
+ * of the License, or (at your option) any later version.
+ */
+
+#include <linux/ip.h>
+#include <linux/ipv6.h>
+#include <linux/tcp.h>
+#include <net/tcp.h>
+#include <net/ip6_checksum.h>
+#include <linux/pm_runtime.h>
+
+#include "netsec.h"
+
+#define WAIT_FW_RDY_TIMEOUT 50
+
+static const u32 desc_ring_irq_status_reg_addr[] = {
+ NETSEC_REG_NRM_TX_STATUS,
+ NETSEC_REG_NRM_RX_STATUS,
+};
+
+static const u32 desc_ads[] = {
+ NETSEC_REG_NRM_TX_CONFIG,
+ NETSEC_REG_NRM_RX_CONFIG,
+};
+
+static const u32 netsec_desc_start_reg_addr_up[] = {
+ NETSEC_REG_NRM_TX_DESC_START_UP,
+ NETSEC_REG_NRM_RX_DESC_START_UP,
+};
+
+static const u32 netsec_desc_start_reg_addr_lw[] = {
+ NETSEC_REG_NRM_TX_DESC_START_LW,
+ NETSEC_REG_NRM_RX_DESC_START_LW,
+};
+
+static int netsec_wait_for_ring_config_ready(struct netsec_priv *priv, int ring)
+{
+ int timeout = WAIT_FW_RDY_TIMEOUT;
+
+ while (--timeout && (netsec_readl(priv, desc_ads[ring]) &
+ NETSEC_REG_DESC_RING_CONFIG_CFG_UP))
+ usleep_range(1000, 2000);
+
+ if (!timeout) {
+ netif_err(priv, hw, priv->ndev,
+ "%s: timeout\n", __func__);
+ return -ETIMEDOUT;
+ }
+
+ return 0;
+}
+
+static u32 netsec_calc_pkt_ctrl_reg_param(const struct netsec_pkt_ctrlaram
+ *pkt_ctrlaram_p)
+{
+ u32 param = NETSEC_PKT_CTRL_REG_MODE_NRM;
+
+ if (pkt_ctrlaram_p->log_chksum_er_flag)
+ param |= NETSEC_PKT_CTRL_REG_LOG_CHKSUM_ER;
+
+ if (pkt_ctrlaram_p->log_hd_imcomplete_flag)
+ param |= NETSEC_PKT_CTRL_REG_LOG_HD_INCOMPLETE;
+
+ if (pkt_ctrlaram_p->log_hd_er_flag)
+ param |= NETSEC_PKT_CTRL_REG_LOG_HD_ER;
+
+ return param;
+}
+
+static int netsec_configure_normal_mode(struct netsec_priv *priv)
+{
+ int ret = 0;
+ u32 value;
+
+ /* save scb set value */
+ priv->scb_set_normal_tx_paddr = (phys_addr_t)netsec_readl(priv,
+ netsec_desc_start_reg_addr_up[NETSEC_RING_TX]) << 32;
+ priv->scb_set_normal_tx_paddr |= (phys_addr_t)netsec_readl(priv,
+ netsec_desc_start_reg_addr_lw[NETSEC_RING_TX]);
+
+ /* set desc_start addr */
+ netsec_writel(priv, netsec_desc_start_reg_addr_up[NETSEC_RING_RX],
+ priv->desc_ring[NETSEC_RING_RX].desc_phys >> 32);
+ netsec_writel(priv, netsec_desc_start_reg_addr_lw[NETSEC_RING_RX],
+ priv->desc_ring[NETSEC_RING_RX].desc_phys & 0xffffffff);
+
+ netsec_writel(priv, netsec_desc_start_reg_addr_up[NETSEC_RING_TX],
+ priv->desc_ring[NETSEC_RING_TX].desc_phys >> 32);
+ netsec_writel(priv, netsec_desc_start_reg_addr_lw[NETSEC_RING_TX],
+ priv->desc_ring[NETSEC_RING_TX].desc_phys & 0xffffffff);
+
+ /* set normal tx desc ring config */
+ value = (cpu_to_le32(1) == 1) << NETSEC_REG_DESC_ENDIAN |
+ NETSEC_REG_DESC_RING_CONFIG_CFG_UP |
+ NETSEC_REG_DESC_RING_CONFIG_CH_RST;
+ netsec_writel(priv, desc_ads[NETSEC_RING_TX], value);
+
+ value = (cpu_to_le32(1) == 1) << NETSEC_REG_DESC_ENDIAN |
+ NETSEC_REG_DESC_RING_CONFIG_CFG_UP |
+ NETSEC_REG_DESC_RING_CONFIG_CH_RST;
+ netsec_writel(priv, desc_ads[NETSEC_RING_RX], value);
+
+ if (netsec_wait_for_ring_config_ready(priv, NETSEC_RING_TX) ||
+ netsec_wait_for_ring_config_ready(priv, NETSEC_RING_RX))
+ return -ETIMEDOUT;
+
+ return ret;
+}
+
+static int netsec_change_mode_to_normal(struct netsec_priv *priv)
+{
+ u32 value;
+
+ priv->scb_pkt_ctrl_reg = netsec_readl(priv, NETSEC_REG_PKT_CTRL);
+
+ value = netsec_calc_pkt_ctrl_reg_param(&priv->param.pkt_ctrlaram);
+
+ if (priv->param.use_jumbo_pkt_flag)
+ value |= NETSEC_PKT_CTRL_REG_EN_JUMBO;
+
+ value |= NETSEC_PKT_CTRL_REG_MODE_NRM;
+
+ /* change to normal mode */
+ netsec_writel(priv, NETSEC_REG_DMA_MH_CTRL, MH_CTRL__MODE_TRANS);
+ netsec_writel(priv, NETSEC_REG_PKT_CTRL, value);
+
+ /* Wait Change mode Complete */
+ usleep_range(2000, 10000);
+
+ return 0;
+}
+
+static int netsec_change_mode_to_taiki(struct netsec_priv *priv)
+{
+ int ret = 0;
+ u32 value;
+
+ netsec_writel(priv, netsec_desc_start_reg_addr_up[NETSEC_RING_TX],
+ priv->scb_set_normal_tx_paddr >> 32);
+ netsec_writel(priv, netsec_desc_start_reg_addr_lw[NETSEC_RING_TX],
+ priv->scb_set_normal_tx_paddr & 0xffffffff);
+
+ value = NETSEC_REG_DESC_RING_CONFIG_CFG_UP |
+ NETSEC_REG_DESC_RING_CONFIG_CH_RST;
+
+ netsec_writel(priv, desc_ads[NETSEC_RING_TX], value);
+
+ if (netsec_wait_for_ring_config_ready(priv, NETSEC_RING_TX))
+ return -ETIMEDOUT;
+
+ netsec_writel(priv, NETSEC_REG_DMA_MH_CTRL, MH_CTRL__MODE_TRANS);
+ netsec_writel(priv, NETSEC_REG_PKT_CTRL, priv->scb_pkt_ctrl_reg);
+
+ /* Wait Change mode Complete */
+ usleep_range(2000, 10000);
+
+ return ret;
+}
+
+static int netsec_clear_modechange_irq(struct netsec_priv *priv, u32 value)
+{
+ netsec_writel(priv, NETSEC_REG_MODE_TRANS_COMP_STATUS,
+ (value & (NETSEC_MODE_TRANS_COMP_IRQ_N2T |
+ NETSEC_MODE_TRANS_COMP_IRQ_T2N)));
+ return 0;
+}
+
+static int netsec_hw_configure_to_normal(struct netsec_priv *priv)
+{
+ int err;
+
+ err = netsec_configure_normal_mode(priv);
+ if (err) {
+ netif_err(priv, drv, priv->ndev,
+ "%s: normal conf fail\n", __func__);
+ return err;
+ }
+ err = netsec_change_mode_to_normal(priv);
+ if (err) {
+ netif_err(priv, drv, priv->ndev,
+ "%s: normal set fail\n", __func__);
+ return err;
+ }
+
+ return err;
+}
+
+static int netsec_hw_configure_to_taiki(struct netsec_priv *priv)
+{
+ int ret;
+
+ ret = netsec_change_mode_to_taiki(priv);
+ if (ret) {
+ netif_err(priv, drv, priv->ndev,
+ "%s: taiki set fail\n", __func__);
+ return ret;
+ }
+
+ /* Clear mode change complete IRQ */
+ ret = netsec_clear_modechange_irq(priv, NETSEC_MODE_TRANS_COMP_IRQ_T2N
+ | NETSEC_MODE_TRANS_COMP_IRQ_N2T);
+
+ if (ret)
+ netif_err(priv, drv, priv->ndev,
+ "%s: clear mode fail\n", __func__);
+
+ return ret;
+}
+
+static void netsec_ring_irq_clr(struct netsec_priv *priv,
+ unsigned int id, u32 value)
+{
+ netsec_writel(priv, desc_ring_irq_status_reg_addr[id],
+ value & (NETSEC_IRQ_EMPTY | NETSEC_IRQ_ERR));
+}
+
+static void netsec_napi_tx_processing(struct napi_struct *napi_p)
+{
+ struct netsec_priv *priv = container_of(napi_p,
+ struct netsec_priv, napi);
+
+ netsec_ring_irq_clr(priv, NETSEC_RING_TX, NETSEC_IRQ_EMPTY);
+ netsec_clean_tx_desc_ring(priv);
+
+ if (netif_queue_stopped(priv->ndev) &&
+ netsec_get_tx_avail_num(priv) >= NETSEC_NETDEV_TX_PKT_SCAT_NUM_MAX)
+ netif_wake_queue(priv->ndev);
+}
+
+int netsec_netdev_napi_poll(struct napi_struct *napi_p, int budget)
+{
+ struct netsec_priv *priv = container_of(napi_p,
+ struct netsec_priv, napi);
+ struct net_device *ndev = priv->ndev;
+ struct netsec_rx_pkt_info rx_info;
+ int ret, done = 0, rx_num = 0;
+ struct netsec_frag_info frag;
+ struct sk_buff *skb;
+ u16 len;
+
+ netsec_napi_tx_processing(napi_p);
+
+ while (done < budget) {
+ if (!rx_num) {
+ rx_num = netsec_get_rx_num(priv);
+ if (!rx_num)
+ break;
+ }
+ done++;
+ rx_num--;
+ ret = netsec_get_rx_pkt_data(priv, &rx_info, &frag, &len, &skb);
+ if (unlikely(ret == -ENOMEM)) {
+ netif_err(priv, drv, priv->ndev,
+ "%s: rx fail %d\n", __func__, ret);
+ ndev->stats.rx_dropped++;
+ continue;
+ }
+ dma_unmap_single(priv->dev, frag.dma_addr, frag.len,
+ DMA_FROM_DEVICE);
+ skb_put(skb, len);
+ skb->protocol = eth_type_trans(skb, priv->ndev);
+
+ if (priv->rx_cksum_offload_flag &&
+ rx_info.rx_cksum_result == NETSEC_RX_CKSUM_OK)
+ skb->ip_summed = CHECKSUM_UNNECESSARY;
+
+ napi_gro_receive(napi_p, skb);
+
+ ndev->stats.rx_packets++;
+ ndev->stats.rx_bytes += len;
+ }
+
+ if (done == budget)
+ return budget;
+
+ napi_complete(napi_p);
+ netsec_writel(priv, NETSEC_REG_INTEN_SET,
+ NETSEC_IRQ_TX | NETSEC_IRQ_RX);
+
+ return done;
+}
+
+static netdev_tx_t netsec_netdev_start_xmit(struct sk_buff *skb,
+ struct net_device *ndev)
+{
+ struct netsec_priv *priv = netdev_priv(ndev);
+ struct netsec_tx_pkt_ctrl tx_ctrl;
+ u16 pend_tx, tso_seg_len = 0;
+ skb_frag_t *frag;
+ int count_frags;
+ int ret, i;
+
+ memset(&tx_ctrl, 0, sizeof(struct netsec_tx_pkt_ctrl));
+
+ netsec_ring_irq_clr(priv, NETSEC_RING_TX, NETSEC_IRQ_EMPTY);
+
+ count_frags = skb_shinfo(skb)->nr_frags + 1;
+
+ if (skb->ip_summed == CHECKSUM_PARTIAL)
+ tx_ctrl.cksum_offload_flag = true;
+
+ if (skb_is_gso(skb))
+ tso_seg_len = skb_shinfo(skb)->gso_size;
+
+ if (tso_seg_len > 0) {
+ if (skb->protocol == htons(ETH_P_IP)) {
+ ip_hdr(skb)->tot_len = 0;
+ tcp_hdr(skb)->check =
+ ~tcp_v4_check(0, ip_hdr(skb)->saddr,
+ ip_hdr(skb)->daddr, 0);
+ } else {
+ ipv6_hdr(skb)->payload_len = 0;
+ tcp_hdr(skb)->check =
+ ~csum_ipv6_magic(&ipv6_hdr(skb)->saddr,
+ &ipv6_hdr(skb)->daddr,
+ 0, IPPROTO_TCP, 0);
+ }
+
+ tx_ctrl.tcp_seg_offload_flag = true;
+ tx_ctrl.tcp_seg_len = tso_seg_len;
+ }
+
+ priv->tx_info[0].dma_addr = dma_map_single(priv->dev, skb->data,
+ skb_headlen(skb),
+ DMA_TO_DEVICE);
+ if (dma_mapping_error(priv->dev, priv->tx_info[0].dma_addr)) {
+ netif_err(priv, drv, priv->ndev,
+ "%s: DMA mapping failed\n", __func__);
+ return NETDEV_TX_OK;
+ }
+ priv->tx_info[0].addr = skb->data;
+ priv->tx_info[0].len = skb_headlen(skb);
+
+ for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) {
+ frag = &skb_shinfo(skb)->frags[i];
+ priv->tx_info[i + 1].dma_addr =
+ skb_frag_dma_map(priv->dev, frag, 0,
+ skb_frag_size(frag), DMA_TO_DEVICE);
+ priv->tx_info[i + 1].addr = skb_frag_address(frag);
+ priv->tx_info[i + 1].len = frag->size;
+ }
+
+ netsec_mark_skb_type(skb, NETSEC_RING_TX);
+
+ ret = netsec_set_tx_pkt_data(priv, &tx_ctrl, count_frags,
+ priv->tx_info, skb);
+ if (ret) {
+ netif_info(priv, drv, priv->ndev,
+ "set tx pkt failed %d\n", ret);
+ for (i = 0; i < count_frags; i++)
+ dma_unmap_single(priv->dev, priv->tx_info[i].dma_addr,
+ priv->tx_info[i].len, DMA_TO_DEVICE);
+ ndev->stats.tx_dropped++;
+
+ return NETDEV_TX_OK;
+ }
+
+ netdev_sent_queue(priv->ndev, skb->len);
+
+ spin_lock(&priv->tx_queue_lock);
+ pend_tx = netsec_get_tx_avail_num(priv);
+
+ if (pend_tx < NETSEC_NETDEV_TX_PKT_SCAT_NUM_MAX) {
+ netsec_ring_irq_enable(priv, NETSEC_RING_TX, NETSEC_IRQ_EMPTY);
+ netif_stop_queue(ndev);
+ goto err;
+ }
+ if (pend_tx <= DESC_NUM - 2) {
+ netsec_ring_irq_enable(priv, NETSEC_RING_TX, NETSEC_IRQ_EMPTY);
+ goto err;
+ }
+ netsec_ring_irq_disable(priv, NETSEC_RING_TX, NETSEC_IRQ_EMPTY);
+
+err:
+ spin_unlock(&priv->tx_queue_lock);
+
+ return NETDEV_TX_OK;
+}
+
+static int netsec_netdev_set_features(struct net_device *ndev,
+ netdev_features_t features)
+{
+ struct netsec_priv *priv = netdev_priv(ndev);
+
+ priv->rx_cksum_offload_flag = !!(features & NETIF_F_RXCSUM);
+
+ return 0;
+}
+
+static void netsec_phy_adjust_link(struct net_device *ndev)
+{
+ struct netsec_priv *priv = netdev_priv(ndev);
+
+ if (priv->actual_link_speed == ndev->phydev->speed &&
+ priv->actual_duplex == ndev->phydev->duplex)
+ return;
+
+ netsec_stop_gmac(priv);
+ netsec_start_gmac(priv);
+}
+
+static irqreturn_t netsec_irq_handler(int irq, void *dev_id)
+{
+ struct netsec_priv *priv = dev_id;
+ u32 status = netsec_readl(priv, NETSEC_REG_TOP_STATUS) &
+ netsec_readl(priv, NETSEC_REG_TOP_INTEN);
+
+ if (!status)
+ return IRQ_NONE;
+
+ if (status & (NETSEC_IRQ_TX | NETSEC_IRQ_RX)) {
+ netsec_writel(priv, NETSEC_REG_INTEN_CLR,
+ status & (NETSEC_IRQ_TX | NETSEC_IRQ_RX));
+ napi_schedule(&priv->napi);
+ }
+
+ return IRQ_HANDLED;
+}
+
+static int netsec_netdev_open(struct net_device *ndev)
+{
+ struct netsec_priv *priv = netdev_priv(ndev);
+ struct phy_device *phydev = NULL;
+ u32 scb_irq_temp;
+ int ret, n;
+
+ scb_irq_temp = netsec_readl(priv, NETSEC_REG_TOP_INTEN);
+
+ for (n = 0; n <= NETSEC_RING_MAX; n++) {
+ ret = netsec_alloc_desc_ring(priv, n);
+ if (ret) {
+ netif_err(priv, probe, priv->ndev,
+ "%s: alloc ring failed\n", __func__);
+ goto err;
+ }
+ }
+
+ ret = netsec_setup_rx_desc(priv, &priv->desc_ring[NETSEC_RING_RX]);
+ if (ret) {
+ netif_err(priv, probe, priv->ndev,
+ "%s: fail setup ring\n", __func__);
+ goto err1;
+ }
+
+ pm_runtime_get_sync(priv->dev);
+
+ netsec_writel(priv, NETSEC_REG_INTEN_CLR, scb_irq_temp);
+
+ ret = netsec_hw_configure_to_normal(priv);
+ if (ret) {
+ netif_err(priv, probe, priv->ndev,
+ "%s: normal fail %d\n", __func__, ret);
+ goto err1;
+ }
+
+ ret = request_irq(priv->ndev->irq, netsec_irq_handler,
+ IRQF_SHARED, "netsec", priv);
+ if (ret) {
+ netif_err(priv, drv, priv->ndev, "request_irq failed\n");
+ goto err1;
+ }
+ priv->irq_registered = true;
+
+ ret = netsec_clean_rx_desc_ring(priv);
+ if (ret) {
+ netif_err(priv, drv, priv->ndev,
+ "%s: clean rx desc fail\n", __func__);
+ goto err2;
+ }
+
+ ret = netsec_clean_tx_desc_ring(priv);
+ if (ret) {
+ netif_err(priv, drv, priv->ndev,
+ "%s: clean tx desc fail\n", __func__);
+ goto err2;
+ }
+
+ netsec_ring_irq_clr(priv, NETSEC_RING_TX, NETSEC_IRQ_EMPTY);
+
+ phydev = of_phy_connect(priv->ndev, priv->phy_np,
+ &netsec_phy_adjust_link, 0,
+ priv->phy_interface);
+ if (!phydev) {
+ netif_err(priv, link, priv->ndev, "missing PHY\n");
+ goto err2;
+ }
+
+ phy_start_aneg(phydev);
+
+ netsec_ring_irq_disable(priv, NETSEC_RING_TX, NETSEC_IRQ_EMPTY);
+
+ netsec_start_gmac(priv);
+ napi_enable(&priv->napi);
+ netif_start_queue(ndev);
+
+ netsec_writel(priv, NETSEC_REG_INTEN_SET,
+ NETSEC_IRQ_TX | NETSEC_IRQ_RX);
+
+ return 0;
+
+err2:
+ pm_runtime_put_sync(priv->dev);
+ free_irq(priv->ndev->irq, priv);
+ priv->irq_registered = false;
+err1:
+ for (n = 0; n <= NETSEC_RING_MAX; n++)
+ netsec_free_desc_ring(priv, &priv->desc_ring[n]);
+err:
+ netsec_writel(priv, NETSEC_REG_INTEN_SET, scb_irq_temp);
+
+ pm_runtime_put_sync(priv->dev);
+
+ return ret;
+}
+
+static int netsec_netdev_stop(struct net_device *ndev)
+{
+ struct netsec_priv *priv = netdev_priv(ndev);
+ int n;
+
+ phy_stop(ndev->phydev);
+ phy_disconnect(ndev->phydev);
+
+ netif_stop_queue(priv->ndev);
+ napi_disable(&priv->napi);
+
+ netsec_writel(priv, NETSEC_REG_INTEN_CLR, ~0);
+ netsec_stop_gmac(priv);
+ WARN_ON(netsec_hw_configure_to_taiki(priv));
+
+ pm_runtime_put_sync(priv->dev);
+
+ for (n = 0; n <= NETSEC_RING_MAX; n++)
+ netsec_free_desc_ring(priv, &priv->desc_ring[n]);
+
+ free_irq(priv->ndev->irq, priv);
+ priv->irq_registered = false;
+
+ return 0;
+}
+
+const struct net_device_ops netsec_netdev_ops = {
+ .ndo_open = netsec_netdev_open,
+ .ndo_stop = netsec_netdev_stop,
+ .ndo_start_xmit = netsec_netdev_start_xmit,
+ .ndo_set_features = netsec_netdev_set_features,
+ .ndo_set_mac_address = eth_mac_addr,
+ .ndo_validate_addr = eth_validate_addr,
+};
new file mode 100644
@@ -0,0 +1,330 @@
+/**
+ * drivers/net/ethernet/socionext/netsec/netsec_platform.c
+ *
+ * Copyright (C) 2013-2014 Fujitsu Semiconductor Limited.
+ * Copyright (C) 2014 Linaro Ltd Andy Green <andy.green@linaro.org>
+ * All rights reserved.
+ *
+ * This program is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU General Public License
+ * as published by the Free Software Foundation; either version 2
+ * of the License, or (at your option) any later version.
+ */
+
+#include <linux/device.h>
+#include <linux/ctype.h>
+#include <linux/netdevice.h>
+#include <linux/types.h>
+#include <linux/bitops.h>
+#include <linux/dma-mapping.h>
+#include <linux/module.h>
+#include <linux/sizes.h>
+#include <linux/platform_device.h>
+#include <linux/clk.h>
+#include <linux/of.h>
+#include <linux/of_device.h>
+#include <linux/of_net.h>
+#include <linux/io.h>
+#include <linux/pm_runtime.h>
+
+#include "netsec.h"
+
+#define NETSEC_F_NETSEC_VER_MAJOR_NUM(x) (x & 0xffff0000)
+
+static int napi_weight = 64;
+unsigned short pause_time = 256;
+
+static int netsec_probe(struct platform_device *pdev)
+{
+ struct net_device *ndev;
+ struct netsec_priv *priv;
+ struct resource *res;
+ const void *mac;
+ bool use_jumbo;
+ u32 hw_ver;
+ int err;
+ int ret;
+
+ ndev = alloc_etherdev(sizeof(*priv));
+ if (!ndev)
+ return -ENOMEM;
+
+ priv = netdev_priv(ndev);
+ priv->ndev = ndev;
+ SET_NETDEV_DEV(ndev, &pdev->dev);
+ platform_set_drvdata(pdev, priv);
+ priv->dev = &pdev->dev;
+
+ priv->msg_enable = NETIF_MSG_TX_ERR | NETIF_MSG_HW | NETIF_MSG_DRV |
+ NETIF_MSG_LINK | NETIF_MSG_PROBE;
+
+ mac = of_get_mac_address(pdev->dev.of_node);
+ if (mac)
+ ether_addr_copy(ndev->dev_addr, mac);
+
+ if (!is_valid_ether_addr(ndev->dev_addr)) {
+ eth_hw_addr_random(ndev);
+ dev_warn(&pdev->dev, "No MAC address found, using random\n");
+ }
+
+ priv->phy_np = of_parse_phandle(pdev->dev.of_node, "phy-handle", 0);
+ if (!priv->phy_np) {
+ netif_err(priv, probe, ndev, "missing phy in DT\n");
+ goto err1;
+ }
+
+ priv->phy_interface = of_get_phy_mode(pdev->dev.of_node);
+ if (priv->phy_interface < 0) {
+ netif_err(priv, probe, ndev,
+ "%s: bad phy-if\n", __func__);
+ goto err1;
+ }
+
+ priv->ioaddr = of_iomap(priv->dev->of_node, 0);
+ if (!priv->ioaddr) {
+ netif_err(priv, probe, ndev, "of_iomap() failed\n");
+ err = -EINVAL;
+ goto err1;
+ }
+
+ res = platform_get_resource(pdev, IORESOURCE_MEM, 1);
+ if (!res) {
+ netif_err(priv, probe, ndev,
+ "Missing rdlar resource\n");
+ goto err1;
+ }
+ priv->rdlar_pa = res->start;
+
+ res = platform_get_resource(pdev, IORESOURCE_MEM, 2);
+ if (!res) {
+ netif_err(priv, probe, ndev,
+ "Missing tdlar resource\n");
+ goto err1;
+ }
+ priv->tdlar_pa = res->start;
+
+ res = platform_get_resource(pdev, IORESOURCE_IRQ, 0);
+ if (!res) {
+ netif_err(priv, probe, ndev,
+ "Missing IRQ resource\n");
+ goto err2;
+ }
+ ndev->irq = res->start;
+
+ while (priv->clock_count < ARRAY_SIZE(priv->clk)) {
+ priv->clk[priv->clock_count] =
+ of_clk_get(pdev->dev.of_node, priv->clock_count);
+ if (IS_ERR(priv->clk[priv->clock_count])) {
+ if (!priv->clock_count) {
+ netif_err(priv, probe, ndev,
+ "Failed to get clock\n");
+ goto err3;
+ }
+ break;
+ }
+ priv->clock_count++;
+ }
+
+ /* disable by default */
+ priv->et_coalesce.rx_coalesce_usecs = 0;
+ priv->et_coalesce.rx_max_coalesced_frames = 1;
+ priv->et_coalesce.tx_coalesce_usecs = 0;
+ priv->et_coalesce.tx_max_coalesced_frames = 1;
+
+ use_jumbo = of_property_read_bool(pdev->dev.of_node, "use-jumbo");
+ priv->param.use_jumbo_pkt_flag = use_jumbo;
+
+ if (priv->param.use_jumbo_pkt_flag)
+ priv->rx_pkt_buf_len = NETSEC_RX_JUMBO_PKT_BUF_LEN;
+ else
+ priv->rx_pkt_buf_len = NETSEC_RX_PKT_BUF_LEN;
+
+ pm_runtime_enable(&pdev->dev);
+ /* runtime_pm coverage just for probe, open/close also cover it */
+ pm_runtime_get_sync(&pdev->dev);
+
+ hw_ver = netsec_readl(priv, NETSEC_REG_F_TAIKI_VER);
+ /* this driver only supports F_TAIKI style NETSEC */
+ if (NETSEC_F_NETSEC_VER_MAJOR_NUM(hw_ver) !=
+ NETSEC_F_NETSEC_VER_MAJOR_NUM(NETSEC_REG_NETSEC_VER_F_TAIKI)) {
+ ret = -ENODEV;
+ goto err3;
+ }
+
+ dev_info(&pdev->dev, "IP rev %d.%d\n", hw_ver >> 16, hw_ver & 0xffff);
+
+ priv->mac_mode.flow_start_th = NETSEC_FLOW_CONTROL_START_THRESHOLD;
+ priv->mac_mode.flow_stop_th = NETSEC_FLOW_CONTROL_STOP_THRESHOLD;
+ priv->mac_mode.pause_time = pause_time;
+ priv->mac_mode.flow_ctrl_enable_flag = false;
+ priv->freq = clk_get_rate(priv->clk[0]);
+
+ netif_napi_add(ndev, &priv->napi, netsec_netdev_napi_poll,
+ napi_weight);
+
+ /* MTU range */
+ ndev->min_mtu = ETH_MIN_MTU;
+ ndev->max_mtu = NETSEC_RX_JUMBO_PKT_BUF_LEN;
+
+ ndev->netdev_ops = &netsec_netdev_ops;
+ ndev->ethtool_ops = &netsec_ethtool_ops;
+ ndev->features = NETIF_F_SG | NETIF_F_IP_CSUM |
+ NETIF_F_IPV6_CSUM | NETIF_F_TSO |
+ NETIF_F_TSO6 | NETIF_F_GSO |
+ NETIF_F_HIGHDMA | NETIF_F_RXCSUM;
+ ndev->hw_features = ndev->features;
+
+ priv->rx_cksum_offload_flag = true;
+ spin_lock_init(&priv->tx_queue_lock);
+
+ err = netsec_mii_register(priv);
+ if (err) {
+ netif_err(priv, probe, ndev,
+ "mii bus registration failed %d\n", err);
+ goto err3;
+ }
+
+ /* disable all other interrupt sources */
+ netsec_writel(priv, NETSEC_REG_INTEN_CLR, ~0);
+ netsec_writel(priv, NETSEC_REG_INTEN_SET,
+ NETSEC_IRQ_TX | NETSEC_IRQ_RX);
+
+ err = register_netdev(ndev);
+ if (err) {
+ netif_err(priv, probe, ndev, "register_netdev() failed\n");
+ goto err4;
+ }
+
+ pm_runtime_put_sync_suspend(&pdev->dev);
+
+ netif_info(priv, probe, ndev, "initialized\n");
+
+ return 0;
+
+err4:
+ netsec_mii_unregister(priv);
+
+err3:
+ pm_runtime_put_sync_suspend(&pdev->dev);
+ pm_runtime_disable(&pdev->dev);
+ while (priv->clock_count > 0) {
+ priv->clock_count--;
+ clk_put(priv->clk[priv->clock_count]);
+ }
+err2:
+ iounmap(priv->ioaddr);
+err1:
+ free_netdev(ndev);
+
+ dev_err(&pdev->dev, "init failed\n");
+
+ return ret;
+}
+
+static int netsec_remove(struct platform_device *pdev)
+{
+ struct netsec_priv *priv = platform_get_drvdata(pdev);
+
+ unregister_netdev(priv->ndev);
+ netsec_mii_unregister(priv);
+ pm_runtime_disable(&pdev->dev);
+ iounmap(priv->ioaddr);
+ free_netdev(priv->ndev);
+
+ return 0;
+}
+
+#ifdef CONFIG_PM
+static int netsec_runtime_suspend(struct device *dev)
+{
+ struct netsec_priv *priv = dev_get_drvdata(dev);
+ int n;
+
+ netif_dbg(priv, drv, priv->ndev, "%s\n", __func__);
+
+ if (priv->irq_registered)
+ disable_irq(priv->ndev->irq);
+
+ netsec_writel(priv, NETSEC_REG_CLK_EN, 0);
+
+ for (n = priv->clock_count - 1; n >= 0; n--)
+ clk_disable_unprepare(priv->clk[n]);
+
+ return 0;
+}
+
+static int netsec_runtime_resume(struct device *dev)
+{
+ struct netsec_priv *priv = dev_get_drvdata(dev);
+ int n;
+
+ netif_dbg(priv, drv, priv->ndev, "%s\n", __func__);
+
+ /* first let the clocks back on */
+
+ for (n = 0; n < priv->clock_count; n++)
+ clk_prepare_enable(priv->clk[n]);
+
+ netsec_writel(priv, NETSEC_REG_CLK_EN, NETSEC_CLK_EN_REG_DOM_D |
+ NETSEC_CLK_EN_REG_DOM_C | NETSEC_CLK_EN_REG_DOM_G);
+
+ if (priv->irq_registered)
+ enable_irq(priv->ndev->irq);
+
+ return 0;
+}
+
+static int netsec_pm_suspend(struct device *dev)
+{
+ struct netsec_priv *priv = dev_get_drvdata(dev);
+
+ netif_dbg(priv, drv, priv->ndev, "%s\n", __func__);
+
+ if (pm_runtime_status_suspended(dev))
+ return 0;
+
+ return netsec_runtime_suspend(dev);
+}
+
+static int netsec_pm_resume(struct device *dev)
+{
+ struct netsec_priv *priv = dev_get_drvdata(dev);
+
+ netif_dbg(priv, drv, priv->ndev, "%s\n", __func__);
+
+ if (pm_runtime_status_suspended(dev))
+ return 0;
+
+ return netsec_runtime_resume(dev);
+}
+#endif
+
+static const struct dev_pm_ops netsec_pm_ops = {
+ SET_SYSTEM_SLEEP_PM_OPS(netsec_pm_suspend, netsec_pm_resume)
+ SET_RUNTIME_PM_OPS(netsec_runtime_suspend, netsec_runtime_resume, NULL)
+};
+
+static const struct of_device_id netsec_dt_ids[] = {
+ {.compatible = "socionext,netsecv5"},
+ { /* sentinel */ }
+};
+
+MODULE_DEVICE_TABLE(of, netsec_dt_ids);
+
+static struct platform_driver netsec_driver = {
+ .probe = netsec_probe,
+ .remove = netsec_remove,
+ .driver = {
+ .name = "netsec",
+ .of_match_table = netsec_dt_ids,
+ .pm = &netsec_pm_ops,
+ },
+};
+
+module_platform_driver(netsec_driver);
+
+MODULE_AUTHOR("Socionext Inc");
+MODULE_DESCRIPTION("NETSEC Ethernet driver");
+MODULE_LICENSE("GPL");
+
+MODULE_ALIAS("platform:netsec");
This driver adds support for Socionext "netsec" IP Gigabit Ethernet + PHY IP used in a variety of their ARM-based ASICs. Signed-off-by: Jassi Brar <jaswinder.singh@linaro.org> --- drivers/net/ethernet/Kconfig | 1 + drivers/net/ethernet/Makefile | 1 + drivers/net/ethernet/socionext/Kconfig | 29 + drivers/net/ethernet/socionext/Makefile | 1 + drivers/net/ethernet/socionext/netsec/Makefile | 6 + drivers/net/ethernet/socionext/netsec/netsec.h | 386 +++++++++++++ .../socionext/netsec/netsec_desc_ring_access.c | 618 +++++++++++++++++++++ .../net/ethernet/socionext/netsec/netsec_ethtool.c | 76 +++ .../ethernet/socionext/netsec/netsec_gmac_access.c | 329 +++++++++++ .../net/ethernet/socionext/netsec/netsec_netdev.c | 558 +++++++++++++++++++ .../ethernet/socionext/netsec/netsec_platform.c | 330 +++++++++++ 11 files changed, 2335 insertions(+) create mode 100644 drivers/net/ethernet/socionext/Kconfig create mode 100644 drivers/net/ethernet/socionext/Makefile create mode 100644 drivers/net/ethernet/socionext/netsec/Makefile create mode 100644 drivers/net/ethernet/socionext/netsec/netsec.h create mode 100644 drivers/net/ethernet/socionext/netsec/netsec_desc_ring_access.c create mode 100644 drivers/net/ethernet/socionext/netsec/netsec_ethtool.c create mode 100644 drivers/net/ethernet/socionext/netsec/netsec_gmac_access.c create mode 100644 drivers/net/ethernet/socionext/netsec/netsec_netdev.c create mode 100644 drivers/net/ethernet/socionext/netsec/netsec_platform.c