From patchwork Wed Aug 30 10:26:11 2017 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Jassi Brar X-Patchwork-Id: 9929251 Return-Path: Received: from mail.wl.linuxfoundation.org (pdx-wl-mail.web.codeaurora.org [172.30.200.125]) by pdx-korg-patchwork.web.codeaurora.org (Postfix) with ESMTP id 03B9A6022E for ; Wed, 30 Aug 2017 10:27:11 +0000 (UTC) Received: from mail.wl.linuxfoundation.org (localhost [127.0.0.1]) by mail.wl.linuxfoundation.org (Postfix) with ESMTP id E55F528429 for ; Wed, 30 Aug 2017 10:27:10 +0000 (UTC) Received: by mail.wl.linuxfoundation.org (Postfix, from userid 486) id D99BA2847A; Wed, 30 Aug 2017 10:27:10 +0000 (UTC) X-Spam-Checker-Version: SpamAssassin 3.3.1 (2010-03-16) on pdx-wl-mail.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-2.6 required=2.0 tests=BAYES_00, DKIM_ADSP_CUSTOM_MED, DKIM_SIGNED, DKIM_VALID, FREEMAIL_FROM, RCVD_IN_DNSWL_LOW autolearn=ham version=3.3.1 Received: from bombadil.infradead.org (bombadil.infradead.org [65.50.211.133]) (using TLSv1.2 with cipher AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.wl.linuxfoundation.org (Postfix) with ESMTPS id D784E28429 for ; Wed, 30 Aug 2017 10:27:07 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; q=dns/txt; c=relaxed/relaxed; d=lists.infradead.org; s=bombadil.20170209; h=Sender: Content-Transfer-Encoding:Content-Type:MIME-Version:Cc:List-Subscribe: List-Help:List-Post:List-Archive:List-Unsubscribe:List-Id:References: In-Reply-To:Message-Id:Date:Subject:To:From:Reply-To:Content-ID: Content-Description:Resent-Date:Resent-From:Resent-Sender:Resent-To:Resent-Cc :Resent-Message-ID:List-Owner; bh=we2fTvMroV+8aTFMN1SQxoOgcS9SFShz+MGBeQmVwkg=; b=AEqbHA7ooPM/SX+zKmo7zJMlHq oYOZmXM8mdqrb5MBjJ98oWXYGaTOqkdIeOwuV4cIkNoLZ6B5+M//3H5rKGaOI80i/RX6qQo1bO2Mw fGTyz9Aj6zPFFlRKbEJJ1mIey4YxGKlCJ6giWMuLGlZXup4gP3zhfErYlzwGwjBuJn5M7ilVs+xgP JQGBm1xjq9NNiWEjziHAz3RSBfRr/xca36a1X1iPvl7LcurzlvwE6PuytqaF/IMMVVC58LaWBK/Uh hn4i7DNadR9800Umg1y5ShPQmthNLsBI3LpSxgj5EqzhWRzMRNK8Jm6MivHWURaPgsAZRZwYfP2aI cgZsqDcA==; Received: from localhost ([127.0.0.1] helo=bombadil.infradead.org) by bombadil.infradead.org with esmtp (Exim 4.87 #1 (Red Hat Linux)) id 1dn0D7-0000OD-D2; Wed, 30 Aug 2017 10:27:01 +0000 Received: from mail-pf0-x242.google.com ([2607:f8b0:400e:c00::242]) by bombadil.infradead.org with esmtps (Exim 4.87 #1 (Red Hat Linux)) id 1dn0Co-0008Mh-Hp for linux-arm-kernel@lists.infradead.org; Wed, 30 Aug 2017 10:26:53 +0000 Received: by mail-pf0-x242.google.com with SMTP id k3so4061505pfc.3 for ; Wed, 30 Aug 2017 03:26:22 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=from:to:cc:subject:date:message-id:in-reply-to:references; bh=2MOxWTdbLZvNlHmtCln5DxBhyp6zobrwm0b422WFvxY=; b=I2Z4NVfIrSWTlqKcmMyxKIZ5i6juin6hh0AoKDWpski8lHQ/+jSR1JtPZOLlrJ2qUW GPIcYeDNE5L8ywnGDtuoFCF0TDjL4CUkM5NkYJ9DPpulk8X/JNb2g68jr+4CkT7i5wCt e8hucBbElFxf08FBFDGFJZJ3Y2BENBnBNOvW/4+N3I8mhNNUti5Ms0OD7D2XHpQdgmjf dSEWHXC74XOuhApIqJP8gq3A8dF9O4HjwUjjMW1MdpQko+V31Ym4e7fCt6G4/XPXqJlu 01P5VY5st5JnWpKDT0K0jWtV1QJ+jpIMkYM0n3F2BbBHGwru0NNydmx36mmpAWuj/l2i /H1w== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references; bh=2MOxWTdbLZvNlHmtCln5DxBhyp6zobrwm0b422WFvxY=; b=sFkRrNo4RUyB/Gf8mnaOP1nWehdWNRO3f+mnRI1ednBTQ6pDGOi7CQYNQ2soZse9RR 4/ZPOXrcSD8bJ5yT9Gt6pFlsrhkmzsNaadpQTWaCJRSMfInOjxmABHzhv2IQ3+Q+Jhgg tMYD84p5FF1BlXqzZc50TLk2eHscrwkvi6jf0y2i87VHdUlaRhf9PsW7Z3i0s9KdMdJm XcP2cSPKivh6uzDxNIP4PD4wValbEU2cOPoe88nn/SuQ90zDx86yTLfScmDE0TYfHlVd /eTWWWtsm8JoOHILGlftL2KNFWVTjmavEP8HdYXAGGMeQ2S8cU7IE+uNqwVKvBxWySgI BSfw== X-Gm-Message-State: AHYfb5h1BhHAoJn8XY6f4KQ5eVk6OoHvR5xX/u6FW0Z3X+XpF3jxmwy6 njvyReiVm77rhg== X-Received: by 10.84.238.198 with SMTP id l6mr1326597pln.365.1504088780947; Wed, 30 Aug 2017 03:26:20 -0700 (PDT) Received: from localhost.localdomain ([27.255.192.36]) by smtp.gmail.com with ESMTPSA id w24sm10085884pfl.80.2017.08.30.03.26.16 (version=TLS1_2 cipher=ECDHE-RSA-AES128-SHA bits=128/128); Wed, 30 Aug 2017 03:26:19 -0700 (PDT) From: Jassi Brar X-Google-Original-From: Jassi Brar To: netdev@vger.kernel.org, devicetree@vger.kernel.org, linux-arm-kernel@lists.infradead.org, davem@davemloft.net Subject: [net-next PATCHv6 2/2] net: socionext: Add NetSec driver Date: Wed, 30 Aug 2017 15:56:11 +0530 Message-Id: <1504088771-6255-1-git-send-email-jaswinder.singh@linaro.org> X-Mailer: git-send-email 2.7.4 In-Reply-To: <1504088657-6102-1-git-send-email-jaswinder.singh@linaro.org> References: <1504088657-6102-1-git-send-email-jaswinder.singh@linaro.org> X-CRM114-Version: 20100106-BlameMichelson ( TRE 0.8.0 (BSD) ) MR-646709E3 X-CRM114-CacheID: sfid-20170830_032643_055792_7BCC7FB9 X-CRM114-Status: GOOD ( 17.72 ) X-BeenThere: linux-arm-kernel@lists.infradead.org X-Mailman-Version: 2.1.21 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: mark.rutland@arm.com, arnd@arndb.de, patches@linaro.org, Jassi Brar , robh+dt@kernel.org, andy@warmcat.com MIME-Version: 1.0 Sender: "linux-arm-kernel" Errors-To: linux-arm-kernel-bounces+patchwork-linux-arm=patchwork.kernel.org@lists.infradead.org X-Virus-Scanned: ClamAV using ClamSMTP This driver adds support for Socionext "netsec" IP Gigabit Ethernet + PHY IP used in a variety of their ARM-based ASICs. Signed-off-by: Jassi Brar --- drivers/net/ethernet/Kconfig | 1 + drivers/net/ethernet/Makefile | 1 + drivers/net/ethernet/socionext/Kconfig | 29 + drivers/net/ethernet/socionext/Makefile | 1 + drivers/net/ethernet/socionext/netsec/Makefile | 6 + drivers/net/ethernet/socionext/netsec/netsec.h | 386 +++++++++++++ .../socionext/netsec/netsec_desc_ring_access.c | 618 +++++++++++++++++++++ .../net/ethernet/socionext/netsec/netsec_ethtool.c | 76 +++ .../ethernet/socionext/netsec/netsec_gmac_access.c | 329 +++++++++++ .../net/ethernet/socionext/netsec/netsec_netdev.c | 558 +++++++++++++++++++ .../ethernet/socionext/netsec/netsec_platform.c | 330 +++++++++++ 11 files changed, 2335 insertions(+) create mode 100644 drivers/net/ethernet/socionext/Kconfig create mode 100644 drivers/net/ethernet/socionext/Makefile create mode 100644 drivers/net/ethernet/socionext/netsec/Makefile create mode 100644 drivers/net/ethernet/socionext/netsec/netsec.h create mode 100644 drivers/net/ethernet/socionext/netsec/netsec_desc_ring_access.c create mode 100644 drivers/net/ethernet/socionext/netsec/netsec_ethtool.c create mode 100644 drivers/net/ethernet/socionext/netsec/netsec_gmac_access.c create mode 100644 drivers/net/ethernet/socionext/netsec/netsec_netdev.c create mode 100644 drivers/net/ethernet/socionext/netsec/netsec_platform.c diff --git a/drivers/net/ethernet/Kconfig b/drivers/net/ethernet/Kconfig index edae15ac..ef23120 100644 --- a/drivers/net/ethernet/Kconfig +++ b/drivers/net/ethernet/Kconfig @@ -169,6 +169,7 @@ source "drivers/net/ethernet/sis/Kconfig" source "drivers/net/ethernet/sfc/Kconfig" source "drivers/net/ethernet/sgi/Kconfig" source "drivers/net/ethernet/smsc/Kconfig" +source "drivers/net/ethernet/socionext/Kconfig" source "drivers/net/ethernet/stmicro/Kconfig" source "drivers/net/ethernet/sun/Kconfig" source "drivers/net/ethernet/tehuti/Kconfig" diff --git a/drivers/net/ethernet/Makefile b/drivers/net/ethernet/Makefile index bf7f450..b2746b1 100644 --- a/drivers/net/ethernet/Makefile +++ b/drivers/net/ethernet/Makefile @@ -80,6 +80,7 @@ obj-$(CONFIG_SFC) += sfc/ obj-$(CONFIG_SFC_FALCON) += sfc/falcon/ obj-$(CONFIG_NET_VENDOR_SGI) += sgi/ obj-$(CONFIG_NET_VENDOR_SMSC) += smsc/ +obj-$(CONFIG_NET_VENDOR_SNI) += socionext/ obj-$(CONFIG_NET_VENDOR_STMICRO) += stmicro/ obj-$(CONFIG_NET_VENDOR_SUN) += sun/ obj-$(CONFIG_NET_VENDOR_TEHUTI) += tehuti/ diff --git a/drivers/net/ethernet/socionext/Kconfig b/drivers/net/ethernet/socionext/Kconfig new file mode 100644 index 0000000..e2bcf90 --- /dev/null +++ b/drivers/net/ethernet/socionext/Kconfig @@ -0,0 +1,29 @@ +# +# Socionext Network device configuration +# + +config NET_VENDOR_SNI + bool "Socionext devices" + default y + ---help--- + If you have a network (Ethernet) card belonging to this class, say Y. + + Note that the answer to this question doesn't directly affect the + the questions about Socionext cards. If you say Y, you will be asked for + your specific card in the following questions. + +if NET_VENDOR_SNI + +config SNI_NETSEC + tristate "NETSEC Driver Support" + depends on OF + select PHYLIB + select MII +help + Enable for NETSEC support of Socionext FGAMC4 IP + Provides Gigabit ethernet support + + To compile this driver as a module, choose M here: the module will be + called netsec. If unsure, say N. + +endif # NET_VENDOR_SNI diff --git a/drivers/net/ethernet/socionext/Makefile b/drivers/net/ethernet/socionext/Makefile new file mode 100644 index 0000000..9555899 --- /dev/null +++ b/drivers/net/ethernet/socionext/Makefile @@ -0,0 +1 @@ +obj-$(CONFIG_SNI_NETSEC) += netsec/ diff --git a/drivers/net/ethernet/socionext/netsec/Makefile b/drivers/net/ethernet/socionext/netsec/Makefile new file mode 100644 index 0000000..42f6bab --- /dev/null +++ b/drivers/net/ethernet/socionext/netsec/Makefile @@ -0,0 +1,6 @@ +obj-m := netsec.o +netsec-objs := netsec_desc_ring_access.o \ + netsec_netdev.o \ + netsec_ethtool.o \ + netsec_platform.o \ + netsec_gmac_access.o diff --git a/drivers/net/ethernet/socionext/netsec/netsec.h b/drivers/net/ethernet/socionext/netsec/netsec.h new file mode 100644 index 0000000..dadf5d9 --- /dev/null +++ b/drivers/net/ethernet/socionext/netsec/netsec.h @@ -0,0 +1,386 @@ +/** + * netsec.h + * + * Copyright (C) 2011 - 2014 Fujitsu Semiconductor Limited. + * Copyright (C) 2014 Linaro Ltd Andy Green + * All rights reserved. + * + * This program is free software; you can redistribute it and/or + * modify it under the terms of the GNU General Public License + * as published by the Free Software Foundation; either version 2 + * of the License, or (at your option) any later version. + */ +#ifndef NETSEC_INTERNAL_H +#define NETSEC_INTERNAL_H + +#include +#include +#include +#include +#include +#include +#include +#include +#include + +#define NETSEC_FLOW_CONTROL_START_THRESHOLD 36 +#define NETSEC_FLOW_CONTROL_STOP_THRESHOLD 48 + +#define NETSEC_CLK_MHZ 1000000 + +#define NETSEC_RX_PKT_BUF_LEN 1522 +#define NETSEC_RX_JUMBO_PKT_BUF_LEN 9022 + +#define NETSEC_NETDEV_TX_PKT_SCAT_NUM_MAX 19 + +#define DESC_NUM 128 + +#define NETSEC_TX_SHIFT_OWN_FIELD 31 +#define NETSEC_TX_SHIFT_LD_FIELD 30 +#define NETSEC_TX_SHIFT_DRID_FIELD 24 +#define NETSEC_TX_SHIFT_PT_FIELD 21 +#define NETSEC_TX_SHIFT_TDRID_FIELD 16 +#define NETSEC_TX_SHIFT_CC_FIELD 15 +#define NETSEC_TX_SHIFT_FS_FIELD 9 +#define NETSEC_TX_LAST 8 +#define NETSEC_TX_SHIFT_CO 7 +#define NETSEC_TX_SHIFT_SO 6 +#define NETSEC_TX_SHIFT_TRS_FIELD 4 + +#define NETSEC_RX_PKT_OWN_FIELD 31 +#define NETSEC_RX_PKT_LD_FIELD 30 +#define NETSEC_RX_PKT_SDRID_FIELD 24 +#define NETSEC_RX_PKT_FR_FIELD 23 +#define NETSEC_RX_PKT_ER_FIELD 21 +#define NETSEC_RX_PKT_ERR_FIELD 16 +#define NETSEC_RX_PKT_TDRID_FIELD 12 +#define NETSEC_RX_PKT_FS_FIELD 9 +#define NETSEC_RX_PKT_LS_FIELD 8 +#define NETSEC_RX_PKT_CO_FIELD 6 + +#define NETSEC_RX_PKT_ERR_MASK 3 + +#define NETSEC_MAX_TX_PKT_LEN 1518 +#define NETSEC_MAX_TX_JUMBO_PKT_LEN 9018 + +enum netsec_rings { + NETSEC_RING_TX, + NETSEC_RING_RX +}; + +#define NETSEC_RING_GMAC 15 +#define NETSEC_RING_MAX 1 + +#define NETSEC_TCP_SEG_LEN_MAX 1460 +#define NETSEC_TCP_JUMBO_SEG_LEN_MAX 8960 + +#define NETSEC_RX_CKSUM_NOTAVAIL 0 +#define NETSEC_RX_CKSUM_OK 1 +#define NETSEC_RX_CKSUM_NG 2 + +#define NETSEC_TOP_IRQ_REG_CODE_LOAD_END BIT(20) +#define NETSEC_IRQ_TRANSITION_COMPLETE BIT(4) +#define NETSEC_IRQ_RX BIT(1) +#define NETSEC_IRQ_TX BIT(0) + +#define NETSEC_IRQ_EMPTY BIT(17) +#define NETSEC_IRQ_ERR BIT(16) +#define NETSEC_IRQ_PKT_CNT BIT(15) +#define NETSEC_IRQ_TIMEUP BIT(14) +#define NETSEC_IRQ_RCV (NETSEC_IRQ_PKT_CNT | NETSEC_IRQ_TIMEUP) + +#define NETSEC_IRQ_TX_DONE BIT(15) +#define NETSEC_IRQ_SND (NETSEC_IRQ_TX_DONE | NETSEC_IRQ_TIMEUP) + +#define NETSEC_MODE_TRANS_COMP_IRQ_N2T BIT(20) +#define NETSEC_MODE_TRANS_COMP_IRQ_T2N BIT(19) + +#define NETSEC_DESC_MIN 2 +#define NETSEC_DESC_MAX 2047 +#define NETSEC_INT_PKTCNT_MAX 2047 + +#define NETSEC_FLOW_START_TH_MAX 95 +#define NETSEC_FLOW_STOP_TH_MAX 95 +#define NETSEC_FLOW_PAUSE_TIME_MIN 5 + +#define NETSEC_CLK_EN_REG_DOM_ALL 0x3f + +#define NETSEC_REG_TOP_STATUS 0x80 +#define NETSEC_REG_TOP_INTEN 0x81 +#define NETSEC_REG_INTEN_SET 0x8d +#define NETSEC_REG_INTEN_CLR 0x8e +#define NETSEC_REG_NRM_TX_STATUS 0x100 +#define NETSEC_REG_NRM_TX_INTEN 0x101 +#define NETSEC_REG_NRM_TX_INTEN_SET 0x10a +#define NETSEC_REG_NRM_TX_INTEN_CLR 0x10b +#define NETSEC_REG_NRM_RX_STATUS 0x110 +#define NETSEC_REG_NRM_RX_INTEN 0x111 +#define NETSEC_REG_NRM_RX_INTEN_SET 0x11a +#define NETSEC_REG_NRM_RX_INTEN_CLR 0x11b +#define NETSEC_REG_RESERVED_RX_DESC_START 0x122 +#define NETSEC_REG_RESERVED_TX_DESC_START 0x132 +#define NETSEC_REG_CLK_EN 0x40 +#define NETSEC_REG_SOFT_RST 0x41 +#define NETSEC_REG_PKT_CTRL 0x50 +#define NETSEC_REG_COM_INIT 0x48 +#define NETSEC_REG_DMA_TMR_CTRL 0x83 +#define NETSEC_REG_F_TAIKI_MC_VER 0x8b +#define NETSEC_REG_F_TAIKI_VER 0x8c +#define NETSEC_REG_DMA_HM_CTRL 0x85 +#define NETSEC_REG_DMA_MH_CTRL 0x88 +#define NETSEC_REG_NRM_TX_PKTCNT 0x104 +#define NETSEC_REG_NRM_TX_DONE_TXINT_PKTCNT 0x106 +#define NETSEC_REG_NRM_RX_RXINT_PKTCNT 0x116 +#define NETSEC_REG_NRM_TX_TXINT_TMR 0x108 +#define NETSEC_REG_NRM_RX_RXINT_TMR 0x118 +#define NETSEC_REG_NRM_TX_DONE_PKTCNT 0x105 +#define NETSEC_REG_NRM_RX_PKTCNT 0x115 +#define NETSEC_REG_NRM_TX_TMR 0x107 +#define NETSEC_REG_NRM_RX_TMR 0x117 +#define NETSEC_REG_NRM_TX_DESC_START_UP 0x10d +#define NETSEC_REG_NRM_TX_DESC_START_LW 0x102 +#define NETSEC_REG_NRM_RX_DESC_START_UP 0x11d +#define NETSEC_REG_NRM_RX_DESC_START_LW 0x112 +#define NETSEC_REG_NRM_TX_CONFIG 0x10c +#define NETSEC_REG_NRM_RX_CONFIG 0x11c +#define MAC_REG_DATA 0x470 +#define MAC_REG_CMD 0x471 +#define MAC_REG_FLOW_TH 0x473 +#define MAC_REG_INTF_SEL 0x475 +#define MAC_REG_DESC_INIT 0x47f +#define MAC_REG_DESC_SOFT_RST 0x481 +#define NETSEC_REG_MODE_TRANS_COMP_STATUS 0x140 +#define GMAC_REG_MCR 0x0000 +#define GMAC_REG_MFFR 0x0004 +#define GMAC_REG_GAR 0x0010 +#define GMAC_REG_GDR 0x0014 +#define GMAC_REG_FCR 0x0018 +#define GMAC_REG_BMR 0x1000 +#define GMAC_REG_RDLAR 0x100c +#define GMAC_REG_TDLAR 0x1010 +#define GMAC_REG_OMR 0x1018 + +#define NETSEC_PKT_CTRL_REG_MODE_NRM BIT(28) +#define NETSEC_PKT_CTRL_REG_EN_JUMBO BIT(27) +#define NETSEC_PKT_CTRL_REG_LOG_CHKSUM_ER BIT(3) +#define NETSEC_PKT_CTRL_REG_LOG_HD_INCOMPLETE BIT(2) +#define NETSEC_PKT_CTRL_REG_LOG_HD_ER BIT(1) +#define NETSEC_PKT_CTRL_REG_DRP_NO_MATCH BIT(0) + +#define NETSEC_CLK_EN_REG_DOM_G BIT(5) +#define NETSEC_CLK_EN_REG_DOM_C BIT(1) +#define NETSEC_CLK_EN_REG_DOM_D BIT(0) + +#define NETSEC_COM_INIT_REG_PKT BIT(1) +#define NETSEC_COM_INIT_REG_CORE BIT(0) + +#define NETSEC_SOFT_RST_REG_RESET 0 +#define NETSEC_SOFT_RST_REG_RUN BIT(31) + +#define NETSEC_DMA_CTRL_REG_STOP 1 +#define MH_CTRL__MODE_TRANS BIT(20) + +#define NETSEC_GMAC_CMD_ST_READ 0 +#define NETSEC_GMAC_CMD_ST_WRITE BIT(28) +#define NETSEC_GMAC_CMD_ST_BUSY BIT(31) + +#define NETSEC_GMAC_BMR_REG_COMMON 0x00412080 +#define NETSEC_GMAC_BMR_REG_RESET 0x00020181 +#define NETSEC_GMAC_BMR_REG_SWR 0x00000001 + +#define NETSEC_GMAC_OMR_REG_ST BIT(13) +#define NETSEC_GMAC_OMR_REG_SR BIT(1) + +#define NETSEC_GMAC_MCR_REG_IBN BIT(30) +#define NETSEC_GMAC_MCR_REG_CST BIT(25) +#define NETSEC_GMAC_MCR_REG_JE BIT(20) +#define NETSEC_MCR_PS BIT(15) +#define NETSEC_GMAC_MCR_REG_FES BIT(14) +#define NETSEC_GMAC_MCR_REG_FULL_DUPLEX_COMMON 0x0000280c +#define NETSEC_GMAC_MCR_REG_HALF_DUPLEX_COMMON 0x0001a00c + +#define NETSEC_FCR_RFE BIT(2) +#define NETSEC_FCR_TFE BIT(1) + +#define NETSEC_GMAC_GAR_REG_GW BIT(1) +#define NETSEC_GMAC_GAR_REG_GB BIT(0) + +#define NETSEC_GMAC_GAR_REG_SHIFT_PA 11 +#define NETSEC_GMAC_GAR_REG_SHIFT_GR 6 +#define GMAC_REG_SHIFT_CR_GAR 2 + +#define NETSEC_GMAC_GAR_REG_CR_25_35_MHZ 2 +#define NETSEC_GMAC_GAR_REG_CR_35_60_MHZ 3 +#define NETSEC_GMAC_GAR_REG_CR_60_100_MHZ 0 +#define NETSEC_GMAC_GAR_REG_CR_100_150_MHZ 1 +#define NETSEC_GMAC_GAR_REG_CR_150_250_MHZ 4 +#define NETSEC_GMAC_GAR_REG_CR_250_300_MHZ 5 + +#define NETSEC_REG_NETSEC_VER_F_TAIKI 0x50000 + +#define NETSEC_REG_DESC_RING_CONFIG_CFG_UP BIT(31) +#define NETSEC_REG_DESC_RING_CONFIG_CH_RST BIT(30) +#define NETSEC_REG_DESC_TMR_MODE 4 +#define NETSEC_REG_DESC_ENDIAN 0 + +#define NETSEC_MAC_DESC_SOFT_RST_SOFT_RST 1 +#define NETSEC_MAC_DESC_INIT_REG_INIT 1 + +/* this is used to interpret a register layout */ +struct netsec_pkt_ctrlaram { + u8 log_chksum_er_flag:1; + u8 log_hd_imcomplete_flag:1; + u8 log_hd_er_flag:1; +}; + +struct netsec_param { + struct netsec_pkt_ctrlaram pkt_ctrlaram; + bool use_jumbo_pkt_flag; +}; + +struct netsec_mac_mode { + u16 flow_start_th; + u16 flow_stop_th; + u16 pause_time; + bool flow_ctrl_enable_flag; +}; + +struct netsec_desc_ring { + spinlock_t spinlock_desc; /* protect descriptor access */ + phys_addr_t desc_phys; + struct netsec_frag_info *frag; + struct sk_buff **priv; + void *ring_vaddr; + enum netsec_rings id; + int len; + u16 tx_done_num; + u16 rx_num; + u16 head; + u16 tail; + bool running; + bool full; +}; + +struct netsec_frag_info { + dma_addr_t dma_addr; + void *addr; + u16 len; +}; + +struct netsec_priv { + struct netsec_desc_ring desc_ring[NETSEC_RING_MAX + 1]; + struct ethtool_coalesce et_coalesce; + struct netsec_mac_mode mac_mode; + struct netsec_param param; + struct napi_struct napi; + phys_addr_t rdlar_pa, tdlar_pa; + phy_interface_t phy_interface; + spinlock_t tx_queue_lock; /* protect transmit queue */ + struct netsec_frag_info tx_info[MAX_SKB_FRAGS]; + struct net_device *ndev; + struct device_node *phy_np; + struct mii_bus *mii_bus; + void __iomem *ioaddr; + struct device *dev; + struct clk *clk[3]; + phys_addr_t scb_set_normal_tx_paddr; + u32 scb_pkt_ctrl_reg; + u32 rx_pkt_buf_len; + u32 msg_enable; + u32 freq; + int actual_link_speed; + int clock_count; + bool rx_cksum_offload_flag; + bool actual_duplex; + bool irq_registered; +}; + +struct netsec_tx_de { + u32 attr; + u32 data_buf_addr_up; + u32 data_buf_addr_lw; + u32 buf_len_info; +}; + +struct netsec_rx_de { + u32 attr; + u32 data_buf_addr_up; + u32 data_buf_addr_lw; + u32 buf_len_info; +}; + +struct netsec_tx_pkt_ctrl { + u16 tcp_seg_len; + bool tcp_seg_offload_flag; + bool cksum_offload_flag; +}; + +struct netsec_rx_pkt_info { + int rx_cksum_result; + int err_code; + bool is_fragmented; + bool err_flag; +}; + +struct netsec_skb_cb { + bool is_rx; +}; + +static inline void netsec_writel(struct netsec_priv *priv, + u32 reg_addr, u32 val) +{ + writel_relaxed(val, priv->ioaddr + (reg_addr << 2)); +} + +static inline u32 netsec_readl(struct netsec_priv *priv, u32 reg_addr) +{ + return readl_relaxed(priv->ioaddr + (reg_addr << 2)); +} + +static inline void netsec_mark_skb_type(struct sk_buff *skb, bool is_rx) +{ + struct netsec_skb_cb *cb = (struct netsec_skb_cb *)skb->cb; + + cb->is_rx = is_rx; +} + +static inline bool skb_is_rx(struct sk_buff *skb) +{ + struct netsec_skb_cb *cb = (struct netsec_skb_cb *)skb->cb; + + return cb->is_rx; +} + +extern const struct net_device_ops netsec_netdev_ops; +extern const struct ethtool_ops netsec_ethtool_ops; + +int netsec_start_gmac(struct netsec_priv *priv); +int netsec_stop_gmac(struct netsec_priv *priv); +int netsec_mii_register(struct netsec_priv *priv); +void netsec_mii_unregister(struct netsec_priv *priv); +int netsec_start_desc_ring(struct netsec_priv *priv, enum netsec_rings id); +void netsec_stop_desc_ring(struct netsec_priv *priv, enum netsec_rings id); +u16 netsec_get_rx_num(struct netsec_priv *priv); +u16 netsec_get_tx_avail_num(struct netsec_priv *priv); +int netsec_clean_tx_desc_ring(struct netsec_priv *priv); +int netsec_clean_rx_desc_ring(struct netsec_priv *priv); +int netsec_set_tx_pkt_data(struct netsec_priv *priv, + const struct netsec_tx_pkt_ctrl *tx_ctrl, + u8 count_frags, const struct netsec_frag_info *info, + struct sk_buff *skb); +int netsec_get_rx_pkt_data(struct netsec_priv *priv, + struct netsec_rx_pkt_info *rxpi, + struct netsec_frag_info *frag, u16 *len, + struct sk_buff **skb); +void netsec_ring_irq_enable(struct netsec_priv *priv, + enum netsec_rings id, u32 i); +void netsec_ring_irq_disable(struct netsec_priv *priv, + enum netsec_rings id, u32 i); +int netsec_alloc_desc_ring(struct netsec_priv *priv, enum netsec_rings id); +void netsec_free_desc_ring(struct netsec_priv *priv, + struct netsec_desc_ring *desc); +int netsec_setup_rx_desc(struct netsec_priv *priv, + struct netsec_desc_ring *desc); +int netsec_netdev_napi_poll(struct napi_struct *napi_p, int budget); + +#endif /* NETSEC_INTERNAL_H */ diff --git a/drivers/net/ethernet/socionext/netsec/netsec_desc_ring_access.c b/drivers/net/ethernet/socionext/netsec/netsec_desc_ring_access.c new file mode 100644 index 0000000..d8e23ca --- /dev/null +++ b/drivers/net/ethernet/socionext/netsec/netsec_desc_ring_access.c @@ -0,0 +1,618 @@ +/** + * drivers/net/ethernet/socionext/netsec/netsec_desc_ring_access.c + * + * Copyright (C) 2011-2014 Fujitsu Semiconductor Limited. + * Copyright (C) 2014 Linaro Ltd Andy Green + * All rights reserved. + * + * This program is free software; you can redistribute it and/or + * modify it under the terms of the GNU General Public License + * as published by the Free Software Foundation; either version 2 + * of the License, or (at your option) any later version. + */ + +#include +#include + +#include "netsec.h" + +static const u32 ads_irq_set[] = { + NETSEC_REG_NRM_TX_INTEN_SET, + NETSEC_REG_NRM_RX_INTEN_SET, +}; + +static const u32 desc_ring_irq_inten_clr_reg_addr[] = { + NETSEC_REG_NRM_TX_INTEN_CLR, + NETSEC_REG_NRM_RX_INTEN_CLR, +}; + +static const u32 int_tmr_reg_addr[] = { + NETSEC_REG_NRM_TX_TXINT_TMR, + NETSEC_REG_NRM_RX_RXINT_TMR, +}; + +static const u32 rx_pkt_cnt_reg_addr[] = { + 0, + NETSEC_REG_NRM_RX_PKTCNT, +}; + +static const u32 tx_pkt_cnt_reg_addr[] = { + NETSEC_REG_NRM_TX_PKTCNT, + 0, +}; + +static const u32 int_pkt_cnt_reg_addr[] = { + NETSEC_REG_NRM_TX_DONE_TXINT_PKTCNT, + NETSEC_REG_NRM_RX_RXINT_PKTCNT, +}; + +static const u32 tx_done_pkt_addr[] = { + NETSEC_REG_NRM_TX_DONE_PKTCNT, + 0, +}; + +static const u32 netsec_desc_mask[] = { + [NETSEC_RING_TX] = NETSEC_GMAC_OMR_REG_ST, + [NETSEC_RING_RX] = NETSEC_GMAC_OMR_REG_SR +}; + +void netsec_ring_irq_enable(struct netsec_priv *priv, + enum netsec_rings id, u32 irqf) +{ + netsec_writel(priv, ads_irq_set[id], irqf); +} + +void netsec_ring_irq_disable(struct netsec_priv *priv, + enum netsec_rings id, u32 irqf) +{ + netsec_writel(priv, desc_ring_irq_inten_clr_reg_addr[id], irqf); +} + +static struct sk_buff *alloc_rx_pkt_buf(struct netsec_priv *priv, + struct netsec_frag_info *info) +{ + struct sk_buff *skb; + + skb = netdev_alloc_skb_ip_align(priv->ndev, info->len); + if (!skb) + return NULL; + + netsec_mark_skb_type(skb, NETSEC_RING_RX); + info->addr = skb->data; + info->dma_addr = dma_map_single(priv->dev, info->addr, info->len, + DMA_FROM_DEVICE); + if (dma_mapping_error(priv->dev, info->dma_addr)) { + dev_kfree_skb(skb); + return NULL; + } + + return skb; +} + +int netsec_alloc_desc_ring(struct netsec_priv *priv, enum netsec_rings id) +{ + struct netsec_desc_ring *desc = &priv->desc_ring[id]; + int ret = 0; + + desc->id = id; + desc->len = sizeof(struct netsec_tx_de); /* rx and tx desc same size */ + + spin_lock_init(&desc->spinlock_desc); + + desc->ring_vaddr = dma_zalloc_coherent(priv->dev, desc->len * DESC_NUM, + &desc->desc_phys, GFP_KERNEL); + if (!desc->ring_vaddr) { + ret = -ENOMEM; + goto err; + } + + desc->frag = kcalloc(DESC_NUM, sizeof(*desc->frag), GFP_KERNEL); + if (!desc->frag) { + ret = -ENOMEM; + goto err; + } + + desc->priv = kcalloc(DESC_NUM, sizeof(struct sk_buff *), GFP_KERNEL); + if (!desc->priv) { + ret = -ENOMEM; + goto err; + } + + return 0; + +err: + netsec_free_desc_ring(priv, desc); + + return ret; +} + +static void netsec_uninit_pkt_desc_ring(struct netsec_priv *priv, + struct netsec_desc_ring *desc) +{ + struct netsec_frag_info *frag; + u32 status; + u16 idx; + + for (idx = 0; idx < DESC_NUM; idx++) { + frag = &desc->frag[idx]; + if (!frag->addr) + continue; + + status = *(u32 *)(desc->ring_vaddr + desc->len * idx); + + dma_unmap_single(priv->dev, frag->dma_addr, frag->len, + skb_is_rx(desc->priv[idx]) ? DMA_FROM_DEVICE : + DMA_TO_DEVICE); + if ((status >> NETSEC_TX_LAST) & 1) + dev_kfree_skb(desc->priv[idx]); + } + + memset(desc->frag, 0, sizeof(struct netsec_frag_info) * DESC_NUM); + memset(desc->priv, 0, sizeof(struct sk_buff *) * DESC_NUM); + memset(desc->ring_vaddr, 0, desc->len * DESC_NUM); +} + +void netsec_free_desc_ring(struct netsec_priv *priv, + struct netsec_desc_ring *desc) +{ + if (desc->ring_vaddr && desc->frag && desc->priv) + netsec_uninit_pkt_desc_ring(priv, desc); + + if (desc->ring_vaddr) { + dma_free_coherent(priv->dev, desc->len * DESC_NUM, + desc->ring_vaddr, desc->desc_phys); + desc->ring_vaddr = NULL; + } + kfree(desc->frag); + desc->frag = NULL; + kfree(desc->priv); + desc->priv = NULL; +} + +static void netsec_set_rx_de(struct netsec_priv *priv, + struct netsec_desc_ring *desc, u16 idx, + const struct netsec_frag_info *info, + struct sk_buff *skb) +{ + struct netsec_rx_de *de = desc->ring_vaddr + desc->len * idx; + u32 attr = 1 << NETSEC_RX_PKT_OWN_FIELD | 1 << NETSEC_RX_PKT_FS_FIELD | + 1 << NETSEC_RX_PKT_LS_FIELD; + + if (idx == DESC_NUM - 1) + attr |= 1 << NETSEC_RX_PKT_LD_FIELD; + + de->data_buf_addr_up = info->dma_addr >> 32; + de->data_buf_addr_lw = info->dma_addr & 0xffffffff; + de->buf_len_info = info->len; + /* desc->attr makes the descriptor live, so it must be physically + * written last after the rest of the descriptor body is already there + */ + wmb(); + de->attr = attr; + + desc->frag[idx].dma_addr = info->dma_addr; + desc->frag[idx].addr = info->addr; + desc->frag[idx].len = info->len; + + desc->priv[idx] = skb; +} + +int netsec_setup_rx_desc(struct netsec_priv *priv, + struct netsec_desc_ring *desc) +{ + struct netsec_frag_info info; + struct sk_buff *skb; + int n; + + info.len = priv->rx_pkt_buf_len; + + for (n = 0; n < DESC_NUM; n++) { + skb = alloc_rx_pkt_buf(priv, &info); + if (!skb) { + netsec_uninit_pkt_desc_ring(priv, desc); + return -ENOMEM; + } + netsec_set_rx_de(priv, desc, n, &info, skb); + } + + return 0; +} + +static void netsec_set_tx_desc_entry(struct netsec_priv *priv, + struct netsec_desc_ring *desc, + const struct netsec_tx_pkt_ctrl *tx_ctrl, + bool first_flag, bool last_flag, + const struct netsec_frag_info *frag, + struct sk_buff *skb) +{ + struct netsec_tx_de tx_desc_entry; + int idx = desc->head; + + memset(&tx_desc_entry, 0, sizeof(struct netsec_tx_de)); + + tx_desc_entry.attr = (1 << NETSEC_TX_SHIFT_OWN_FIELD) | + (desc->id << NETSEC_TX_SHIFT_DRID_FIELD) | + (1 << NETSEC_TX_SHIFT_PT_FIELD) | + (NETSEC_RING_GMAC << NETSEC_TX_SHIFT_TDRID_FIELD) | + (first_flag << NETSEC_TX_SHIFT_FS_FIELD) | + (last_flag << NETSEC_TX_LAST) | + (tx_ctrl->cksum_offload_flag << + NETSEC_TX_SHIFT_CO) | + (tx_ctrl->tcp_seg_offload_flag << + NETSEC_TX_SHIFT_SO) | + (1 << NETSEC_TX_SHIFT_TRS_FIELD); + if (idx == DESC_NUM - 1) + tx_desc_entry.attr |= (1 << NETSEC_TX_SHIFT_LD_FIELD); + + tx_desc_entry.data_buf_addr_up = frag->dma_addr >> 32; + tx_desc_entry.data_buf_addr_lw = frag->dma_addr & 0xffffffff; + tx_desc_entry.buf_len_info = (tx_ctrl->tcp_seg_len << 16) | frag->len; + + memcpy(desc->ring_vaddr + (desc->len * idx), &tx_desc_entry, desc->len); + + desc->frag[idx].dma_addr = frag->dma_addr; + desc->frag[idx].addr = frag->addr; + desc->frag[idx].len = frag->len; + + desc->priv[idx] = skb; +} + +static void netsec_get_rx_de(struct netsec_priv *priv, + struct netsec_desc_ring *desc, u16 idx, + struct netsec_rx_pkt_info *rxpi, + struct netsec_frag_info *frag, u16 *len, + struct sk_buff **skb) +{ + struct netsec_rx_de de; + + memset(&de, 0, sizeof(struct netsec_rx_de)); + memset(rxpi, 0, sizeof(struct netsec_rx_pkt_info)); + memcpy(&de, ((void *)desc->ring_vaddr + desc->len * idx), desc->len); + + dev_dbg(priv->dev, "%08x\n", *(u32 *)&de); + *len = de.buf_len_info >> 16; + + rxpi->is_fragmented = (de.attr >> NETSEC_RX_PKT_FR_FIELD) & 1; + rxpi->err_flag = (de.attr >> NETSEC_RX_PKT_ER_FIELD) & 1; + rxpi->rx_cksum_result = (de.attr >> NETSEC_RX_PKT_CO_FIELD) & 3; + rxpi->err_code = (de.attr >> NETSEC_RX_PKT_ERR_FIELD) & + NETSEC_RX_PKT_ERR_MASK; + memcpy(frag, &desc->frag[idx], sizeof(*frag)); + *skb = desc->priv[idx]; +} + +static void netsec_inc_desc_head_idx(struct netsec_priv *priv, + struct netsec_desc_ring *desc, u16 inc) +{ + u32 sum; + + sum = desc->head + inc; + + if (sum >= DESC_NUM) + sum -= DESC_NUM; + + desc->head = sum; + desc->full = desc->head == desc->tail; +} + +static void netsec_inc_desc_tail_idx(struct netsec_priv *priv, + struct netsec_desc_ring *desc) +{ + u32 sum; + + sum = desc->tail + 1; + + if (sum >= DESC_NUM) + sum -= DESC_NUM; + + desc->tail = sum; + desc->full = false; +} + +static u16 netsec_get_tx_avail_num_sub(struct netsec_priv *priv, + const struct netsec_desc_ring *desc) +{ + if (desc->full) + return 0; + + if (desc->tail > desc->head) + return desc->tail - desc->head; + + return DESC_NUM + desc->tail - desc->head; +} + +static u16 netsec_get_tx_done_num_sub(struct netsec_priv *priv, + struct netsec_desc_ring *desc) +{ + desc->tx_done_num += netsec_readl(priv, tx_done_pkt_addr[desc->id]); + + return desc->tx_done_num; +} + +static int netsec_set_irq_coalesce_param(struct netsec_priv *priv, + enum netsec_rings id) +{ + int max_frames, tmr; + + switch (id) { + case NETSEC_RING_TX: + max_frames = priv->et_coalesce.tx_max_coalesced_frames; + tmr = priv->et_coalesce.tx_coalesce_usecs; + break; + case NETSEC_RING_RX: + max_frames = priv->et_coalesce.rx_max_coalesced_frames; + tmr = priv->et_coalesce.rx_coalesce_usecs; + break; + default: + return -EINVAL; + } + + netsec_writel(priv, int_pkt_cnt_reg_addr[id], max_frames); + netsec_writel(priv, int_tmr_reg_addr[id], ((tmr != 0) << 31) | tmr); + + return 0; +} + +int netsec_start_desc_ring(struct netsec_priv *priv, enum netsec_rings id) +{ + struct netsec_desc_ring *desc = &priv->desc_ring[id]; + int ret = 0; + + spin_lock_bh(&desc->spinlock_desc); + + if (desc->running) { + ret = -EBUSY; + goto err; + } + + switch (desc->id) { + case NETSEC_RING_RX: + netsec_writel(priv, ads_irq_set[id], NETSEC_IRQ_RCV); + break; + case NETSEC_RING_TX: + netsec_writel(priv, ads_irq_set[id], NETSEC_IRQ_EMPTY); + break; + } + + netsec_set_irq_coalesce_param(priv, desc->id); + desc->running = true; + +err: + spin_unlock_bh(&desc->spinlock_desc); + + return ret; +} + +void netsec_stop_desc_ring(struct netsec_priv *priv, enum netsec_rings id) +{ + struct netsec_desc_ring *desc = &priv->desc_ring[id]; + + spin_lock_bh(&desc->spinlock_desc); + if (desc->running) + netsec_writel(priv, desc_ring_irq_inten_clr_reg_addr[id], + NETSEC_IRQ_RCV | NETSEC_IRQ_EMPTY | + NETSEC_IRQ_SND); + + desc->running = false; + spin_unlock_bh(&desc->spinlock_desc); +} + +u16 netsec_get_rx_num(struct netsec_priv *priv) +{ + struct netsec_desc_ring *desc = &priv->desc_ring[NETSEC_RING_RX]; + u32 result; + + spin_lock(&desc->spinlock_desc); + if (desc->running) { + result = netsec_readl(priv, + rx_pkt_cnt_reg_addr[NETSEC_RING_RX]); + desc->rx_num += result; + if (result) + netsec_inc_desc_head_idx(priv, desc, result); + } + spin_unlock(&desc->spinlock_desc); + + return desc->rx_num; +} + +u16 netsec_get_tx_avail_num(struct netsec_priv *priv) +{ + struct netsec_desc_ring *desc = &priv->desc_ring[NETSEC_RING_TX]; + u16 result; + + spin_lock(&desc->spinlock_desc); + + if (!desc->running) { + netif_err(priv, drv, priv->ndev, + "%s: not running tx desc\n", __func__); + result = 0; + goto err; + } + + result = netsec_get_tx_avail_num_sub(priv, desc); + +err: + spin_unlock(&desc->spinlock_desc); + + return result; +} + +int netsec_clean_tx_desc_ring(struct netsec_priv *priv) +{ + struct netsec_desc_ring *desc = &priv->desc_ring[NETSEC_RING_TX]; + unsigned int pkts = 0, bytes = 0; + struct netsec_frag_info *frag; + struct netsec_tx_de *entry; + bool is_last; + + spin_lock(&desc->spinlock_desc); + + netsec_get_tx_done_num_sub(priv, desc); + + while ((desc->tail != desc->head || desc->full) && desc->tx_done_num) { + frag = &desc->frag[desc->tail]; + entry = desc->ring_vaddr + desc->len * desc->tail; + is_last = (entry->attr >> NETSEC_TX_LAST) & 1; + + dma_unmap_single(priv->dev, frag->dma_addr, frag->len, + DMA_TO_DEVICE); + if (is_last) { + pkts++; + bytes += desc->priv[desc->tail]->len; + dev_kfree_skb(desc->priv[desc->tail]); + } + memset(frag, 0, sizeof(*frag)); + netsec_inc_desc_tail_idx(priv, desc); + + if (is_last) + desc->tx_done_num--; + } + + spin_unlock(&desc->spinlock_desc); + + priv->ndev->stats.tx_packets += pkts; + priv->ndev->stats.tx_bytes += bytes; + + netdev_completed_queue(priv->ndev, pkts, bytes); + + return 0; +} + +int netsec_clean_rx_desc_ring(struct netsec_priv *priv) +{ + struct netsec_desc_ring *desc = &priv->desc_ring[NETSEC_RING_RX]; + + spin_lock(&desc->spinlock_desc); + + while (desc->full || (desc->tail != desc->head)) { + netsec_set_rx_de(priv, desc, desc->tail, + &desc->frag[desc->tail], + desc->priv[desc->tail]); + desc->rx_num--; + netsec_inc_desc_tail_idx(priv, desc); + } + + spin_unlock(&desc->spinlock_desc); + + return 0; +} + +int netsec_set_tx_pkt_data(struct netsec_priv *priv, + const struct netsec_tx_pkt_ctrl *tx_ctrl, + u8 count_frags, const struct netsec_frag_info *info, + struct sk_buff *skb) +{ + struct netsec_desc_ring *desc; + u32 sum_len = 0; + unsigned int i; + int ret = 0; + + if (tx_ctrl->tcp_seg_offload_flag && !tx_ctrl->cksum_offload_flag) + return -EINVAL; + + if (tx_ctrl->tcp_seg_offload_flag) { + if (tx_ctrl->tcp_seg_len == 0) + return -EINVAL; + + if (priv->param.use_jumbo_pkt_flag) { + if (tx_ctrl->tcp_seg_len > NETSEC_TCP_JUMBO_SEG_LEN_MAX) + return -EINVAL; + } else { + if (tx_ctrl->tcp_seg_len > NETSEC_TCP_SEG_LEN_MAX) + return -EINVAL; + } + } else { + if (tx_ctrl->tcp_seg_len) + return -EINVAL; + } + + if (!count_frags) + return -ERANGE; + + for (i = 0; i < count_frags; i++) { + if ((info[i].len == 0) || (info[i].len > 0xffff)) { + netif_err(priv, drv, priv->ndev, + "%s: bad info len\n", __func__); + return -EINVAL; + } + sum_len += info[i].len; + } + + if (!tx_ctrl->tcp_seg_offload_flag) { + if (priv->param.use_jumbo_pkt_flag) { + if (sum_len > NETSEC_MAX_TX_JUMBO_PKT_LEN) + return -EINVAL; + } else { + if (sum_len > NETSEC_MAX_TX_PKT_LEN) + return -EINVAL; + } + } + + desc = &priv->desc_ring[NETSEC_RING_TX]; + spin_lock(&desc->spinlock_desc); + + if (!desc->running) { + ret = -ENODEV; + goto end; + } + + smp_rmb(); /* we need to see a consistent view of pending tx count */ + if (count_frags > netsec_get_tx_avail_num_sub(priv, desc)) { + ret = -EBUSY; + goto end; + } + + for (i = 0; i < count_frags; i++) { + netsec_set_tx_desc_entry(priv, desc, tx_ctrl, i == 0, + i == count_frags - 1, &info[i], skb); + netsec_inc_desc_head_idx(priv, desc, 1); + } + + wmb(); /* ensure the descriptor is flushed */ + netsec_writel(priv, tx_pkt_cnt_reg_addr[NETSEC_RING_TX], 1); + +end: + spin_unlock(&desc->spinlock_desc); + + return ret; +} + +int netsec_get_rx_pkt_data(struct netsec_priv *priv, + struct netsec_rx_pkt_info *rxpi, + struct netsec_frag_info *frag, u16 *len, + struct sk_buff **skb) +{ + struct netsec_desc_ring *desc = &priv->desc_ring[NETSEC_RING_RX]; + struct netsec_frag_info info; + struct sk_buff *tmp_skb; + int ret = 0; + + spin_lock(&desc->spinlock_desc); + + if (desc->rx_num == 0) { + dev_err(priv->dev, "%s 0 len rx\n", __func__); + ret = -EINVAL; + goto err; + } + + info.len = priv->rx_pkt_buf_len; + rmb(); /* we need to ensure we only see current data in descriptor */ + tmp_skb = alloc_rx_pkt_buf(priv, &info); + if (!tmp_skb) { + netsec_set_rx_de(priv, desc, desc->tail, + &desc->frag[desc->tail], + desc->priv[desc->tail]); + ret = -ENOMEM; + } else { + netsec_get_rx_de(priv, desc, desc->tail, rxpi, frag, len, skb); + netsec_set_rx_de(priv, desc, desc->tail, &info, tmp_skb); + } + + netsec_inc_desc_tail_idx(priv, desc); + desc->rx_num--; + +err: + spin_unlock(&desc->spinlock_desc); + + return ret; +} diff --git a/drivers/net/ethernet/socionext/netsec/netsec_ethtool.c b/drivers/net/ethernet/socionext/netsec/netsec_ethtool.c new file mode 100644 index 0000000..91f737d --- /dev/null +++ b/drivers/net/ethernet/socionext/netsec/netsec_ethtool.c @@ -0,0 +1,76 @@ +/** + * drivers/net/ethernet/socionext/netsec/netsec_ethtool.c + * + * Copyright (C) 2013-2014 Fujitsu Semiconductor Limited. + * Copyright (C) 2014 Linaro Ltd Andy Green + * All rights reserved. + * + * This program is free software; you can redistribute it and/or + * modify it under the terms of the GNU General Public License + * as published by the Free Software Foundation; either version 2 + * of the License, or (at your option) any later version. + */ + +#include "netsec.h" + +static void netsec_et_get_drvinfo(struct net_device *net_device, + struct ethtool_drvinfo *info) +{ + strlcpy(info->driver, "netsec", sizeof(info->driver)); + strlcpy(info->bus_info, dev_name(net_device->dev.parent), + sizeof(info->bus_info)); +} + +static int netsec_et_get_coalesce(struct net_device *net_device, + struct ethtool_coalesce *et_coalesce) +{ + struct netsec_priv *priv = netdev_priv(net_device); + + *et_coalesce = priv->et_coalesce; + + return 0; +} + +static int netsec_et_set_coalesce(struct net_device *net_device, + struct ethtool_coalesce *et_coalesce) +{ + struct netsec_priv *priv = netdev_priv(net_device); + + if (et_coalesce->rx_max_coalesced_frames > NETSEC_INT_PKTCNT_MAX) + return -EINVAL; + if (et_coalesce->tx_max_coalesced_frames > NETSEC_INT_PKTCNT_MAX) + return -EINVAL; + if (!et_coalesce->rx_max_coalesced_frames) + return -EINVAL; + if (!et_coalesce->tx_max_coalesced_frames) + return -EINVAL; + + priv->et_coalesce = *et_coalesce; + + return 0; +} + +static u32 netsec_et_get_msglevel(struct net_device *dev) +{ + struct netsec_priv *priv = netdev_priv(dev); + + return priv->msg_enable; +} + +static void netsec_et_set_msglevel(struct net_device *dev, u32 datum) +{ + struct netsec_priv *priv = netdev_priv(dev); + + priv->msg_enable = datum; +} + +const struct ethtool_ops netsec_ethtool_ops = { + .get_drvinfo = netsec_et_get_drvinfo, + .get_link_ksettings = phy_ethtool_get_link_ksettings, + .set_link_ksettings = phy_ethtool_set_link_ksettings, + .get_link = ethtool_op_get_link, + .get_coalesce = netsec_et_get_coalesce, + .set_coalesce = netsec_et_set_coalesce, + .get_msglevel = netsec_et_get_msglevel, + .set_msglevel = netsec_et_set_msglevel, +}; diff --git a/drivers/net/ethernet/socionext/netsec/netsec_gmac_access.c b/drivers/net/ethernet/socionext/netsec/netsec_gmac_access.c new file mode 100644 index 0000000..415c7b4 --- /dev/null +++ b/drivers/net/ethernet/socionext/netsec/netsec_gmac_access.c @@ -0,0 +1,329 @@ +/** + * drivers/net/ethernet/socionext/netsec/netsec_gmac_access.c + * + * Copyright (C) 2011-2014 Fujitsu Semiconductor Limited. + * Copyright (C) 2014 Linaro Ltd Andy Green + * All rights reserved. + * + * This program is free software; you can redistribute it and/or + * modify it under the terms of the GNU General Public License + * as published by the Free Software Foundation; either version 2 + * of the License, or (at your option) any later version. + */ +#include "netsec.h" + +#define TIMEOUT_SPINS_MAC 1000 +#define TIMEOUT_SECONDARY_MS_MAC 100 + +static u32 netsec_clk_type(u32 freq) +{ + if (freq < 35 * NETSEC_CLK_MHZ) + return NETSEC_GMAC_GAR_REG_CR_25_35_MHZ; + if (freq < 60 * NETSEC_CLK_MHZ) + return NETSEC_GMAC_GAR_REG_CR_35_60_MHZ; + if (freq < 100 * NETSEC_CLK_MHZ) + return NETSEC_GMAC_GAR_REG_CR_60_100_MHZ; + if (freq < 150 * NETSEC_CLK_MHZ) + return NETSEC_GMAC_GAR_REG_CR_100_150_MHZ; + if (freq < 250 * NETSEC_CLK_MHZ) + return NETSEC_GMAC_GAR_REG_CR_150_250_MHZ; + + return NETSEC_GMAC_GAR_REG_CR_250_300_MHZ; +} + +static int netsec_wait_while_busy(struct netsec_priv *priv, u32 addr, u32 mask) +{ + u32 timeout = TIMEOUT_SPINS_MAC; + + while (--timeout && netsec_readl(priv, addr) & mask) + ; + if (timeout) + return 0; + + timeout = TIMEOUT_SECONDARY_MS_MAC; + while (--timeout && netsec_readl(priv, addr) & mask) + usleep_range(1000, 2000); + + if (timeout) + return 0; + + netdev_WARN(priv->ndev, "%s: timeout\n", __func__); + + return -ETIMEDOUT; +} + +static int netsec_mac_write(struct netsec_priv *priv, u32 addr, u32 value) +{ + netsec_writel(priv, MAC_REG_DATA, value); + netsec_writel(priv, MAC_REG_CMD, addr | NETSEC_GMAC_CMD_ST_WRITE); + return netsec_wait_while_busy(priv, + MAC_REG_CMD, NETSEC_GMAC_CMD_ST_BUSY); +} + +static int netsec_mac_read(struct netsec_priv *priv, u32 addr, u32 *read) +{ + int ret; + + netsec_writel(priv, MAC_REG_CMD, addr | NETSEC_GMAC_CMD_ST_READ); + ret = netsec_wait_while_busy(priv, + MAC_REG_CMD, NETSEC_GMAC_CMD_ST_BUSY); + if (ret) + return ret; + + *read = netsec_readl(priv, MAC_REG_DATA); + + return 0; +} + +static int netsec_mac_wait_while_busy(struct netsec_priv *priv, + u32 addr, u32 mask) +{ + u32 timeout = TIMEOUT_SPINS_MAC; + int ret, data; + + do { + ret = netsec_mac_read(priv, addr, &data); + if (ret) + break; + } while (--timeout && (data & mask)); + + if (timeout) + return 0; + + timeout = TIMEOUT_SECONDARY_MS_MAC; + do { + usleep_range(1000, 2000); + + ret = netsec_mac_read(priv, addr, &data); + if (ret) + break; + } while (--timeout && (data & mask)); + + if (timeout && !ret) + return 0; + + netdev_WARN(priv->ndev, "%s: timeout\n", __func__); + + return -ETIMEDOUT; +} + +static int netsec_mac_update_to_phy_state(struct netsec_priv *priv) +{ + struct phy_device *phydev = priv->ndev->phydev; + u32 value = 0; + + value = phydev->duplex ? NETSEC_GMAC_MCR_REG_FULL_DUPLEX_COMMON : + NETSEC_GMAC_MCR_REG_HALF_DUPLEX_COMMON; + + if (phydev->speed != SPEED_1000) + value |= NETSEC_MCR_PS; + + if ((priv->phy_interface != PHY_INTERFACE_MODE_GMII) && + (phydev->speed == SPEED_100)) + value |= NETSEC_GMAC_MCR_REG_FES; + + value |= NETSEC_GMAC_MCR_REG_CST | NETSEC_GMAC_MCR_REG_JE; + + if (priv->phy_interface == PHY_INTERFACE_MODE_RGMII) + value |= NETSEC_GMAC_MCR_REG_IBN; + + if (netsec_mac_write(priv, GMAC_REG_MCR, value)) + return -ETIMEDOUT; + + priv->actual_link_speed = phydev->speed; + priv->actual_duplex = phydev->duplex; + netif_info(priv, drv, priv->ndev, "%s: %uMbps, duplex:%d\n", + __func__, phydev->speed, phydev->duplex); + + return 0; +} + +/* NB netsec_start_gmac() only called from adjust_link */ + +int netsec_start_gmac(struct netsec_priv *priv) +{ + struct phy_device *phydev = priv->ndev->phydev; + u32 value = 0; + int ret; + + if (priv->desc_ring[NETSEC_RING_TX].running && + priv->desc_ring[NETSEC_RING_RX].running) + return 0; + + if (!priv->desc_ring[NETSEC_RING_RX].running && + !priv->desc_ring[NETSEC_RING_TX].running) { + if (phydev->speed != SPEED_1000) + value = (NETSEC_GMAC_MCR_REG_CST | + NETSEC_GMAC_MCR_REG_HALF_DUPLEX_COMMON); + + if (netsec_mac_write(priv, GMAC_REG_MCR, value)) + return -ETIMEDOUT; + if (netsec_mac_write(priv, GMAC_REG_BMR, + NETSEC_GMAC_BMR_REG_RESET)) + return -ETIMEDOUT; + + /* Wait soft reset */ + usleep_range(1000, 5000); + + ret = netsec_mac_read(priv, GMAC_REG_BMR, &value); + if (ret) + return ret; + if (value & NETSEC_GMAC_BMR_REG_SWR) + return -EAGAIN; + + netsec_writel(priv, MAC_REG_DESC_SOFT_RST, 1); + if (netsec_wait_while_busy(priv, MAC_REG_DESC_SOFT_RST, 1)) + return -ETIMEDOUT; + + netsec_writel(priv, MAC_REG_DESC_INIT, 1); + if (netsec_wait_while_busy(priv, MAC_REG_DESC_INIT, 1)) + return -ETIMEDOUT; + + if (netsec_mac_write(priv, GMAC_REG_BMR, + NETSEC_GMAC_BMR_REG_COMMON)) + return -ETIMEDOUT; + if (netsec_mac_write(priv, GMAC_REG_RDLAR, priv->rdlar_pa)) + return -ETIMEDOUT; + if (netsec_mac_write(priv, GMAC_REG_TDLAR, priv->tdlar_pa)) + return -ETIMEDOUT; + if (netsec_mac_write(priv, GMAC_REG_MFFR, 0x80000001)) + return -ETIMEDOUT; + + ret = netsec_mac_update_to_phy_state(priv); + if (ret) + return ret; + + if (priv->mac_mode.flow_ctrl_enable_flag) { + netsec_writel(priv, MAC_REG_FLOW_TH, + (priv->mac_mode.flow_stop_th << 16) | + priv->mac_mode.flow_start_th); + if (netsec_mac_write(priv, GMAC_REG_FCR, + (priv->mac_mode.pause_time << 16) | + NETSEC_FCR_RFE | NETSEC_FCR_TFE)) + return -ETIMEDOUT; + } + } + + ret = netsec_mac_read(priv, GMAC_REG_OMR, &value); + if (ret) + return ret; + + if (!priv->desc_ring[NETSEC_RING_RX].running) { + value |= NETSEC_GMAC_OMR_REG_SR; + netsec_start_desc_ring(priv, NETSEC_RING_RX); + } + if (!priv->desc_ring[NETSEC_RING_TX].running) { + value |= NETSEC_GMAC_OMR_REG_ST; + netsec_start_desc_ring(priv, NETSEC_RING_TX); + } + + if (netsec_mac_write(priv, GMAC_REG_OMR, value)) + return -ETIMEDOUT; + + netsec_writel(priv, NETSEC_REG_INTEN_SET, + NETSEC_IRQ_TX | NETSEC_IRQ_RX); + + return 0; +} + +int netsec_stop_gmac(struct netsec_priv *priv) +{ + u32 value; + int ret; + + ret = netsec_mac_read(priv, GMAC_REG_OMR, &value); + if (ret) + return ret; + + if (priv->desc_ring[NETSEC_RING_RX].running) { + value &= ~NETSEC_GMAC_OMR_REG_SR; + netsec_stop_desc_ring(priv, NETSEC_RING_RX); + } + if (priv->desc_ring[NETSEC_RING_TX].running) { + value &= ~NETSEC_GMAC_OMR_REG_ST; + netsec_stop_desc_ring(priv, NETSEC_RING_TX); + } + + priv->actual_link_speed = 0; + priv->actual_duplex = false; + + return netsec_mac_write(priv, GMAC_REG_OMR, value); +} + +static int netsec_phy_write(struct mii_bus *bus, + int phy_addr, int reg, u16 val) +{ + struct netsec_priv *priv = bus->priv; + + if (netsec_mac_write(priv, GMAC_REG_GDR, val)) + return -ETIMEDOUT; + if (netsec_mac_write(priv, GMAC_REG_GAR, + phy_addr << NETSEC_GMAC_GAR_REG_SHIFT_PA | + reg << NETSEC_GMAC_GAR_REG_SHIFT_GR | + NETSEC_GMAC_GAR_REG_GW | NETSEC_GMAC_GAR_REG_GB) | + (netsec_clk_type(priv->freq) << + GMAC_REG_SHIFT_CR_GAR)) + return -ETIMEDOUT; + + return netsec_mac_wait_while_busy(priv, GMAC_REG_GAR, + NETSEC_GMAC_GAR_REG_GB); +} + +static int netsec_phy_read(struct mii_bus *bus, int phy_addr, int reg_addr) +{ + struct netsec_priv *priv = bus->priv; + u32 data; + int ret; + + if (netsec_mac_write(priv, GMAC_REG_GAR, NETSEC_GMAC_GAR_REG_GB | + phy_addr << NETSEC_GMAC_GAR_REG_SHIFT_PA | + reg_addr << NETSEC_GMAC_GAR_REG_SHIFT_GR | + (netsec_clk_type(priv->freq) << + GMAC_REG_SHIFT_CR_GAR))) + return -ETIMEDOUT; + + ret = netsec_mac_wait_while_busy(priv, GMAC_REG_GAR, + NETSEC_GMAC_GAR_REG_GB); + if (ret) + return 0; + + ret = netsec_mac_read(priv, GMAC_REG_GDR, &data); + if (ret) + return ret; + + return data; +} + +int netsec_mii_register(struct netsec_priv *priv) +{ + struct mii_bus *bus = mdiobus_alloc(); + struct resource res; + int ret; + + if (!bus) + return -ENOMEM; + + of_address_to_resource(priv->dev->of_node, 0, &res); + snprintf(bus->id, MII_BUS_ID_SIZE, "%s", priv->dev->of_node->full_name); + bus->priv = priv; + bus->name = "SNI NETSEC MDIO"; + bus->read = netsec_phy_read; + bus->write = netsec_phy_write; + bus->parent = priv->dev; + priv->mii_bus = bus; + + ret = of_mdiobus_register(bus, priv->dev->of_node); + if (ret) { + mdiobus_free(bus); + return ret; + } + + return 0; +} + +void netsec_mii_unregister(struct netsec_priv *priv) +{ + mdiobus_unregister(priv->mii_bus); + mdiobus_free(priv->mii_bus); + priv->mii_bus = NULL; +} diff --git a/drivers/net/ethernet/socionext/netsec/netsec_netdev.c b/drivers/net/ethernet/socionext/netsec/netsec_netdev.c new file mode 100644 index 0000000..ff418c1 --- /dev/null +++ b/drivers/net/ethernet/socionext/netsec/netsec_netdev.c @@ -0,0 +1,558 @@ +/** + * drivers/net/ethernet/socionext/netsec/netsec_netdev.c + * + * Copyright (C) 2013-2014 Fujitsu Semiconductor Limited. + * Copyright (C) 2014 Linaro Ltd Andy Green + * All rights reserved. + * + * This program is free software; you can redistribute it and/or + * modify it under the terms of the GNU General Public License + * as published by the Free Software Foundation; either version 2 + * of the License, or (at your option) any later version. + */ + +#include +#include +#include +#include +#include +#include + +#include "netsec.h" + +#define WAIT_FW_RDY_TIMEOUT 50 + +static const u32 desc_ring_irq_status_reg_addr[] = { + NETSEC_REG_NRM_TX_STATUS, + NETSEC_REG_NRM_RX_STATUS, +}; + +static const u32 desc_ads[] = { + NETSEC_REG_NRM_TX_CONFIG, + NETSEC_REG_NRM_RX_CONFIG, +}; + +static const u32 netsec_desc_start_reg_addr_up[] = { + NETSEC_REG_NRM_TX_DESC_START_UP, + NETSEC_REG_NRM_RX_DESC_START_UP, +}; + +static const u32 netsec_desc_start_reg_addr_lw[] = { + NETSEC_REG_NRM_TX_DESC_START_LW, + NETSEC_REG_NRM_RX_DESC_START_LW, +}; + +static int netsec_wait_for_ring_config_ready(struct netsec_priv *priv, int ring) +{ + int timeout = WAIT_FW_RDY_TIMEOUT; + + while (--timeout && (netsec_readl(priv, desc_ads[ring]) & + NETSEC_REG_DESC_RING_CONFIG_CFG_UP)) + usleep_range(1000, 2000); + + if (!timeout) { + netif_err(priv, hw, priv->ndev, + "%s: timeout\n", __func__); + return -ETIMEDOUT; + } + + return 0; +} + +static u32 netsec_calc_pkt_ctrl_reg_param(const struct netsec_pkt_ctrlaram + *pkt_ctrlaram_p) +{ + u32 param = NETSEC_PKT_CTRL_REG_MODE_NRM; + + if (pkt_ctrlaram_p->log_chksum_er_flag) + param |= NETSEC_PKT_CTRL_REG_LOG_CHKSUM_ER; + + if (pkt_ctrlaram_p->log_hd_imcomplete_flag) + param |= NETSEC_PKT_CTRL_REG_LOG_HD_INCOMPLETE; + + if (pkt_ctrlaram_p->log_hd_er_flag) + param |= NETSEC_PKT_CTRL_REG_LOG_HD_ER; + + return param; +} + +static int netsec_configure_normal_mode(struct netsec_priv *priv) +{ + int ret = 0; + u32 value; + + /* save scb set value */ + priv->scb_set_normal_tx_paddr = (phys_addr_t)netsec_readl(priv, + netsec_desc_start_reg_addr_up[NETSEC_RING_TX]) << 32; + priv->scb_set_normal_tx_paddr |= (phys_addr_t)netsec_readl(priv, + netsec_desc_start_reg_addr_lw[NETSEC_RING_TX]); + + /* set desc_start addr */ + netsec_writel(priv, netsec_desc_start_reg_addr_up[NETSEC_RING_RX], + priv->desc_ring[NETSEC_RING_RX].desc_phys >> 32); + netsec_writel(priv, netsec_desc_start_reg_addr_lw[NETSEC_RING_RX], + priv->desc_ring[NETSEC_RING_RX].desc_phys & 0xffffffff); + + netsec_writel(priv, netsec_desc_start_reg_addr_up[NETSEC_RING_TX], + priv->desc_ring[NETSEC_RING_TX].desc_phys >> 32); + netsec_writel(priv, netsec_desc_start_reg_addr_lw[NETSEC_RING_TX], + priv->desc_ring[NETSEC_RING_TX].desc_phys & 0xffffffff); + + /* set normal tx desc ring config */ + value = (cpu_to_le32(1) == 1) << NETSEC_REG_DESC_ENDIAN | + NETSEC_REG_DESC_RING_CONFIG_CFG_UP | + NETSEC_REG_DESC_RING_CONFIG_CH_RST; + netsec_writel(priv, desc_ads[NETSEC_RING_TX], value); + + value = (cpu_to_le32(1) == 1) << NETSEC_REG_DESC_ENDIAN | + NETSEC_REG_DESC_RING_CONFIG_CFG_UP | + NETSEC_REG_DESC_RING_CONFIG_CH_RST; + netsec_writel(priv, desc_ads[NETSEC_RING_RX], value); + + if (netsec_wait_for_ring_config_ready(priv, NETSEC_RING_TX) || + netsec_wait_for_ring_config_ready(priv, NETSEC_RING_RX)) + return -ETIMEDOUT; + + return ret; +} + +static int netsec_change_mode_to_normal(struct netsec_priv *priv) +{ + u32 value; + + priv->scb_pkt_ctrl_reg = netsec_readl(priv, NETSEC_REG_PKT_CTRL); + + value = netsec_calc_pkt_ctrl_reg_param(&priv->param.pkt_ctrlaram); + + if (priv->param.use_jumbo_pkt_flag) + value |= NETSEC_PKT_CTRL_REG_EN_JUMBO; + + value |= NETSEC_PKT_CTRL_REG_MODE_NRM; + + /* change to normal mode */ + netsec_writel(priv, NETSEC_REG_DMA_MH_CTRL, MH_CTRL__MODE_TRANS); + netsec_writel(priv, NETSEC_REG_PKT_CTRL, value); + + /* Wait Change mode Complete */ + usleep_range(2000, 10000); + + return 0; +} + +static int netsec_change_mode_to_taiki(struct netsec_priv *priv) +{ + int ret = 0; + u32 value; + + netsec_writel(priv, netsec_desc_start_reg_addr_up[NETSEC_RING_TX], + priv->scb_set_normal_tx_paddr >> 32); + netsec_writel(priv, netsec_desc_start_reg_addr_lw[NETSEC_RING_TX], + priv->scb_set_normal_tx_paddr & 0xffffffff); + + value = NETSEC_REG_DESC_RING_CONFIG_CFG_UP | + NETSEC_REG_DESC_RING_CONFIG_CH_RST; + + netsec_writel(priv, desc_ads[NETSEC_RING_TX], value); + + if (netsec_wait_for_ring_config_ready(priv, NETSEC_RING_TX)) + return -ETIMEDOUT; + + netsec_writel(priv, NETSEC_REG_DMA_MH_CTRL, MH_CTRL__MODE_TRANS); + netsec_writel(priv, NETSEC_REG_PKT_CTRL, priv->scb_pkt_ctrl_reg); + + /* Wait Change mode Complete */ + usleep_range(2000, 10000); + + return ret; +} + +static int netsec_clear_modechange_irq(struct netsec_priv *priv, u32 value) +{ + netsec_writel(priv, NETSEC_REG_MODE_TRANS_COMP_STATUS, + (value & (NETSEC_MODE_TRANS_COMP_IRQ_N2T | + NETSEC_MODE_TRANS_COMP_IRQ_T2N))); + return 0; +} + +static int netsec_hw_configure_to_normal(struct netsec_priv *priv) +{ + int err; + + err = netsec_configure_normal_mode(priv); + if (err) { + netif_err(priv, drv, priv->ndev, + "%s: normal conf fail\n", __func__); + return err; + } + err = netsec_change_mode_to_normal(priv); + if (err) { + netif_err(priv, drv, priv->ndev, + "%s: normal set fail\n", __func__); + return err; + } + + return err; +} + +static int netsec_hw_configure_to_taiki(struct netsec_priv *priv) +{ + int ret; + + ret = netsec_change_mode_to_taiki(priv); + if (ret) { + netif_err(priv, drv, priv->ndev, + "%s: taiki set fail\n", __func__); + return ret; + } + + /* Clear mode change complete IRQ */ + ret = netsec_clear_modechange_irq(priv, NETSEC_MODE_TRANS_COMP_IRQ_T2N + | NETSEC_MODE_TRANS_COMP_IRQ_N2T); + + if (ret) + netif_err(priv, drv, priv->ndev, + "%s: clear mode fail\n", __func__); + + return ret; +} + +static void netsec_ring_irq_clr(struct netsec_priv *priv, + unsigned int id, u32 value) +{ + netsec_writel(priv, desc_ring_irq_status_reg_addr[id], + value & (NETSEC_IRQ_EMPTY | NETSEC_IRQ_ERR)); +} + +static void netsec_napi_tx_processing(struct napi_struct *napi_p) +{ + struct netsec_priv *priv = container_of(napi_p, + struct netsec_priv, napi); + + netsec_ring_irq_clr(priv, NETSEC_RING_TX, NETSEC_IRQ_EMPTY); + netsec_clean_tx_desc_ring(priv); + + if (netif_queue_stopped(priv->ndev) && + netsec_get_tx_avail_num(priv) >= NETSEC_NETDEV_TX_PKT_SCAT_NUM_MAX) + netif_wake_queue(priv->ndev); +} + +int netsec_netdev_napi_poll(struct napi_struct *napi_p, int budget) +{ + struct netsec_priv *priv = container_of(napi_p, + struct netsec_priv, napi); + struct net_device *ndev = priv->ndev; + struct netsec_rx_pkt_info rx_info; + int ret, done = 0, rx_num = 0; + struct netsec_frag_info frag; + struct sk_buff *skb; + u16 len; + + netsec_napi_tx_processing(napi_p); + + while (done < budget) { + if (!rx_num) { + rx_num = netsec_get_rx_num(priv); + if (!rx_num) + break; + } + done++; + rx_num--; + ret = netsec_get_rx_pkt_data(priv, &rx_info, &frag, &len, &skb); + if (unlikely(ret == -ENOMEM)) { + netif_err(priv, drv, priv->ndev, + "%s: rx fail %d\n", __func__, ret); + ndev->stats.rx_dropped++; + continue; + } + dma_unmap_single(priv->dev, frag.dma_addr, frag.len, + DMA_FROM_DEVICE); + skb_put(skb, len); + skb->protocol = eth_type_trans(skb, priv->ndev); + + if (priv->rx_cksum_offload_flag && + rx_info.rx_cksum_result == NETSEC_RX_CKSUM_OK) + skb->ip_summed = CHECKSUM_UNNECESSARY; + + napi_gro_receive(napi_p, skb); + + ndev->stats.rx_packets++; + ndev->stats.rx_bytes += len; + } + + if (done == budget) + return budget; + + napi_complete(napi_p); + netsec_writel(priv, NETSEC_REG_INTEN_SET, + NETSEC_IRQ_TX | NETSEC_IRQ_RX); + + return done; +} + +static netdev_tx_t netsec_netdev_start_xmit(struct sk_buff *skb, + struct net_device *ndev) +{ + struct netsec_priv *priv = netdev_priv(ndev); + struct netsec_tx_pkt_ctrl tx_ctrl; + u16 pend_tx, tso_seg_len = 0; + skb_frag_t *frag; + int count_frags; + int ret, i; + + memset(&tx_ctrl, 0, sizeof(struct netsec_tx_pkt_ctrl)); + + netsec_ring_irq_clr(priv, NETSEC_RING_TX, NETSEC_IRQ_EMPTY); + + count_frags = skb_shinfo(skb)->nr_frags + 1; + + if (skb->ip_summed == CHECKSUM_PARTIAL) + tx_ctrl.cksum_offload_flag = true; + + if (skb_is_gso(skb)) + tso_seg_len = skb_shinfo(skb)->gso_size; + + if (tso_seg_len > 0) { + if (skb->protocol == htons(ETH_P_IP)) { + ip_hdr(skb)->tot_len = 0; + tcp_hdr(skb)->check = + ~tcp_v4_check(0, ip_hdr(skb)->saddr, + ip_hdr(skb)->daddr, 0); + } else { + ipv6_hdr(skb)->payload_len = 0; + tcp_hdr(skb)->check = + ~csum_ipv6_magic(&ipv6_hdr(skb)->saddr, + &ipv6_hdr(skb)->daddr, + 0, IPPROTO_TCP, 0); + } + + tx_ctrl.tcp_seg_offload_flag = true; + tx_ctrl.tcp_seg_len = tso_seg_len; + } + + priv->tx_info[0].dma_addr = dma_map_single(priv->dev, skb->data, + skb_headlen(skb), + DMA_TO_DEVICE); + if (dma_mapping_error(priv->dev, priv->tx_info[0].dma_addr)) { + netif_err(priv, drv, priv->ndev, + "%s: DMA mapping failed\n", __func__); + return NETDEV_TX_OK; + } + priv->tx_info[0].addr = skb->data; + priv->tx_info[0].len = skb_headlen(skb); + + for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { + frag = &skb_shinfo(skb)->frags[i]; + priv->tx_info[i + 1].dma_addr = + skb_frag_dma_map(priv->dev, frag, 0, + skb_frag_size(frag), DMA_TO_DEVICE); + priv->tx_info[i + 1].addr = skb_frag_address(frag); + priv->tx_info[i + 1].len = frag->size; + } + + netsec_mark_skb_type(skb, NETSEC_RING_TX); + + ret = netsec_set_tx_pkt_data(priv, &tx_ctrl, count_frags, + priv->tx_info, skb); + if (ret) { + netif_info(priv, drv, priv->ndev, + "set tx pkt failed %d\n", ret); + for (i = 0; i < count_frags; i++) + dma_unmap_single(priv->dev, priv->tx_info[i].dma_addr, + priv->tx_info[i].len, DMA_TO_DEVICE); + ndev->stats.tx_dropped++; + + return NETDEV_TX_OK; + } + + netdev_sent_queue(priv->ndev, skb->len); + + spin_lock(&priv->tx_queue_lock); + pend_tx = netsec_get_tx_avail_num(priv); + + if (pend_tx < NETSEC_NETDEV_TX_PKT_SCAT_NUM_MAX) { + netsec_ring_irq_enable(priv, NETSEC_RING_TX, NETSEC_IRQ_EMPTY); + netif_stop_queue(ndev); + goto err; + } + if (pend_tx <= DESC_NUM - 2) { + netsec_ring_irq_enable(priv, NETSEC_RING_TX, NETSEC_IRQ_EMPTY); + goto err; + } + netsec_ring_irq_disable(priv, NETSEC_RING_TX, NETSEC_IRQ_EMPTY); + +err: + spin_unlock(&priv->tx_queue_lock); + + return NETDEV_TX_OK; +} + +static int netsec_netdev_set_features(struct net_device *ndev, + netdev_features_t features) +{ + struct netsec_priv *priv = netdev_priv(ndev); + + priv->rx_cksum_offload_flag = !!(features & NETIF_F_RXCSUM); + + return 0; +} + +static void netsec_phy_adjust_link(struct net_device *ndev) +{ + struct netsec_priv *priv = netdev_priv(ndev); + + if (priv->actual_link_speed == ndev->phydev->speed && + priv->actual_duplex == ndev->phydev->duplex) + return; + + netsec_stop_gmac(priv); + netsec_start_gmac(priv); +} + +static irqreturn_t netsec_irq_handler(int irq, void *dev_id) +{ + struct netsec_priv *priv = dev_id; + u32 status = netsec_readl(priv, NETSEC_REG_TOP_STATUS) & + netsec_readl(priv, NETSEC_REG_TOP_INTEN); + + if (!status) + return IRQ_NONE; + + if (status & (NETSEC_IRQ_TX | NETSEC_IRQ_RX)) { + netsec_writel(priv, NETSEC_REG_INTEN_CLR, + status & (NETSEC_IRQ_TX | NETSEC_IRQ_RX)); + napi_schedule(&priv->napi); + } + + return IRQ_HANDLED; +} + +static int netsec_netdev_open(struct net_device *ndev) +{ + struct netsec_priv *priv = netdev_priv(ndev); + struct phy_device *phydev = NULL; + u32 scb_irq_temp; + int ret, n; + + scb_irq_temp = netsec_readl(priv, NETSEC_REG_TOP_INTEN); + + for (n = 0; n <= NETSEC_RING_MAX; n++) { + ret = netsec_alloc_desc_ring(priv, n); + if (ret) { + netif_err(priv, probe, priv->ndev, + "%s: alloc ring failed\n", __func__); + goto err; + } + } + + ret = netsec_setup_rx_desc(priv, &priv->desc_ring[NETSEC_RING_RX]); + if (ret) { + netif_err(priv, probe, priv->ndev, + "%s: fail setup ring\n", __func__); + goto err1; + } + + pm_runtime_get_sync(priv->dev); + + netsec_writel(priv, NETSEC_REG_INTEN_CLR, scb_irq_temp); + + ret = netsec_hw_configure_to_normal(priv); + if (ret) { + netif_err(priv, probe, priv->ndev, + "%s: normal fail %d\n", __func__, ret); + goto err1; + } + + ret = request_irq(priv->ndev->irq, netsec_irq_handler, + IRQF_SHARED, "netsec", priv); + if (ret) { + netif_err(priv, drv, priv->ndev, "request_irq failed\n"); + goto err1; + } + priv->irq_registered = true; + + ret = netsec_clean_rx_desc_ring(priv); + if (ret) { + netif_err(priv, drv, priv->ndev, + "%s: clean rx desc fail\n", __func__); + goto err2; + } + + ret = netsec_clean_tx_desc_ring(priv); + if (ret) { + netif_err(priv, drv, priv->ndev, + "%s: clean tx desc fail\n", __func__); + goto err2; + } + + netsec_ring_irq_clr(priv, NETSEC_RING_TX, NETSEC_IRQ_EMPTY); + + phydev = of_phy_connect(priv->ndev, priv->phy_np, + &netsec_phy_adjust_link, 0, + priv->phy_interface); + if (!phydev) { + netif_err(priv, link, priv->ndev, "missing PHY\n"); + goto err2; + } + + phy_start_aneg(phydev); + + netsec_ring_irq_disable(priv, NETSEC_RING_TX, NETSEC_IRQ_EMPTY); + + netsec_start_gmac(priv); + napi_enable(&priv->napi); + netif_start_queue(ndev); + + netsec_writel(priv, NETSEC_REG_INTEN_SET, + NETSEC_IRQ_TX | NETSEC_IRQ_RX); + + return 0; + +err2: + pm_runtime_put_sync(priv->dev); + free_irq(priv->ndev->irq, priv); + priv->irq_registered = false; +err1: + for (n = 0; n <= NETSEC_RING_MAX; n++) + netsec_free_desc_ring(priv, &priv->desc_ring[n]); +err: + netsec_writel(priv, NETSEC_REG_INTEN_SET, scb_irq_temp); + + pm_runtime_put_sync(priv->dev); + + return ret; +} + +static int netsec_netdev_stop(struct net_device *ndev) +{ + struct netsec_priv *priv = netdev_priv(ndev); + int n; + + phy_stop(ndev->phydev); + phy_disconnect(ndev->phydev); + + netif_stop_queue(priv->ndev); + napi_disable(&priv->napi); + + netsec_writel(priv, NETSEC_REG_INTEN_CLR, ~0); + netsec_stop_gmac(priv); + WARN_ON(netsec_hw_configure_to_taiki(priv)); + + pm_runtime_put_sync(priv->dev); + + for (n = 0; n <= NETSEC_RING_MAX; n++) + netsec_free_desc_ring(priv, &priv->desc_ring[n]); + + free_irq(priv->ndev->irq, priv); + priv->irq_registered = false; + + return 0; +} + +const struct net_device_ops netsec_netdev_ops = { + .ndo_open = netsec_netdev_open, + .ndo_stop = netsec_netdev_stop, + .ndo_start_xmit = netsec_netdev_start_xmit, + .ndo_set_features = netsec_netdev_set_features, + .ndo_set_mac_address = eth_mac_addr, + .ndo_validate_addr = eth_validate_addr, +}; diff --git a/drivers/net/ethernet/socionext/netsec/netsec_platform.c b/drivers/net/ethernet/socionext/netsec/netsec_platform.c new file mode 100644 index 0000000..d5c19f7 --- /dev/null +++ b/drivers/net/ethernet/socionext/netsec/netsec_platform.c @@ -0,0 +1,330 @@ +/** + * drivers/net/ethernet/socionext/netsec/netsec_platform.c + * + * Copyright (C) 2013-2014 Fujitsu Semiconductor Limited. + * Copyright (C) 2014 Linaro Ltd Andy Green + * All rights reserved. + * + * This program is free software; you can redistribute it and/or + * modify it under the terms of the GNU General Public License + * as published by the Free Software Foundation; either version 2 + * of the License, or (at your option) any later version. + */ + +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include + +#include "netsec.h" + +#define NETSEC_F_NETSEC_VER_MAJOR_NUM(x) (x & 0xffff0000) + +static int napi_weight = 64; +unsigned short pause_time = 256; + +static int netsec_probe(struct platform_device *pdev) +{ + struct net_device *ndev; + struct netsec_priv *priv; + struct resource *res; + const void *mac; + bool use_jumbo; + u32 hw_ver; + int err; + int ret; + + ndev = alloc_etherdev(sizeof(*priv)); + if (!ndev) + return -ENOMEM; + + priv = netdev_priv(ndev); + priv->ndev = ndev; + SET_NETDEV_DEV(ndev, &pdev->dev); + platform_set_drvdata(pdev, priv); + priv->dev = &pdev->dev; + + priv->msg_enable = NETIF_MSG_TX_ERR | NETIF_MSG_HW | NETIF_MSG_DRV | + NETIF_MSG_LINK | NETIF_MSG_PROBE; + + mac = of_get_mac_address(pdev->dev.of_node); + if (mac) + ether_addr_copy(ndev->dev_addr, mac); + + if (!is_valid_ether_addr(ndev->dev_addr)) { + eth_hw_addr_random(ndev); + dev_warn(&pdev->dev, "No MAC address found, using random\n"); + } + + priv->phy_np = of_parse_phandle(pdev->dev.of_node, "phy-handle", 0); + if (!priv->phy_np) { + netif_err(priv, probe, ndev, "missing phy in DT\n"); + goto err1; + } + + priv->phy_interface = of_get_phy_mode(pdev->dev.of_node); + if (priv->phy_interface < 0) { + netif_err(priv, probe, ndev, + "%s: bad phy-if\n", __func__); + goto err1; + } + + priv->ioaddr = of_iomap(priv->dev->of_node, 0); + if (!priv->ioaddr) { + netif_err(priv, probe, ndev, "of_iomap() failed\n"); + err = -EINVAL; + goto err1; + } + + res = platform_get_resource(pdev, IORESOURCE_MEM, 1); + if (!res) { + netif_err(priv, probe, ndev, + "Missing rdlar resource\n"); + goto err1; + } + priv->rdlar_pa = res->start; + + res = platform_get_resource(pdev, IORESOURCE_MEM, 2); + if (!res) { + netif_err(priv, probe, ndev, + "Missing tdlar resource\n"); + goto err1; + } + priv->tdlar_pa = res->start; + + res = platform_get_resource(pdev, IORESOURCE_IRQ, 0); + if (!res) { + netif_err(priv, probe, ndev, + "Missing IRQ resource\n"); + goto err2; + } + ndev->irq = res->start; + + while (priv->clock_count < ARRAY_SIZE(priv->clk)) { + priv->clk[priv->clock_count] = + of_clk_get(pdev->dev.of_node, priv->clock_count); + if (IS_ERR(priv->clk[priv->clock_count])) { + if (!priv->clock_count) { + netif_err(priv, probe, ndev, + "Failed to get clock\n"); + goto err3; + } + break; + } + priv->clock_count++; + } + + /* disable by default */ + priv->et_coalesce.rx_coalesce_usecs = 0; + priv->et_coalesce.rx_max_coalesced_frames = 1; + priv->et_coalesce.tx_coalesce_usecs = 0; + priv->et_coalesce.tx_max_coalesced_frames = 1; + + use_jumbo = of_property_read_bool(pdev->dev.of_node, "use-jumbo"); + priv->param.use_jumbo_pkt_flag = use_jumbo; + + if (priv->param.use_jumbo_pkt_flag) + priv->rx_pkt_buf_len = NETSEC_RX_JUMBO_PKT_BUF_LEN; + else + priv->rx_pkt_buf_len = NETSEC_RX_PKT_BUF_LEN; + + pm_runtime_enable(&pdev->dev); + /* runtime_pm coverage just for probe, open/close also cover it */ + pm_runtime_get_sync(&pdev->dev); + + hw_ver = netsec_readl(priv, NETSEC_REG_F_TAIKI_VER); + /* this driver only supports F_TAIKI style NETSEC */ + if (NETSEC_F_NETSEC_VER_MAJOR_NUM(hw_ver) != + NETSEC_F_NETSEC_VER_MAJOR_NUM(NETSEC_REG_NETSEC_VER_F_TAIKI)) { + ret = -ENODEV; + goto err3; + } + + dev_info(&pdev->dev, "IP rev %d.%d\n", hw_ver >> 16, hw_ver & 0xffff); + + priv->mac_mode.flow_start_th = NETSEC_FLOW_CONTROL_START_THRESHOLD; + priv->mac_mode.flow_stop_th = NETSEC_FLOW_CONTROL_STOP_THRESHOLD; + priv->mac_mode.pause_time = pause_time; + priv->mac_mode.flow_ctrl_enable_flag = false; + priv->freq = clk_get_rate(priv->clk[0]); + + netif_napi_add(ndev, &priv->napi, netsec_netdev_napi_poll, + napi_weight); + + /* MTU range */ + ndev->min_mtu = ETH_MIN_MTU; + ndev->max_mtu = NETSEC_RX_JUMBO_PKT_BUF_LEN; + + ndev->netdev_ops = &netsec_netdev_ops; + ndev->ethtool_ops = &netsec_ethtool_ops; + ndev->features = NETIF_F_SG | NETIF_F_IP_CSUM | + NETIF_F_IPV6_CSUM | NETIF_F_TSO | + NETIF_F_TSO6 | NETIF_F_GSO | + NETIF_F_HIGHDMA | NETIF_F_RXCSUM; + ndev->hw_features = ndev->features; + + priv->rx_cksum_offload_flag = true; + spin_lock_init(&priv->tx_queue_lock); + + err = netsec_mii_register(priv); + if (err) { + netif_err(priv, probe, ndev, + "mii bus registration failed %d\n", err); + goto err3; + } + + /* disable all other interrupt sources */ + netsec_writel(priv, NETSEC_REG_INTEN_CLR, ~0); + netsec_writel(priv, NETSEC_REG_INTEN_SET, + NETSEC_IRQ_TX | NETSEC_IRQ_RX); + + err = register_netdev(ndev); + if (err) { + netif_err(priv, probe, ndev, "register_netdev() failed\n"); + goto err4; + } + + pm_runtime_put_sync_suspend(&pdev->dev); + + netif_info(priv, probe, ndev, "initialized\n"); + + return 0; + +err4: + netsec_mii_unregister(priv); + +err3: + pm_runtime_put_sync_suspend(&pdev->dev); + pm_runtime_disable(&pdev->dev); + while (priv->clock_count > 0) { + priv->clock_count--; + clk_put(priv->clk[priv->clock_count]); + } +err2: + iounmap(priv->ioaddr); +err1: + free_netdev(ndev); + + dev_err(&pdev->dev, "init failed\n"); + + return ret; +} + +static int netsec_remove(struct platform_device *pdev) +{ + struct netsec_priv *priv = platform_get_drvdata(pdev); + + unregister_netdev(priv->ndev); + netsec_mii_unregister(priv); + pm_runtime_disable(&pdev->dev); + iounmap(priv->ioaddr); + free_netdev(priv->ndev); + + return 0; +} + +#ifdef CONFIG_PM +static int netsec_runtime_suspend(struct device *dev) +{ + struct netsec_priv *priv = dev_get_drvdata(dev); + int n; + + netif_dbg(priv, drv, priv->ndev, "%s\n", __func__); + + if (priv->irq_registered) + disable_irq(priv->ndev->irq); + + netsec_writel(priv, NETSEC_REG_CLK_EN, 0); + + for (n = priv->clock_count - 1; n >= 0; n--) + clk_disable_unprepare(priv->clk[n]); + + return 0; +} + +static int netsec_runtime_resume(struct device *dev) +{ + struct netsec_priv *priv = dev_get_drvdata(dev); + int n; + + netif_dbg(priv, drv, priv->ndev, "%s\n", __func__); + + /* first let the clocks back on */ + + for (n = 0; n < priv->clock_count; n++) + clk_prepare_enable(priv->clk[n]); + + netsec_writel(priv, NETSEC_REG_CLK_EN, NETSEC_CLK_EN_REG_DOM_D | + NETSEC_CLK_EN_REG_DOM_C | NETSEC_CLK_EN_REG_DOM_G); + + if (priv->irq_registered) + enable_irq(priv->ndev->irq); + + return 0; +} + +static int netsec_pm_suspend(struct device *dev) +{ + struct netsec_priv *priv = dev_get_drvdata(dev); + + netif_dbg(priv, drv, priv->ndev, "%s\n", __func__); + + if (pm_runtime_status_suspended(dev)) + return 0; + + return netsec_runtime_suspend(dev); +} + +static int netsec_pm_resume(struct device *dev) +{ + struct netsec_priv *priv = dev_get_drvdata(dev); + + netif_dbg(priv, drv, priv->ndev, "%s\n", __func__); + + if (pm_runtime_status_suspended(dev)) + return 0; + + return netsec_runtime_resume(dev); +} +#endif + +static const struct dev_pm_ops netsec_pm_ops = { + SET_SYSTEM_SLEEP_PM_OPS(netsec_pm_suspend, netsec_pm_resume) + SET_RUNTIME_PM_OPS(netsec_runtime_suspend, netsec_runtime_resume, NULL) +}; + +static const struct of_device_id netsec_dt_ids[] = { + {.compatible = "socionext,netsecv5"}, + { /* sentinel */ } +}; + +MODULE_DEVICE_TABLE(of, netsec_dt_ids); + +static struct platform_driver netsec_driver = { + .probe = netsec_probe, + .remove = netsec_remove, + .driver = { + .name = "netsec", + .of_match_table = netsec_dt_ids, + .pm = &netsec_pm_ops, + }, +}; + +module_platform_driver(netsec_driver); + +MODULE_AUTHOR("Socionext Inc"); +MODULE_DESCRIPTION("NETSEC Ethernet driver"); +MODULE_LICENSE("GPL"); + +MODULE_ALIAS("platform:netsec");