From patchwork Tue Feb 9 09:05:54 2021 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Loic Poulain X-Patchwork-Id: 12077533 X-Patchwork-Delegate: kuba@kernel.org Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-18.8 required=3.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_CR_TRAILER, INCLUDES_PATCH,MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED, USER_AGENT_GIT autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 77B60C433E0 for ; Tue, 9 Feb 2021 09:00:38 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 3B7F464EA6 for ; Tue, 9 Feb 2021 09:00:38 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S230310AbhBIJAb (ORCPT ); Tue, 9 Feb 2021 04:00:31 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:54856 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S230170AbhBII6q (ORCPT ); Tue, 9 Feb 2021 03:58:46 -0500 Received: from mail-wr1-x436.google.com (mail-wr1-x436.google.com [IPv6:2a00:1450:4864:20::436]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id DA2B7C06178A for ; Tue, 9 Feb 2021 00:57:57 -0800 (PST) Received: by mail-wr1-x436.google.com with SMTP id n6so7815422wrv.8 for ; Tue, 09 Feb 2021 00:57:57 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linaro.org; s=google; h=from:to:cc:subject:date:message-id:in-reply-to:references; bh=D1xoggkKUYPm/up9HUqtS9Xiw3MkdDwvnq8xK4ouEnc=; b=DEMrIs+7nG1vykmUxnuuBvgSEP+iWHxBRCzZtsw0kFiLmpVsMMXbP0jkNnZ4ZMFdvG am2V1JM3QfksypjkgRgkwNtTBhtlkQmgFxx5r6UfDRDtaa81dm7AAAbttGol/ISFym3v JMeML9dXEcaVEG0BoIgi7c5qrBi1sy0c7OpQJUm/EfWBl6jP9pjpG2LZfdBPf93OsJni JFtsm33LEylTQjjFza6c/u8z6wDNtz/61c5iL4oid6nBWgt3aQLJTaREo8Ec/JYWq9E8 3/VL/2twbW6IXHeSQJdDE0ULN5wx/FR+VS8wk/xbIjJ77EAN/ajpriglscpk+V+wmi2k jVdQ== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references; bh=D1xoggkKUYPm/up9HUqtS9Xiw3MkdDwvnq8xK4ouEnc=; b=LM0heQdDemkwwZ56+HcuHRyBB/2TAuZkbJuYTm7L5iPKyXwbleUUD+WqvVYqobLpsU oZE6txwl79l5NDuzvkVXXnOFgh1QCriFxNuDGv3Yo/+Is5tK95DRG9Q78jtd/svxHu03 S60b5d4ruxUhOVKCngs4G1Nv30OicSDVIoQQpH5mmuai9k4u1hZ6pLxqR1GtzXngwQBB BlJDsa/GbVZheFtpernPwUXJfAmaNbCvVo1bX5vrUGqbIAJ4n+EOPSS21BMnuxrT/En/ 5alxo0l0oL6/D4VRoYW79hFcf5cwzlMRfFFhaNAvID044BSmnmLmCTF6hcVrPS9roM2u CkvQ== X-Gm-Message-State: AOAM532u7bW2NWdwkPDClPN84o61NM5mynFtEZ4tdWdL543pHpZFhdNq NMHwfe6BTV4T4De/B9l1bG6AWo4/kxUli+Oo X-Google-Smtp-Source: ABdhPJwNV+VUd5Vf7PuNNn7qarhW1t9EIqe8yK4wSO3huC0VM4vwUr22ANuiASKu5CZTuVA0pbxewA== X-Received: by 2002:adf:c40a:: with SMTP id v10mr10676723wrf.10.1612861076528; Tue, 09 Feb 2021 00:57:56 -0800 (PST) Received: from localhost.localdomain ([88.122.66.28]) by smtp.gmail.com with ESMTPSA id d3sm38348693wrp.79.2021.02.09.00.57.55 (version=TLS1_2 cipher=ECDHE-ECDSA-AES128-GCM-SHA256 bits=128/128); Tue, 09 Feb 2021 00:57:56 -0800 (PST) From: Loic Poulain To: kuba@kernel.org, davem@davemloft.net Cc: netdev@vger.kernel.org, bjorn@mork.no, dcbw@redhat.com, carl.yin@quectel.com, mpearson@lenovo.com, cchen50@lenovo.com, jwjiang@lenovo.com, ivan.zhang@quectel.com, naveen.kumar@quectel.com, ivan.mikhanchuk@quectel.com, Loic Poulain Subject: [PATCH net-next v5 1/5] net: mhi: Add protocol support Date: Tue, 9 Feb 2021 10:05:54 +0100 Message-Id: <1612861558-14487-2-git-send-email-loic.poulain@linaro.org> X-Mailer: git-send-email 2.7.4 In-Reply-To: <1612861558-14487-1-git-send-email-loic.poulain@linaro.org> References: <1612861558-14487-1-git-send-email-loic.poulain@linaro.org> Precedence: bulk List-ID: X-Mailing-List: netdev@vger.kernel.org X-Patchwork-Delegate: kuba@kernel.org MHI can transport different protocols, some are handled at upper level, like IP and QMAP(rmnet/netlink), but others will need to be inside MHI net driver, like mbim. This change adds support for protocol rx and tx_fixup callbacks registration, that can be used to encode/decode the targeted protocol. Signed-off-by: Loic Poulain --- drivers/net/mhi_net.c | 69 ++++++++++++++++++++++++++++++++++++++++++--------- 1 file changed, 57 insertions(+), 12 deletions(-) diff --git a/drivers/net/mhi_net.c b/drivers/net/mhi_net.c index 8800991..b92c2e1 100644 --- a/drivers/net/mhi_net.c +++ b/drivers/net/mhi_net.c @@ -34,11 +34,24 @@ struct mhi_net_dev { struct net_device *ndev; struct sk_buff *skbagg_head; struct sk_buff *skbagg_tail; + const struct mhi_net_proto *proto; + void *proto_data; struct delayed_work rx_refill; struct mhi_net_stats stats; u32 rx_queue_sz; }; +struct mhi_net_proto { + int (*init)(struct mhi_net_dev *mhi_netdev); + struct sk_buff * (*tx_fixup)(struct mhi_net_dev *mhi_netdev, struct sk_buff *skb); + void (*rx)(struct mhi_net_dev *mhi_netdev, struct sk_buff *skb); +}; + +struct mhi_device_info { + const char *netname; + const struct mhi_net_proto *proto; +}; + static int mhi_ndo_open(struct net_device *ndev) { struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); @@ -68,26 +81,35 @@ static int mhi_ndo_stop(struct net_device *ndev) static int mhi_ndo_xmit(struct sk_buff *skb, struct net_device *ndev) { struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); + const struct mhi_net_proto *proto = mhi_netdev->proto; struct mhi_device *mdev = mhi_netdev->mdev; int err; + if (proto && proto->tx_fixup) { + skb = proto->tx_fixup(mhi_netdev, skb); + if (unlikely(!skb)) + goto exit_drop; + } + err = mhi_queue_skb(mdev, DMA_TO_DEVICE, skb, skb->len, MHI_EOT); if (unlikely(err)) { net_err_ratelimited("%s: Failed to queue TX buf (%d)\n", ndev->name, err); - - u64_stats_update_begin(&mhi_netdev->stats.tx_syncp); - u64_stats_inc(&mhi_netdev->stats.tx_dropped); - u64_stats_update_end(&mhi_netdev->stats.tx_syncp); - - /* drop the packet */ dev_kfree_skb_any(skb); + goto exit_drop; } if (mhi_queue_is_full(mdev, DMA_TO_DEVICE)) netif_stop_queue(ndev); return NETDEV_TX_OK; + +exit_drop: + u64_stats_update_begin(&mhi_netdev->stats.tx_syncp); + u64_stats_inc(&mhi_netdev->stats.tx_dropped); + u64_stats_update_end(&mhi_netdev->stats.tx_syncp); + + return NETDEV_TX_OK; } static void mhi_ndo_get_stats64(struct net_device *ndev, @@ -164,6 +186,7 @@ static void mhi_net_dl_callback(struct mhi_device *mhi_dev, struct mhi_result *mhi_res) { struct mhi_net_dev *mhi_netdev = dev_get_drvdata(&mhi_dev->dev); + const struct mhi_net_proto *proto = mhi_netdev->proto; struct sk_buff *skb = mhi_res->buf_addr; int free_desc_count; @@ -220,7 +243,10 @@ static void mhi_net_dl_callback(struct mhi_device *mhi_dev, break; } - netif_rx(skb); + if (proto && proto->rx) + proto->rx(mhi_netdev, skb); + else + netif_rx(skb); } /* Refill if RX buffers queue becomes low */ @@ -302,14 +328,14 @@ static struct device_type wwan_type = { static int mhi_net_probe(struct mhi_device *mhi_dev, const struct mhi_device_id *id) { - const char *netname = (char *)id->driver_data; + const struct mhi_device_info *info = (struct mhi_device_info *)id->driver_data; struct device *dev = &mhi_dev->dev; struct mhi_net_dev *mhi_netdev; struct net_device *ndev; int err; - ndev = alloc_netdev(sizeof(*mhi_netdev), netname, NET_NAME_PREDICTABLE, - mhi_net_setup); + ndev = alloc_netdev(sizeof(*mhi_netdev), info->netname, + NET_NAME_PREDICTABLE, mhi_net_setup); if (!ndev) return -ENOMEM; @@ -318,6 +344,7 @@ static int mhi_net_probe(struct mhi_device *mhi_dev, mhi_netdev->ndev = ndev; mhi_netdev->mdev = mhi_dev; mhi_netdev->skbagg_head = NULL; + mhi_netdev->proto = info->proto; SET_NETDEV_DEV(ndev, &mhi_dev->dev); SET_NETDEV_DEVTYPE(ndev, &wwan_type); @@ -337,8 +364,16 @@ static int mhi_net_probe(struct mhi_device *mhi_dev, if (err) goto out_err; + if (mhi_netdev->proto) { + err = mhi_netdev->proto->init(mhi_netdev); + if (err) + goto out_err_proto; + } + return 0; +out_err_proto: + unregister_netdev(ndev); out_err: free_netdev(ndev); return err; @@ -358,9 +393,19 @@ static void mhi_net_remove(struct mhi_device *mhi_dev) free_netdev(mhi_netdev->ndev); } +static const struct mhi_device_info mhi_hwip0 = { + .netname = "mhi_hwip%d", +}; + +static const struct mhi_device_info mhi_swip0 = { + .netname = "mhi_swip%d", +}; + static const struct mhi_device_id mhi_net_id_table[] = { - { .chan = "IP_HW0", .driver_data = (kernel_ulong_t)"mhi_hwip%d" }, - { .chan = "IP_SW0", .driver_data = (kernel_ulong_t)"mhi_swip%d" }, + /* Hardware accelerated data PATH (to modem IPA), protocol agnostic */ + { .chan = "IP_HW0", .driver_data = (kernel_ulong_t)&mhi_hwip0 }, + /* Software data PATH (to modem CPU) */ + { .chan = "IP_SW0", .driver_data = (kernel_ulong_t)&mhi_swip0 }, {} }; MODULE_DEVICE_TABLE(mhi, mhi_net_id_table); From patchwork Tue Feb 9 09:05:55 2021 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Loic Poulain X-Patchwork-Id: 12077535 X-Patchwork-Delegate: kuba@kernel.org Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-18.8 required=3.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_CR_TRAILER, INCLUDES_PATCH,MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED, USER_AGENT_GIT autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 941FAC433DB for ; Tue, 9 Feb 2021 09:01:23 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 4087664DEE for ; Tue, 9 Feb 2021 09:01:23 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S230245AbhBIJBB (ORCPT ); Tue, 9 Feb 2021 04:01:01 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:54868 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S230177AbhBII6q (ORCPT ); Tue, 9 Feb 2021 03:58:46 -0500 Received: from mail-wr1-x42f.google.com (mail-wr1-x42f.google.com [IPv6:2a00:1450:4864:20::42f]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 1FE49C06178B for ; Tue, 9 Feb 2021 00:58:00 -0800 (PST) Received: by mail-wr1-x42f.google.com with SMTP id g10so20682987wrx.1 for ; Tue, 09 Feb 2021 00:58:00 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linaro.org; s=google; h=from:to:cc:subject:date:message-id:in-reply-to:references; bh=cf0LfcfRp3bcgJncg1NSTu2ZamMvFp0Ok+t82TG7INQ=; b=wkXWJVHk1a1ASiHT4GG8k4wYKwOHUIGnYRfUub15qz2cmdZVk7ES4MQ9sdw9PZ5+mJ TyM/atkZ20KuKdRrHHW2GfiuG8trvCAmf12efEeIuH8oBkNbKuDuKLDW6AliFdZ+npMy VK1qxsFbe1l5RxVrFCq5Gt3fM5Zdsfsan1XjuciWd9CHZSdmrZ8rwxqkj3QHvxpUkUSw AtvgsNe9fsvLpKFDGhL5W5/4rbgTvFQANmUNqM70yigUPHjYwmkUi+goPScJmKSHt1Do T9wce+KnxbOsvnSwXs/RPrD3AMfyjY+0eStzYt9YDoRNT9sxxfekIKwIMJuBvhZ7HyiF PQeg== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references; bh=cf0LfcfRp3bcgJncg1NSTu2ZamMvFp0Ok+t82TG7INQ=; b=ZDC85ewLU3BW5FSEX3J2OuIOoyq6nr3sI70IWmhcYN56oEmSqBF9pUpkblZ8t/5aTV JdzNrwDGKyba8g/HRlvAT6zoaJip11FmB5a2/Q1Ch2vTbQpZwn7XQxYDa8fUNxRb0HQ4 RzJVMv2lRPmPTp0S22weLK0/edM3go6+Jnx7dsTHiAmv5bUj+CqPZMV/dpqeyefilx5S 9KNvEBYQCBLJN1MPZvcCkVPPjQd4eH6N4NF/ja/y6IKUeSl9Qows8QOtcIA4pkTz310I XXvZIBKxJQGhMZQ4E9j1fWpRkkm5DZ+Ca8J480JPb9pOX5n5IvaakVPwZL3yu0UfvU6W s8EQ== X-Gm-Message-State: AOAM53031zKL0PhTBGY0In/PXgho2LIwSHA6PdurtLycYBLUQDAXgvHj oUvBPbJ2MPQeHy05F+Vp+QseiMJk1hB2TCM2 X-Google-Smtp-Source: ABdhPJz/cR1XKN/oGV+Wyv2w7fZ68IW+CJtdhAw6PBx6Sc7aF368l2n949hZYH9EbyDD6D+dHnBstw== X-Received: by 2002:a5d:4f84:: with SMTP id d4mr23940308wru.374.1612861077973; Tue, 09 Feb 2021 00:57:57 -0800 (PST) Received: from localhost.localdomain ([88.122.66.28]) by smtp.gmail.com with ESMTPSA id d3sm38348693wrp.79.2021.02.09.00.57.56 (version=TLS1_2 cipher=ECDHE-ECDSA-AES128-GCM-SHA256 bits=128/128); Tue, 09 Feb 2021 00:57:57 -0800 (PST) From: Loic Poulain To: kuba@kernel.org, davem@davemloft.net Cc: netdev@vger.kernel.org, bjorn@mork.no, dcbw@redhat.com, carl.yin@quectel.com, mpearson@lenovo.com, cchen50@lenovo.com, jwjiang@lenovo.com, ivan.zhang@quectel.com, naveen.kumar@quectel.com, ivan.mikhanchuk@quectel.com, Loic Poulain Subject: [PATCH net-next v5 2/5] net: mhi: Add dedicated folder Date: Tue, 9 Feb 2021 10:05:55 +0100 Message-Id: <1612861558-14487-3-git-send-email-loic.poulain@linaro.org> X-Mailer: git-send-email 2.7.4 In-Reply-To: <1612861558-14487-1-git-send-email-loic.poulain@linaro.org> References: <1612861558-14487-1-git-send-email-loic.poulain@linaro.org> Precedence: bulk List-ID: X-Mailing-List: netdev@vger.kernel.org X-Patchwork-Delegate: kuba@kernel.org Create a dedicated mhi directory for mhi-net, mhi-net is going to be split into differente files (for additional protocol support). Signed-off-by: Loic Poulain --- drivers/net/Makefile | 2 +- drivers/net/mhi/Makefile | 3 + drivers/net/mhi/net.c | 429 +++++++++++++++++++++++++++++++++++++++++++++++ drivers/net/mhi_net.c | 429 ----------------------------------------------- 4 files changed, 433 insertions(+), 430 deletions(-) create mode 100644 drivers/net/mhi/Makefile create mode 100644 drivers/net/mhi/net.c delete mode 100644 drivers/net/mhi_net.c diff --git a/drivers/net/Makefile b/drivers/net/Makefile index 36e2e41..f4990ff 100644 --- a/drivers/net/Makefile +++ b/drivers/net/Makefile @@ -36,7 +36,7 @@ obj-$(CONFIG_GTP) += gtp.o obj-$(CONFIG_NLMON) += nlmon.o obj-$(CONFIG_NET_VRF) += vrf.o obj-$(CONFIG_VSOCKMON) += vsockmon.o -obj-$(CONFIG_MHI_NET) += mhi_net.o +obj-$(CONFIG_MHI_NET) += mhi/ # # Networking Drivers diff --git a/drivers/net/mhi/Makefile b/drivers/net/mhi/Makefile new file mode 100644 index 0000000..0acf989 --- /dev/null +++ b/drivers/net/mhi/Makefile @@ -0,0 +1,3 @@ +obj-$(CONFIG_MHI_NET) += mhi_net.o + +mhi_net-y := net.o diff --git a/drivers/net/mhi/net.c b/drivers/net/mhi/net.c new file mode 100644 index 0000000..b92c2e1 --- /dev/null +++ b/drivers/net/mhi/net.c @@ -0,0 +1,429 @@ +// SPDX-License-Identifier: GPL-2.0-or-later +/* MHI Network driver - Network over MHI bus + * + * Copyright (C) 2020 Linaro Ltd + */ + +#include +#include +#include +#include +#include +#include +#include + +#define MHI_NET_MIN_MTU ETH_MIN_MTU +#define MHI_NET_MAX_MTU 0xffff +#define MHI_NET_DEFAULT_MTU 0x4000 + +struct mhi_net_stats { + u64_stats_t rx_packets; + u64_stats_t rx_bytes; + u64_stats_t rx_errors; + u64_stats_t rx_dropped; + u64_stats_t tx_packets; + u64_stats_t tx_bytes; + u64_stats_t tx_errors; + u64_stats_t tx_dropped; + struct u64_stats_sync tx_syncp; + struct u64_stats_sync rx_syncp; +}; + +struct mhi_net_dev { + struct mhi_device *mdev; + struct net_device *ndev; + struct sk_buff *skbagg_head; + struct sk_buff *skbagg_tail; + const struct mhi_net_proto *proto; + void *proto_data; + struct delayed_work rx_refill; + struct mhi_net_stats stats; + u32 rx_queue_sz; +}; + +struct mhi_net_proto { + int (*init)(struct mhi_net_dev *mhi_netdev); + struct sk_buff * (*tx_fixup)(struct mhi_net_dev *mhi_netdev, struct sk_buff *skb); + void (*rx)(struct mhi_net_dev *mhi_netdev, struct sk_buff *skb); +}; + +struct mhi_device_info { + const char *netname; + const struct mhi_net_proto *proto; +}; + +static int mhi_ndo_open(struct net_device *ndev) +{ + struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); + + /* Feed the rx buffer pool */ + schedule_delayed_work(&mhi_netdev->rx_refill, 0); + + /* Carrier is established via out-of-band channel (e.g. qmi) */ + netif_carrier_on(ndev); + + netif_start_queue(ndev); + + return 0; +} + +static int mhi_ndo_stop(struct net_device *ndev) +{ + struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); + + netif_stop_queue(ndev); + netif_carrier_off(ndev); + cancel_delayed_work_sync(&mhi_netdev->rx_refill); + + return 0; +} + +static int mhi_ndo_xmit(struct sk_buff *skb, struct net_device *ndev) +{ + struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); + const struct mhi_net_proto *proto = mhi_netdev->proto; + struct mhi_device *mdev = mhi_netdev->mdev; + int err; + + if (proto && proto->tx_fixup) { + skb = proto->tx_fixup(mhi_netdev, skb); + if (unlikely(!skb)) + goto exit_drop; + } + + err = mhi_queue_skb(mdev, DMA_TO_DEVICE, skb, skb->len, MHI_EOT); + if (unlikely(err)) { + net_err_ratelimited("%s: Failed to queue TX buf (%d)\n", + ndev->name, err); + dev_kfree_skb_any(skb); + goto exit_drop; + } + + if (mhi_queue_is_full(mdev, DMA_TO_DEVICE)) + netif_stop_queue(ndev); + + return NETDEV_TX_OK; + +exit_drop: + u64_stats_update_begin(&mhi_netdev->stats.tx_syncp); + u64_stats_inc(&mhi_netdev->stats.tx_dropped); + u64_stats_update_end(&mhi_netdev->stats.tx_syncp); + + return NETDEV_TX_OK; +} + +static void mhi_ndo_get_stats64(struct net_device *ndev, + struct rtnl_link_stats64 *stats) +{ + struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); + unsigned int start; + + do { + start = u64_stats_fetch_begin_irq(&mhi_netdev->stats.rx_syncp); + stats->rx_packets = u64_stats_read(&mhi_netdev->stats.rx_packets); + stats->rx_bytes = u64_stats_read(&mhi_netdev->stats.rx_bytes); + stats->rx_errors = u64_stats_read(&mhi_netdev->stats.rx_errors); + stats->rx_dropped = u64_stats_read(&mhi_netdev->stats.rx_dropped); + } while (u64_stats_fetch_retry_irq(&mhi_netdev->stats.rx_syncp, start)); + + do { + start = u64_stats_fetch_begin_irq(&mhi_netdev->stats.tx_syncp); + stats->tx_packets = u64_stats_read(&mhi_netdev->stats.tx_packets); + stats->tx_bytes = u64_stats_read(&mhi_netdev->stats.tx_bytes); + stats->tx_errors = u64_stats_read(&mhi_netdev->stats.tx_errors); + stats->tx_dropped = u64_stats_read(&mhi_netdev->stats.tx_dropped); + } while (u64_stats_fetch_retry_irq(&mhi_netdev->stats.tx_syncp, start)); +} + +static const struct net_device_ops mhi_netdev_ops = { + .ndo_open = mhi_ndo_open, + .ndo_stop = mhi_ndo_stop, + .ndo_start_xmit = mhi_ndo_xmit, + .ndo_get_stats64 = mhi_ndo_get_stats64, +}; + +static void mhi_net_setup(struct net_device *ndev) +{ + ndev->header_ops = NULL; /* No header */ + ndev->type = ARPHRD_RAWIP; + ndev->hard_header_len = 0; + ndev->addr_len = 0; + ndev->flags = IFF_POINTOPOINT | IFF_NOARP; + ndev->netdev_ops = &mhi_netdev_ops; + ndev->mtu = MHI_NET_DEFAULT_MTU; + ndev->min_mtu = MHI_NET_MIN_MTU; + ndev->max_mtu = MHI_NET_MAX_MTU; + ndev->tx_queue_len = 1000; +} + +static struct sk_buff *mhi_net_skb_agg(struct mhi_net_dev *mhi_netdev, + struct sk_buff *skb) +{ + struct sk_buff *head = mhi_netdev->skbagg_head; + struct sk_buff *tail = mhi_netdev->skbagg_tail; + + /* This is non-paged skb chaining using frag_list */ + if (!head) { + mhi_netdev->skbagg_head = skb; + return skb; + } + + if (!skb_shinfo(head)->frag_list) + skb_shinfo(head)->frag_list = skb; + else + tail->next = skb; + + head->len += skb->len; + head->data_len += skb->len; + head->truesize += skb->truesize; + + mhi_netdev->skbagg_tail = skb; + + return mhi_netdev->skbagg_head; +} + +static void mhi_net_dl_callback(struct mhi_device *mhi_dev, + struct mhi_result *mhi_res) +{ + struct mhi_net_dev *mhi_netdev = dev_get_drvdata(&mhi_dev->dev); + const struct mhi_net_proto *proto = mhi_netdev->proto; + struct sk_buff *skb = mhi_res->buf_addr; + int free_desc_count; + + free_desc_count = mhi_get_free_desc_count(mhi_dev, DMA_FROM_DEVICE); + + if (unlikely(mhi_res->transaction_status)) { + switch (mhi_res->transaction_status) { + case -EOVERFLOW: + /* Packet can not fit in one MHI buffer and has been + * split over multiple MHI transfers, do re-aggregation. + * That usually means the device side MTU is larger than + * the host side MTU/MRU. Since this is not optimal, + * print a warning (once). + */ + netdev_warn_once(mhi_netdev->ndev, + "Fragmented packets received, fix MTU?\n"); + skb_put(skb, mhi_res->bytes_xferd); + mhi_net_skb_agg(mhi_netdev, skb); + break; + case -ENOTCONN: + /* MHI layer stopping/resetting the DL channel */ + dev_kfree_skb_any(skb); + return; + default: + /* Unknown error, simply drop */ + dev_kfree_skb_any(skb); + u64_stats_update_begin(&mhi_netdev->stats.rx_syncp); + u64_stats_inc(&mhi_netdev->stats.rx_errors); + u64_stats_update_end(&mhi_netdev->stats.rx_syncp); + } + } else { + skb_put(skb, mhi_res->bytes_xferd); + + if (mhi_netdev->skbagg_head) { + /* Aggregate the final fragment */ + skb = mhi_net_skb_agg(mhi_netdev, skb); + mhi_netdev->skbagg_head = NULL; + } + + u64_stats_update_begin(&mhi_netdev->stats.rx_syncp); + u64_stats_inc(&mhi_netdev->stats.rx_packets); + u64_stats_add(&mhi_netdev->stats.rx_bytes, skb->len); + u64_stats_update_end(&mhi_netdev->stats.rx_syncp); + + switch (skb->data[0] & 0xf0) { + case 0x40: + skb->protocol = htons(ETH_P_IP); + break; + case 0x60: + skb->protocol = htons(ETH_P_IPV6); + break; + default: + skb->protocol = htons(ETH_P_MAP); + break; + } + + if (proto && proto->rx) + proto->rx(mhi_netdev, skb); + else + netif_rx(skb); + } + + /* Refill if RX buffers queue becomes low */ + if (free_desc_count >= mhi_netdev->rx_queue_sz / 2) + schedule_delayed_work(&mhi_netdev->rx_refill, 0); +} + +static void mhi_net_ul_callback(struct mhi_device *mhi_dev, + struct mhi_result *mhi_res) +{ + struct mhi_net_dev *mhi_netdev = dev_get_drvdata(&mhi_dev->dev); + struct net_device *ndev = mhi_netdev->ndev; + struct mhi_device *mdev = mhi_netdev->mdev; + struct sk_buff *skb = mhi_res->buf_addr; + + /* Hardware has consumed the buffer, so free the skb (which is not + * freed by the MHI stack) and perform accounting. + */ + dev_consume_skb_any(skb); + + u64_stats_update_begin(&mhi_netdev->stats.tx_syncp); + if (unlikely(mhi_res->transaction_status)) { + + /* MHI layer stopping/resetting the UL channel */ + if (mhi_res->transaction_status == -ENOTCONN) { + u64_stats_update_end(&mhi_netdev->stats.tx_syncp); + return; + } + + u64_stats_inc(&mhi_netdev->stats.tx_errors); + } else { + u64_stats_inc(&mhi_netdev->stats.tx_packets); + u64_stats_add(&mhi_netdev->stats.tx_bytes, mhi_res->bytes_xferd); + } + u64_stats_update_end(&mhi_netdev->stats.tx_syncp); + + if (netif_queue_stopped(ndev) && !mhi_queue_is_full(mdev, DMA_TO_DEVICE)) + netif_wake_queue(ndev); +} + +static void mhi_net_rx_refill_work(struct work_struct *work) +{ + struct mhi_net_dev *mhi_netdev = container_of(work, struct mhi_net_dev, + rx_refill.work); + struct net_device *ndev = mhi_netdev->ndev; + struct mhi_device *mdev = mhi_netdev->mdev; + int size = READ_ONCE(ndev->mtu); + struct sk_buff *skb; + int err; + + while (!mhi_queue_is_full(mdev, DMA_FROM_DEVICE)) { + skb = netdev_alloc_skb(ndev, size); + if (unlikely(!skb)) + break; + + err = mhi_queue_skb(mdev, DMA_FROM_DEVICE, skb, size, MHI_EOT); + if (unlikely(err)) { + net_err_ratelimited("%s: Failed to queue RX buf (%d)\n", + ndev->name, err); + kfree_skb(skb); + break; + } + + /* Do not hog the CPU if rx buffers are consumed faster than + * queued (unlikely). + */ + cond_resched(); + } + + /* If we're still starved of rx buffers, reschedule later */ + if (mhi_get_free_desc_count(mdev, DMA_FROM_DEVICE) == mhi_netdev->rx_queue_sz) + schedule_delayed_work(&mhi_netdev->rx_refill, HZ / 2); +} + +static struct device_type wwan_type = { + .name = "wwan", +}; + +static int mhi_net_probe(struct mhi_device *mhi_dev, + const struct mhi_device_id *id) +{ + const struct mhi_device_info *info = (struct mhi_device_info *)id->driver_data; + struct device *dev = &mhi_dev->dev; + struct mhi_net_dev *mhi_netdev; + struct net_device *ndev; + int err; + + ndev = alloc_netdev(sizeof(*mhi_netdev), info->netname, + NET_NAME_PREDICTABLE, mhi_net_setup); + if (!ndev) + return -ENOMEM; + + mhi_netdev = netdev_priv(ndev); + dev_set_drvdata(dev, mhi_netdev); + mhi_netdev->ndev = ndev; + mhi_netdev->mdev = mhi_dev; + mhi_netdev->skbagg_head = NULL; + mhi_netdev->proto = info->proto; + SET_NETDEV_DEV(ndev, &mhi_dev->dev); + SET_NETDEV_DEVTYPE(ndev, &wwan_type); + + INIT_DELAYED_WORK(&mhi_netdev->rx_refill, mhi_net_rx_refill_work); + u64_stats_init(&mhi_netdev->stats.rx_syncp); + u64_stats_init(&mhi_netdev->stats.tx_syncp); + + /* Start MHI channels */ + err = mhi_prepare_for_transfer(mhi_dev); + if (err) + goto out_err; + + /* Number of transfer descriptors determines size of the queue */ + mhi_netdev->rx_queue_sz = mhi_get_free_desc_count(mhi_dev, DMA_FROM_DEVICE); + + err = register_netdev(ndev); + if (err) + goto out_err; + + if (mhi_netdev->proto) { + err = mhi_netdev->proto->init(mhi_netdev); + if (err) + goto out_err_proto; + } + + return 0; + +out_err_proto: + unregister_netdev(ndev); +out_err: + free_netdev(ndev); + return err; +} + +static void mhi_net_remove(struct mhi_device *mhi_dev) +{ + struct mhi_net_dev *mhi_netdev = dev_get_drvdata(&mhi_dev->dev); + + unregister_netdev(mhi_netdev->ndev); + + mhi_unprepare_from_transfer(mhi_netdev->mdev); + + if (mhi_netdev->skbagg_head) + kfree_skb(mhi_netdev->skbagg_head); + + free_netdev(mhi_netdev->ndev); +} + +static const struct mhi_device_info mhi_hwip0 = { + .netname = "mhi_hwip%d", +}; + +static const struct mhi_device_info mhi_swip0 = { + .netname = "mhi_swip%d", +}; + +static const struct mhi_device_id mhi_net_id_table[] = { + /* Hardware accelerated data PATH (to modem IPA), protocol agnostic */ + { .chan = "IP_HW0", .driver_data = (kernel_ulong_t)&mhi_hwip0 }, + /* Software data PATH (to modem CPU) */ + { .chan = "IP_SW0", .driver_data = (kernel_ulong_t)&mhi_swip0 }, + {} +}; +MODULE_DEVICE_TABLE(mhi, mhi_net_id_table); + +static struct mhi_driver mhi_net_driver = { + .probe = mhi_net_probe, + .remove = mhi_net_remove, + .dl_xfer_cb = mhi_net_dl_callback, + .ul_xfer_cb = mhi_net_ul_callback, + .id_table = mhi_net_id_table, + .driver = { + .name = "mhi_net", + .owner = THIS_MODULE, + }, +}; + +module_mhi_driver(mhi_net_driver); + +MODULE_AUTHOR("Loic Poulain "); +MODULE_DESCRIPTION("Network over MHI"); +MODULE_LICENSE("GPL v2"); diff --git a/drivers/net/mhi_net.c b/drivers/net/mhi_net.c deleted file mode 100644 index b92c2e1..0000000 --- a/drivers/net/mhi_net.c +++ /dev/null @@ -1,429 +0,0 @@ -// SPDX-License-Identifier: GPL-2.0-or-later -/* MHI Network driver - Network over MHI bus - * - * Copyright (C) 2020 Linaro Ltd - */ - -#include -#include -#include -#include -#include -#include -#include - -#define MHI_NET_MIN_MTU ETH_MIN_MTU -#define MHI_NET_MAX_MTU 0xffff -#define MHI_NET_DEFAULT_MTU 0x4000 - -struct mhi_net_stats { - u64_stats_t rx_packets; - u64_stats_t rx_bytes; - u64_stats_t rx_errors; - u64_stats_t rx_dropped; - u64_stats_t tx_packets; - u64_stats_t tx_bytes; - u64_stats_t tx_errors; - u64_stats_t tx_dropped; - struct u64_stats_sync tx_syncp; - struct u64_stats_sync rx_syncp; -}; - -struct mhi_net_dev { - struct mhi_device *mdev; - struct net_device *ndev; - struct sk_buff *skbagg_head; - struct sk_buff *skbagg_tail; - const struct mhi_net_proto *proto; - void *proto_data; - struct delayed_work rx_refill; - struct mhi_net_stats stats; - u32 rx_queue_sz; -}; - -struct mhi_net_proto { - int (*init)(struct mhi_net_dev *mhi_netdev); - struct sk_buff * (*tx_fixup)(struct mhi_net_dev *mhi_netdev, struct sk_buff *skb); - void (*rx)(struct mhi_net_dev *mhi_netdev, struct sk_buff *skb); -}; - -struct mhi_device_info { - const char *netname; - const struct mhi_net_proto *proto; -}; - -static int mhi_ndo_open(struct net_device *ndev) -{ - struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); - - /* Feed the rx buffer pool */ - schedule_delayed_work(&mhi_netdev->rx_refill, 0); - - /* Carrier is established via out-of-band channel (e.g. qmi) */ - netif_carrier_on(ndev); - - netif_start_queue(ndev); - - return 0; -} - -static int mhi_ndo_stop(struct net_device *ndev) -{ - struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); - - netif_stop_queue(ndev); - netif_carrier_off(ndev); - cancel_delayed_work_sync(&mhi_netdev->rx_refill); - - return 0; -} - -static int mhi_ndo_xmit(struct sk_buff *skb, struct net_device *ndev) -{ - struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); - const struct mhi_net_proto *proto = mhi_netdev->proto; - struct mhi_device *mdev = mhi_netdev->mdev; - int err; - - if (proto && proto->tx_fixup) { - skb = proto->tx_fixup(mhi_netdev, skb); - if (unlikely(!skb)) - goto exit_drop; - } - - err = mhi_queue_skb(mdev, DMA_TO_DEVICE, skb, skb->len, MHI_EOT); - if (unlikely(err)) { - net_err_ratelimited("%s: Failed to queue TX buf (%d)\n", - ndev->name, err); - dev_kfree_skb_any(skb); - goto exit_drop; - } - - if (mhi_queue_is_full(mdev, DMA_TO_DEVICE)) - netif_stop_queue(ndev); - - return NETDEV_TX_OK; - -exit_drop: - u64_stats_update_begin(&mhi_netdev->stats.tx_syncp); - u64_stats_inc(&mhi_netdev->stats.tx_dropped); - u64_stats_update_end(&mhi_netdev->stats.tx_syncp); - - return NETDEV_TX_OK; -} - -static void mhi_ndo_get_stats64(struct net_device *ndev, - struct rtnl_link_stats64 *stats) -{ - struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); - unsigned int start; - - do { - start = u64_stats_fetch_begin_irq(&mhi_netdev->stats.rx_syncp); - stats->rx_packets = u64_stats_read(&mhi_netdev->stats.rx_packets); - stats->rx_bytes = u64_stats_read(&mhi_netdev->stats.rx_bytes); - stats->rx_errors = u64_stats_read(&mhi_netdev->stats.rx_errors); - stats->rx_dropped = u64_stats_read(&mhi_netdev->stats.rx_dropped); - } while (u64_stats_fetch_retry_irq(&mhi_netdev->stats.rx_syncp, start)); - - do { - start = u64_stats_fetch_begin_irq(&mhi_netdev->stats.tx_syncp); - stats->tx_packets = u64_stats_read(&mhi_netdev->stats.tx_packets); - stats->tx_bytes = u64_stats_read(&mhi_netdev->stats.tx_bytes); - stats->tx_errors = u64_stats_read(&mhi_netdev->stats.tx_errors); - stats->tx_dropped = u64_stats_read(&mhi_netdev->stats.tx_dropped); - } while (u64_stats_fetch_retry_irq(&mhi_netdev->stats.tx_syncp, start)); -} - -static const struct net_device_ops mhi_netdev_ops = { - .ndo_open = mhi_ndo_open, - .ndo_stop = mhi_ndo_stop, - .ndo_start_xmit = mhi_ndo_xmit, - .ndo_get_stats64 = mhi_ndo_get_stats64, -}; - -static void mhi_net_setup(struct net_device *ndev) -{ - ndev->header_ops = NULL; /* No header */ - ndev->type = ARPHRD_RAWIP; - ndev->hard_header_len = 0; - ndev->addr_len = 0; - ndev->flags = IFF_POINTOPOINT | IFF_NOARP; - ndev->netdev_ops = &mhi_netdev_ops; - ndev->mtu = MHI_NET_DEFAULT_MTU; - ndev->min_mtu = MHI_NET_MIN_MTU; - ndev->max_mtu = MHI_NET_MAX_MTU; - ndev->tx_queue_len = 1000; -} - -static struct sk_buff *mhi_net_skb_agg(struct mhi_net_dev *mhi_netdev, - struct sk_buff *skb) -{ - struct sk_buff *head = mhi_netdev->skbagg_head; - struct sk_buff *tail = mhi_netdev->skbagg_tail; - - /* This is non-paged skb chaining using frag_list */ - if (!head) { - mhi_netdev->skbagg_head = skb; - return skb; - } - - if (!skb_shinfo(head)->frag_list) - skb_shinfo(head)->frag_list = skb; - else - tail->next = skb; - - head->len += skb->len; - head->data_len += skb->len; - head->truesize += skb->truesize; - - mhi_netdev->skbagg_tail = skb; - - return mhi_netdev->skbagg_head; -} - -static void mhi_net_dl_callback(struct mhi_device *mhi_dev, - struct mhi_result *mhi_res) -{ - struct mhi_net_dev *mhi_netdev = dev_get_drvdata(&mhi_dev->dev); - const struct mhi_net_proto *proto = mhi_netdev->proto; - struct sk_buff *skb = mhi_res->buf_addr; - int free_desc_count; - - free_desc_count = mhi_get_free_desc_count(mhi_dev, DMA_FROM_DEVICE); - - if (unlikely(mhi_res->transaction_status)) { - switch (mhi_res->transaction_status) { - case -EOVERFLOW: - /* Packet can not fit in one MHI buffer and has been - * split over multiple MHI transfers, do re-aggregation. - * That usually means the device side MTU is larger than - * the host side MTU/MRU. Since this is not optimal, - * print a warning (once). - */ - netdev_warn_once(mhi_netdev->ndev, - "Fragmented packets received, fix MTU?\n"); - skb_put(skb, mhi_res->bytes_xferd); - mhi_net_skb_agg(mhi_netdev, skb); - break; - case -ENOTCONN: - /* MHI layer stopping/resetting the DL channel */ - dev_kfree_skb_any(skb); - return; - default: - /* Unknown error, simply drop */ - dev_kfree_skb_any(skb); - u64_stats_update_begin(&mhi_netdev->stats.rx_syncp); - u64_stats_inc(&mhi_netdev->stats.rx_errors); - u64_stats_update_end(&mhi_netdev->stats.rx_syncp); - } - } else { - skb_put(skb, mhi_res->bytes_xferd); - - if (mhi_netdev->skbagg_head) { - /* Aggregate the final fragment */ - skb = mhi_net_skb_agg(mhi_netdev, skb); - mhi_netdev->skbagg_head = NULL; - } - - u64_stats_update_begin(&mhi_netdev->stats.rx_syncp); - u64_stats_inc(&mhi_netdev->stats.rx_packets); - u64_stats_add(&mhi_netdev->stats.rx_bytes, skb->len); - u64_stats_update_end(&mhi_netdev->stats.rx_syncp); - - switch (skb->data[0] & 0xf0) { - case 0x40: - skb->protocol = htons(ETH_P_IP); - break; - case 0x60: - skb->protocol = htons(ETH_P_IPV6); - break; - default: - skb->protocol = htons(ETH_P_MAP); - break; - } - - if (proto && proto->rx) - proto->rx(mhi_netdev, skb); - else - netif_rx(skb); - } - - /* Refill if RX buffers queue becomes low */ - if (free_desc_count >= mhi_netdev->rx_queue_sz / 2) - schedule_delayed_work(&mhi_netdev->rx_refill, 0); -} - -static void mhi_net_ul_callback(struct mhi_device *mhi_dev, - struct mhi_result *mhi_res) -{ - struct mhi_net_dev *mhi_netdev = dev_get_drvdata(&mhi_dev->dev); - struct net_device *ndev = mhi_netdev->ndev; - struct mhi_device *mdev = mhi_netdev->mdev; - struct sk_buff *skb = mhi_res->buf_addr; - - /* Hardware has consumed the buffer, so free the skb (which is not - * freed by the MHI stack) and perform accounting. - */ - dev_consume_skb_any(skb); - - u64_stats_update_begin(&mhi_netdev->stats.tx_syncp); - if (unlikely(mhi_res->transaction_status)) { - - /* MHI layer stopping/resetting the UL channel */ - if (mhi_res->transaction_status == -ENOTCONN) { - u64_stats_update_end(&mhi_netdev->stats.tx_syncp); - return; - } - - u64_stats_inc(&mhi_netdev->stats.tx_errors); - } else { - u64_stats_inc(&mhi_netdev->stats.tx_packets); - u64_stats_add(&mhi_netdev->stats.tx_bytes, mhi_res->bytes_xferd); - } - u64_stats_update_end(&mhi_netdev->stats.tx_syncp); - - if (netif_queue_stopped(ndev) && !mhi_queue_is_full(mdev, DMA_TO_DEVICE)) - netif_wake_queue(ndev); -} - -static void mhi_net_rx_refill_work(struct work_struct *work) -{ - struct mhi_net_dev *mhi_netdev = container_of(work, struct mhi_net_dev, - rx_refill.work); - struct net_device *ndev = mhi_netdev->ndev; - struct mhi_device *mdev = mhi_netdev->mdev; - int size = READ_ONCE(ndev->mtu); - struct sk_buff *skb; - int err; - - while (!mhi_queue_is_full(mdev, DMA_FROM_DEVICE)) { - skb = netdev_alloc_skb(ndev, size); - if (unlikely(!skb)) - break; - - err = mhi_queue_skb(mdev, DMA_FROM_DEVICE, skb, size, MHI_EOT); - if (unlikely(err)) { - net_err_ratelimited("%s: Failed to queue RX buf (%d)\n", - ndev->name, err); - kfree_skb(skb); - break; - } - - /* Do not hog the CPU if rx buffers are consumed faster than - * queued (unlikely). - */ - cond_resched(); - } - - /* If we're still starved of rx buffers, reschedule later */ - if (mhi_get_free_desc_count(mdev, DMA_FROM_DEVICE) == mhi_netdev->rx_queue_sz) - schedule_delayed_work(&mhi_netdev->rx_refill, HZ / 2); -} - -static struct device_type wwan_type = { - .name = "wwan", -}; - -static int mhi_net_probe(struct mhi_device *mhi_dev, - const struct mhi_device_id *id) -{ - const struct mhi_device_info *info = (struct mhi_device_info *)id->driver_data; - struct device *dev = &mhi_dev->dev; - struct mhi_net_dev *mhi_netdev; - struct net_device *ndev; - int err; - - ndev = alloc_netdev(sizeof(*mhi_netdev), info->netname, - NET_NAME_PREDICTABLE, mhi_net_setup); - if (!ndev) - return -ENOMEM; - - mhi_netdev = netdev_priv(ndev); - dev_set_drvdata(dev, mhi_netdev); - mhi_netdev->ndev = ndev; - mhi_netdev->mdev = mhi_dev; - mhi_netdev->skbagg_head = NULL; - mhi_netdev->proto = info->proto; - SET_NETDEV_DEV(ndev, &mhi_dev->dev); - SET_NETDEV_DEVTYPE(ndev, &wwan_type); - - INIT_DELAYED_WORK(&mhi_netdev->rx_refill, mhi_net_rx_refill_work); - u64_stats_init(&mhi_netdev->stats.rx_syncp); - u64_stats_init(&mhi_netdev->stats.tx_syncp); - - /* Start MHI channels */ - err = mhi_prepare_for_transfer(mhi_dev); - if (err) - goto out_err; - - /* Number of transfer descriptors determines size of the queue */ - mhi_netdev->rx_queue_sz = mhi_get_free_desc_count(mhi_dev, DMA_FROM_DEVICE); - - err = register_netdev(ndev); - if (err) - goto out_err; - - if (mhi_netdev->proto) { - err = mhi_netdev->proto->init(mhi_netdev); - if (err) - goto out_err_proto; - } - - return 0; - -out_err_proto: - unregister_netdev(ndev); -out_err: - free_netdev(ndev); - return err; -} - -static void mhi_net_remove(struct mhi_device *mhi_dev) -{ - struct mhi_net_dev *mhi_netdev = dev_get_drvdata(&mhi_dev->dev); - - unregister_netdev(mhi_netdev->ndev); - - mhi_unprepare_from_transfer(mhi_netdev->mdev); - - if (mhi_netdev->skbagg_head) - kfree_skb(mhi_netdev->skbagg_head); - - free_netdev(mhi_netdev->ndev); -} - -static const struct mhi_device_info mhi_hwip0 = { - .netname = "mhi_hwip%d", -}; - -static const struct mhi_device_info mhi_swip0 = { - .netname = "mhi_swip%d", -}; - -static const struct mhi_device_id mhi_net_id_table[] = { - /* Hardware accelerated data PATH (to modem IPA), protocol agnostic */ - { .chan = "IP_HW0", .driver_data = (kernel_ulong_t)&mhi_hwip0 }, - /* Software data PATH (to modem CPU) */ - { .chan = "IP_SW0", .driver_data = (kernel_ulong_t)&mhi_swip0 }, - {} -}; -MODULE_DEVICE_TABLE(mhi, mhi_net_id_table); - -static struct mhi_driver mhi_net_driver = { - .probe = mhi_net_probe, - .remove = mhi_net_remove, - .dl_xfer_cb = mhi_net_dl_callback, - .ul_xfer_cb = mhi_net_ul_callback, - .id_table = mhi_net_id_table, - .driver = { - .name = "mhi_net", - .owner = THIS_MODULE, - }, -}; - -module_mhi_driver(mhi_net_driver); - -MODULE_AUTHOR("Loic Poulain "); -MODULE_DESCRIPTION("Network over MHI"); -MODULE_LICENSE("GPL v2"); From patchwork Tue Feb 9 09:05:56 2021 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Loic Poulain X-Patchwork-Id: 12077543 X-Patchwork-Delegate: kuba@kernel.org Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-18.8 required=3.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_CR_TRAILER, INCLUDES_PATCH,MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED, USER_AGENT_GIT autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 1FDC2C433E0 for ; Tue, 9 Feb 2021 09:02:20 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id BE8DF64E99 for ; Tue, 9 Feb 2021 09:02:19 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S230227AbhBIJBk (ORCPT ); Tue, 9 Feb 2021 04:01:40 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:54870 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S230183AbhBII6q (ORCPT ); Tue, 9 Feb 2021 03:58:46 -0500 Received: from mail-wr1-x436.google.com (mail-wr1-x436.google.com [IPv6:2a00:1450:4864:20::436]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 4F9C8C06178C for ; Tue, 9 Feb 2021 00:58:00 -0800 (PST) Received: by mail-wr1-x436.google.com with SMTP id b3so20641702wrj.5 for ; Tue, 09 Feb 2021 00:58:00 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linaro.org; s=google; h=from:to:cc:subject:date:message-id:in-reply-to:references; bh=nUxAXdLjhkJNUqiBcTNdD1orykfV5lPBWV6f7eDBEEk=; b=og0zYNFuIfnOyzDZ+rF4cq10kfF2FN328svbr8HZVYuYvn332y0r8UUG8iPVwejzAU kais7gmM0lW1Rw6nigqK1urngPaCjqxbBZdAH10oHPSBgyYQiDJRKqNCZtYEO8UJctrV mP9OIxwaJPyjb4n4iftCSOYq9yxTiQtqiz6DNMzPab2mmC0kqkYuQ8hWfyodVQRSmsdy gIDOgHBRAqi95KOFuNxUMh7QHzQKOtdEyQOvwoqFpKpOx4a+vgZusL9SOfkL1C9BP70c WQ/u7rYLUXn+Bd1RqETSubtdAkL2BbbXXLGCw3iT0IMz+PU7hHeXy6lB8gIqdFVoIuAa /hiw== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references; bh=nUxAXdLjhkJNUqiBcTNdD1orykfV5lPBWV6f7eDBEEk=; b=qHuqqfSMzOqag2he2SwccjgvhB3LfAfHem9INCDOLNHtbVLc/1nO1BPW/o3bIQXbR1 7A6O0+3ieDf8dIe0Hbqk1h+H/XJ7QVwVrpgEon4mx6ZbeUBOV2uY9tepFYsh36RnGxSD 236FVzUgzBM8FmUW2j6VKwIpRkmnPxElxRkDnmK4ktL4jOfQcoHHoKqOP/JNlz9n/FtI MXfjskrH8H4+nHDbpPHUrMCkei2Y1cyr9T/Hl9QPUYwgKVAUBAC1SJhDZyC8qloHHANI kd3s6Mm23ZPwdM/09B5UAXDodCL+q/50tEWrQn+/Xe16FUvfK6FHt3Cwhb+C85Ixyd2a 5qyQ== X-Gm-Message-State: AOAM533LFLtiCRMNBJGYALaEoco/qfvh+uCVAOc8Kw7sXLsObnUdFUD7 nkj8OLi5Apx97a/jBqYkFvQAeQ== X-Google-Smtp-Source: ABdhPJxoAkNZE4VJncT3E3NsI9XK2XcG2SvLJ8EYTbV0odgHwffJaY/rMsfY9Br2UH2uK7ADPe7w5A== X-Received: by 2002:a05:6000:cd:: with SMTP id q13mr24186024wrx.138.1612861079021; Tue, 09 Feb 2021 00:57:59 -0800 (PST) Received: from localhost.localdomain ([88.122.66.28]) by smtp.gmail.com with ESMTPSA id d3sm38348693wrp.79.2021.02.09.00.57.58 (version=TLS1_2 cipher=ECDHE-ECDSA-AES128-GCM-SHA256 bits=128/128); Tue, 09 Feb 2021 00:57:58 -0800 (PST) From: Loic Poulain To: kuba@kernel.org, davem@davemloft.net Cc: netdev@vger.kernel.org, bjorn@mork.no, dcbw@redhat.com, carl.yin@quectel.com, mpearson@lenovo.com, cchen50@lenovo.com, jwjiang@lenovo.com, ivan.zhang@quectel.com, naveen.kumar@quectel.com, ivan.mikhanchuk@quectel.com, Loic Poulain Subject: [PATCH net-next v5 3/5] net: mhi: Create mhi.h Date: Tue, 9 Feb 2021 10:05:56 +0100 Message-Id: <1612861558-14487-4-git-send-email-loic.poulain@linaro.org> X-Mailer: git-send-email 2.7.4 In-Reply-To: <1612861558-14487-1-git-send-email-loic.poulain@linaro.org> References: <1612861558-14487-1-git-send-email-loic.poulain@linaro.org> Precedence: bulk List-ID: X-Mailing-List: netdev@vger.kernel.org X-Patchwork-Delegate: kuba@kernel.org Move mhi-net shared structures to mhi header, that will be used by upcoming proto(s). Signed-off-by: Loic Poulain --- drivers/net/mhi/mhi.h | 36 ++++++++++++++++++++++++++++++++++++ drivers/net/mhi/net.c | 33 ++------------------------------- 2 files changed, 38 insertions(+), 31 deletions(-) create mode 100644 drivers/net/mhi/mhi.h diff --git a/drivers/net/mhi/mhi.h b/drivers/net/mhi/mhi.h new file mode 100644 index 0000000..5050e4a --- /dev/null +++ b/drivers/net/mhi/mhi.h @@ -0,0 +1,36 @@ +/* SPDX-License-Identifier: GPL-2.0-or-later */ +/* MHI Network driver - Network over MHI bus + * + * Copyright (C) 2021 Linaro Ltd + */ + +struct mhi_net_stats { + u64_stats_t rx_packets; + u64_stats_t rx_bytes; + u64_stats_t rx_errors; + u64_stats_t rx_dropped; + u64_stats_t tx_packets; + u64_stats_t tx_bytes; + u64_stats_t tx_errors; + u64_stats_t tx_dropped; + struct u64_stats_sync tx_syncp; + struct u64_stats_sync rx_syncp; +}; + +struct mhi_net_dev { + struct mhi_device *mdev; + struct net_device *ndev; + struct sk_buff *skbagg_head; + struct sk_buff *skbagg_tail; + const struct mhi_net_proto *proto; + void *proto_data; + struct delayed_work rx_refill; + struct mhi_net_stats stats; + u32 rx_queue_sz; +}; + +struct mhi_net_proto { + int (*init)(struct mhi_net_dev *mhi_netdev); + struct sk_buff * (*tx_fixup)(struct mhi_net_dev *mhi_netdev, struct sk_buff *skb); + void (*rx)(struct mhi_net_dev *mhi_netdev, struct sk_buff *skb); +}; diff --git a/drivers/net/mhi/net.c b/drivers/net/mhi/net.c index b92c2e1..58b4b7c 100644 --- a/drivers/net/mhi/net.c +++ b/drivers/net/mhi/net.c @@ -12,41 +12,12 @@ #include #include +#include "mhi.h" + #define MHI_NET_MIN_MTU ETH_MIN_MTU #define MHI_NET_MAX_MTU 0xffff #define MHI_NET_DEFAULT_MTU 0x4000 -struct mhi_net_stats { - u64_stats_t rx_packets; - u64_stats_t rx_bytes; - u64_stats_t rx_errors; - u64_stats_t rx_dropped; - u64_stats_t tx_packets; - u64_stats_t tx_bytes; - u64_stats_t tx_errors; - u64_stats_t tx_dropped; - struct u64_stats_sync tx_syncp; - struct u64_stats_sync rx_syncp; -}; - -struct mhi_net_dev { - struct mhi_device *mdev; - struct net_device *ndev; - struct sk_buff *skbagg_head; - struct sk_buff *skbagg_tail; - const struct mhi_net_proto *proto; - void *proto_data; - struct delayed_work rx_refill; - struct mhi_net_stats stats; - u32 rx_queue_sz; -}; - -struct mhi_net_proto { - int (*init)(struct mhi_net_dev *mhi_netdev); - struct sk_buff * (*tx_fixup)(struct mhi_net_dev *mhi_netdev, struct sk_buff *skb); - void (*rx)(struct mhi_net_dev *mhi_netdev, struct sk_buff *skb); -}; - struct mhi_device_info { const char *netname; const struct mhi_net_proto *proto; From patchwork Tue Feb 9 09:05:57 2021 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Loic Poulain X-Patchwork-Id: 12077547 X-Patchwork-Delegate: kuba@kernel.org Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-18.8 required=3.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_CR_TRAILER, INCLUDES_PATCH,MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED, USER_AGENT_GIT autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 1788AC433DB for ; Tue, 9 Feb 2021 09:09:26 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id D696264EB9 for ; Tue, 9 Feb 2021 09:09:25 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S230479AbhBIJHt (ORCPT ); Tue, 9 Feb 2021 04:07:49 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:55276 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S230198AbhBIJAg (ORCPT ); Tue, 9 Feb 2021 04:00:36 -0500 Received: from mail-wr1-x42d.google.com (mail-wr1-x42d.google.com [IPv6:2a00:1450:4864:20::42d]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 63DD6C061793 for ; Tue, 9 Feb 2021 00:58:01 -0800 (PST) Received: by mail-wr1-x42d.google.com with SMTP id q7so20593532wre.13 for ; Tue, 09 Feb 2021 00:58:01 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linaro.org; s=google; h=from:to:cc:subject:date:message-id:in-reply-to:references; bh=d/vcUFElLAuaSN3VBWBjg1pNshUOyxf0nCgvwWAXx70=; b=R6rF8ziL4TxdR5RXMcmtNQq8ZQMT0sTTuBOzJ3Bh9mnHEvNX0YpjoHmaqojcWS26TB WZcRCIlA7HigJkSAOjybd35eM455fUycY7n58z2pQL7Q5cNXggQcGeY4bCmUvSpOeYxG CjKYFARbirAB4Ye3OlGmcZjYsyQIybj97QkPlDu/N7cwuZW/DZN4zspKDbMdm4ScmNqE lOOZi89JPwI6A2FKEuBqeKfiNndOmfnOCqvXFSlpK9Trbm7uv8HyAn9H9eEhENzdgBZM D2ZGA7AZyZOvMlgA+UwK/eYsTFMlAHw4XlxGc8nYmrDP+lVjnqqKU5ldn0i4drvIuCyG WY1g== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references; bh=d/vcUFElLAuaSN3VBWBjg1pNshUOyxf0nCgvwWAXx70=; b=Z1REGWJR7Gyl7uq8TmMjTAfnBDa3JzAM0O526u28Xolaj5nmp0Yw2jhGPIE7IpA7R+ 9zgoR6ylqptzP6y8zmiq+bhtlaDOtRfTmLNTLjjXVCCN1e5ZLFfDEemC4pwgOI7HQVFo J8Pu1FMKtw5AOTouT1BALxjwWnpwR1cpWv7gsCXsZA/dryiJSR9ldI1p7wMtYmYqIxQM f5SsH3r1HwTNvBr2317zspEuh+AT2anIdFHz5P5odQMfFmOHt03FD0g5mElZvQUPTpo1 gjVqa9GRbgDFntHr28/1PdAlLNWM7i4gfYqPKeznmdHXbK82uUhWO21Hs2dMf+uaCu0B ZFYw== X-Gm-Message-State: AOAM532XaGW25oeEgTmKNMU/e4kGKNKuuTiYMks84XDuihb+CVYARAsP PBGzgAhq3XrysHP44aOw2DQ4MQ== X-Google-Smtp-Source: ABdhPJx2XDoI+HnDtYH5TmjCJGXZ9e/L1dKwNM348mCd1VbXTMqQotTeLNIk2xOaiv1m4JiY+iEF0A== X-Received: by 2002:a5d:5010:: with SMTP id e16mr1995305wrt.202.1612861080034; Tue, 09 Feb 2021 00:58:00 -0800 (PST) Received: from localhost.localdomain ([88.122.66.28]) by smtp.gmail.com with ESMTPSA id d3sm38348693wrp.79.2021.02.09.00.57.59 (version=TLS1_2 cipher=ECDHE-ECDSA-AES128-GCM-SHA256 bits=128/128); Tue, 09 Feb 2021 00:57:59 -0800 (PST) From: Loic Poulain To: kuba@kernel.org, davem@davemloft.net Cc: netdev@vger.kernel.org, bjorn@mork.no, dcbw@redhat.com, carl.yin@quectel.com, mpearson@lenovo.com, cchen50@lenovo.com, jwjiang@lenovo.com, ivan.zhang@quectel.com, naveen.kumar@quectel.com, ivan.mikhanchuk@quectel.com, Loic Poulain Subject: [PATCH net-next v5 4/5] net: mhi: Add rx_length_errors stat Date: Tue, 9 Feb 2021 10:05:57 +0100 Message-Id: <1612861558-14487-5-git-send-email-loic.poulain@linaro.org> X-Mailer: git-send-email 2.7.4 In-Reply-To: <1612861558-14487-1-git-send-email-loic.poulain@linaro.org> References: <1612861558-14487-1-git-send-email-loic.poulain@linaro.org> Precedence: bulk List-ID: X-Mailing-List: netdev@vger.kernel.org X-Patchwork-Delegate: kuba@kernel.org This can be used by proto when packet len is incorrect. Signed-off-by: Loic Poulain --- drivers/net/mhi/mhi.h | 1 + drivers/net/mhi/net.c | 1 + 2 files changed, 2 insertions(+) diff --git a/drivers/net/mhi/mhi.h b/drivers/net/mhi/mhi.h index 5050e4a..82210e0 100644 --- a/drivers/net/mhi/mhi.h +++ b/drivers/net/mhi/mhi.h @@ -9,6 +9,7 @@ struct mhi_net_stats { u64_stats_t rx_bytes; u64_stats_t rx_errors; u64_stats_t rx_dropped; + u64_stats_t rx_length_errors; u64_stats_t tx_packets; u64_stats_t tx_bytes; u64_stats_t tx_errors; diff --git a/drivers/net/mhi/net.c b/drivers/net/mhi/net.c index 58b4b7c..44cbfb3 100644 --- a/drivers/net/mhi/net.c +++ b/drivers/net/mhi/net.c @@ -95,6 +95,7 @@ static void mhi_ndo_get_stats64(struct net_device *ndev, stats->rx_bytes = u64_stats_read(&mhi_netdev->stats.rx_bytes); stats->rx_errors = u64_stats_read(&mhi_netdev->stats.rx_errors); stats->rx_dropped = u64_stats_read(&mhi_netdev->stats.rx_dropped); + stats->rx_length_errors = u64_stats_read(&mhi_netdev->stats.rx_length_errors); } while (u64_stats_fetch_retry_irq(&mhi_netdev->stats.rx_syncp, start)); do { From patchwork Tue Feb 9 09:05:58 2021 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 8bit X-Patchwork-Submitter: Loic Poulain X-Patchwork-Id: 12077545 X-Patchwork-Delegate: kuba@kernel.org Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-18.8 required=3.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_CR_TRAILER, INCLUDES_PATCH,MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED, USER_AGENT_GIT autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 8B4A1C433DB for ; Tue, 9 Feb 2021 09:07:32 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 700DF64E99 for ; Tue, 9 Feb 2021 09:07:31 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S230382AbhBIJDc (ORCPT ); Tue, 9 Feb 2021 04:03:32 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:55278 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S230315AbhBIJAg (ORCPT ); Tue, 9 Feb 2021 04:00:36 -0500 Received: from mail-wr1-x434.google.com (mail-wr1-x434.google.com [IPv6:2a00:1450:4864:20::434]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 7F2EDC061794 for ; Tue, 9 Feb 2021 00:58:02 -0800 (PST) Received: by mail-wr1-x434.google.com with SMTP id v14so5066346wro.7 for ; Tue, 09 Feb 2021 00:58:02 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linaro.org; s=google; h=from:to:cc:subject:date:message-id:in-reply-to:references :mime-version:content-transfer-encoding; bh=zUJnTsvIoxdlhfi2ZGPj+BzO4F/bHluhePET0mqOh1w=; b=EwFTyvL2M/4PMZpaYqUsXlSveQ00uUUmaTWbFMWB5k5xpwvGDEfx2qrLgI/xUF3znE uzg5p6R1bNbEjaTSkRnHCVM3Nfu7EIGij8T/ZLIP1WJW+5d8mIKEbCHwp22LdQve5bDQ QCgWruu0HcVw+9mxfsuO3/30QYBYLFOsnqOiQeuLRIxOqxJlqzE7igZnTpuxXCAQ1EpD g37EKUHBg2uRbuTlis0TvInuwcgRt9FDQ18vpC5JwOKgreFdQb857pDJOSwHdL+hdc5E lW7adPshCl5FiT4FmZbPcB53wLck/A1FgrBPYsvBfZJDYCUVvyMQRCDfcJpjLOO0wdps 6Ejw== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references:mime-version:content-transfer-encoding; bh=zUJnTsvIoxdlhfi2ZGPj+BzO4F/bHluhePET0mqOh1w=; b=pJ/Te40s1WofSGK+OUQ/eqhWnkGgocltGBZdMe4VmlAvyeX2zjiS2z5xADnMaa7Uc2 01A5qNb/K6U5hVrwfBtWoM93Wf/OAwG6DFfckZv9Mg3U6pCiRxRx3iOvyDHEZc+rHQ0f sdDpe6J14OeHBM43DXHpF6zDOZr2O9H3m0HlK/c6SdIypVBFmLRDw15/VCiFuPW+EfoU SYrANyK1gICGiAtTLL6v6wf8hV5POx+8y9rSw0iRdsU4OJWb+aaq7OwaOYU03eg23kjT EfPjvxPHlvwtQSw4z+ImvFglKWK8mnPRPPPyU4J4sCyI2KgMRz08khMLO8rCDmmrLeMg SjBg== X-Gm-Message-State: AOAM533m/NPSa6Xx7S/1+H665oGAd8LW4SpgMBgzszhx+Nj5gLmIGHQY bm4kEh06By6/ZOIbObaCym7GNQ== X-Google-Smtp-Source: ABdhPJyfSuUavfSQMmUzvwe0NfmS4IIE5K40pLmW1VB7bJzplgmcGT3KKtkZ2wv7GwR+FyEKvDXTFw== X-Received: by 2002:adf:a41d:: with SMTP id d29mr11758132wra.196.1612861081175; Tue, 09 Feb 2021 00:58:01 -0800 (PST) Received: from localhost.localdomain ([88.122.66.28]) by smtp.gmail.com with ESMTPSA id d3sm38348693wrp.79.2021.02.09.00.58.00 (version=TLS1_2 cipher=ECDHE-ECDSA-AES128-GCM-SHA256 bits=128/128); Tue, 09 Feb 2021 00:58:00 -0800 (PST) From: Loic Poulain To: kuba@kernel.org, davem@davemloft.net Cc: netdev@vger.kernel.org, bjorn@mork.no, dcbw@redhat.com, carl.yin@quectel.com, mpearson@lenovo.com, cchen50@lenovo.com, jwjiang@lenovo.com, ivan.zhang@quectel.com, naveen.kumar@quectel.com, ivan.mikhanchuk@quectel.com, Loic Poulain Subject: [PATCH net-next v5 5/5] net: mhi: Add mbim proto Date: Tue, 9 Feb 2021 10:05:58 +0100 Message-Id: <1612861558-14487-6-git-send-email-loic.poulain@linaro.org> X-Mailer: git-send-email 2.7.4 In-Reply-To: <1612861558-14487-1-git-send-email-loic.poulain@linaro.org> References: <1612861558-14487-1-git-send-email-loic.poulain@linaro.org> MIME-Version: 1.0 Precedence: bulk List-ID: X-Mailing-List: netdev@vger.kernel.org X-Patchwork-Delegate: kuba@kernel.org MBIM has initially been specified by USB-IF for transporting data (IP) between a modem and a host over USB. However some modern modems also support MBIM over PCIe (via MHI). In the same way as QMAP(rmnet), it allows to aggregate IP packets and to perform context multiplexing. This change adds minimal MBIM data transport support to MHI, allowing to support MBIM only modems. MBIM being based on USB NCM, it reuses and copy some helpers/functions from the USB stack (cdc-ncm, cdc-mbim). Note that is a subset of the CDC-MBIM specification, supporting only transport of network data (IP), there is no support for DSS. Moreover the multi-session (for multi-pdn) is not supported in this initial version, but will be added latter, and aligned with the cdc-mbim solution (VLAN tags). This code has been inspired from the mhi_mbim downstream implementation (Carl Yin ). Signed-off-by: Loic Poulain --- drivers/net/mhi/Makefile | 2 +- drivers/net/mhi/mhi.h | 3 + drivers/net/mhi/net.c | 7 ++ drivers/net/mhi/proto_mbim.c | 293 +++++++++++++++++++++++++++++++++++++++++++ 4 files changed, 304 insertions(+), 1 deletion(-) create mode 100644 drivers/net/mhi/proto_mbim.c diff --git a/drivers/net/mhi/Makefile b/drivers/net/mhi/Makefile index 0acf989..f71b9f8 100644 --- a/drivers/net/mhi/Makefile +++ b/drivers/net/mhi/Makefile @@ -1,3 +1,3 @@ obj-$(CONFIG_MHI_NET) += mhi_net.o -mhi_net-y := net.o +mhi_net-y := net.o proto_mbim.o diff --git a/drivers/net/mhi/mhi.h b/drivers/net/mhi/mhi.h index 82210e0..12e7407 100644 --- a/drivers/net/mhi/mhi.h +++ b/drivers/net/mhi/mhi.h @@ -28,6 +28,7 @@ struct mhi_net_dev { struct delayed_work rx_refill; struct mhi_net_stats stats; u32 rx_queue_sz; + int msg_enable; }; struct mhi_net_proto { @@ -35,3 +36,5 @@ struct mhi_net_proto { struct sk_buff * (*tx_fixup)(struct mhi_net_dev *mhi_netdev, struct sk_buff *skb); void (*rx)(struct mhi_net_dev *mhi_netdev, struct sk_buff *skb); }; + +extern const struct mhi_net_proto proto_mbim; diff --git a/drivers/net/mhi/net.c b/drivers/net/mhi/net.c index 44cbfb3..f599608 100644 --- a/drivers/net/mhi/net.c +++ b/drivers/net/mhi/net.c @@ -373,11 +373,18 @@ static const struct mhi_device_info mhi_swip0 = { .netname = "mhi_swip%d", }; +static const struct mhi_device_info mhi_hwip0_mbim = { + .netname = "mhi_mbim%d", + .proto = &proto_mbim, +}; + static const struct mhi_device_id mhi_net_id_table[] = { /* Hardware accelerated data PATH (to modem IPA), protocol agnostic */ { .chan = "IP_HW0", .driver_data = (kernel_ulong_t)&mhi_hwip0 }, /* Software data PATH (to modem CPU) */ { .chan = "IP_SW0", .driver_data = (kernel_ulong_t)&mhi_swip0 }, + /* Hardware accelerated data PATH (to modem IPA), MBIM protocol */ + { .chan = "IP_HW0_MBIM", .driver_data = (kernel_ulong_t)&mhi_hwip0_mbim }, {} }; MODULE_DEVICE_TABLE(mhi, mhi_net_id_table); diff --git a/drivers/net/mhi/proto_mbim.c b/drivers/net/mhi/proto_mbim.c new file mode 100644 index 0000000..75b5484 --- /dev/null +++ b/drivers/net/mhi/proto_mbim.c @@ -0,0 +1,293 @@ +// SPDX-License-Identifier: GPL-2.0-or-later +/* MHI Network driver - Network over MHI bus + * + * Copyright (C) 2021 Linaro Ltd + * + * This driver copy some code from cdc_ncm, which is: + * Copyright (C) ST-Ericsson 2010-2012 + * and cdc_mbim, which is: + * Copyright (c) 2012 Smith Micro Software, Inc. + * Copyright (c) 2012 Bjørn Mork + * + */ + +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include + +#include "mhi.h" + +#define MBIM_NDP16_SIGN_MASK 0x00ffffff + +struct mbim_context { + u16 rx_seq; + u16 tx_seq; +}; + +static void __mbim_length_errors_inc(struct mhi_net_dev *dev) +{ + u64_stats_update_begin(&dev->stats.rx_syncp); + u64_stats_inc(&dev->stats.rx_length_errors); + u64_stats_update_end(&dev->stats.rx_syncp); +} + +static void __mbim_errors_inc(struct mhi_net_dev *dev) +{ + u64_stats_update_begin(&dev->stats.rx_syncp); + u64_stats_inc(&dev->stats.rx_errors); + u64_stats_update_end(&dev->stats.rx_syncp); +} + +static int mbim_rx_verify_nth16(struct sk_buff *skb) +{ + struct mhi_net_dev *dev = netdev_priv(skb->dev); + struct mbim_context *ctx = dev->proto_data; + struct usb_cdc_ncm_nth16 *nth16; + int len; + + if (skb->len < sizeof(struct usb_cdc_ncm_nth16) + + sizeof(struct usb_cdc_ncm_ndp16)) { + netif_dbg(dev, rx_err, dev->ndev, "frame too short\n"); + __mbim_length_errors_inc(dev); + return -EINVAL; + } + + nth16 = (struct usb_cdc_ncm_nth16 *)skb->data; + + if (nth16->dwSignature != cpu_to_le32(USB_CDC_NCM_NTH16_SIGN)) { + netif_dbg(dev, rx_err, dev->ndev, + "invalid NTH16 signature <%#010x>\n", + le32_to_cpu(nth16->dwSignature)); + __mbim_errors_inc(dev); + return -EINVAL; + } + + /* No limit on the block length, except the size of the data pkt */ + len = le16_to_cpu(nth16->wBlockLength); + if (len > skb->len) { + netif_dbg(dev, rx_err, dev->ndev, + "NTB does not fit into the skb %u/%u\n", len, + skb->len); + __mbim_length_errors_inc(dev); + return -EINVAL; + } + + if (ctx->rx_seq + 1 != le16_to_cpu(nth16->wSequence) && + (ctx->rx_seq || le16_to_cpu(nth16->wSequence)) && + !(ctx->rx_seq == 0xffff && !le16_to_cpu(nth16->wSequence))) { + netif_dbg(dev, rx_err, dev->ndev, + "sequence number glitch prev=%d curr=%d\n", + ctx->rx_seq, le16_to_cpu(nth16->wSequence)); + } + ctx->rx_seq = le16_to_cpu(nth16->wSequence); + + return le16_to_cpu(nth16->wNdpIndex); +} + +static int mbim_rx_verify_ndp16(struct sk_buff *skb, int ndpoffset) +{ + struct mhi_net_dev *dev = netdev_priv(skb->dev); + struct usb_cdc_ncm_ndp16 *ndp16; + int ret; + + if (ndpoffset + sizeof(struct usb_cdc_ncm_ndp16) > skb->len) { + netif_dbg(dev, rx_err, dev->ndev, "invalid NDP offset <%u>\n", + ndpoffset); + return -EINVAL; + } + + ndp16 = (struct usb_cdc_ncm_ndp16 *)(skb->data + ndpoffset); + + if (le16_to_cpu(ndp16->wLength) < USB_CDC_NCM_NDP16_LENGTH_MIN) { + netif_dbg(dev, rx_err, dev->ndev, "invalid DPT16 length <%u>\n", + le16_to_cpu(ndp16->wLength)); + return -EINVAL; + } + + ret = ((le16_to_cpu(ndp16->wLength) - sizeof(struct usb_cdc_ncm_ndp16)) + / sizeof(struct usb_cdc_ncm_dpe16)); + ret--; /* Last entry is always a NULL terminator */ + + if (sizeof(struct usb_cdc_ncm_ndp16) + + ret * sizeof(struct usb_cdc_ncm_dpe16) > skb->len) { + netif_dbg(dev, rx_err, dev->ndev, + "Invalid nframes = %d\n", ret); + return -EINVAL; + } + + return ret; +} + +static void mbim_rx(struct mhi_net_dev *mhi_netdev, struct sk_buff *skb) +{ + struct net_device *ndev = mhi_netdev->ndev; + int ndpoffset; + + if (skb_linearize(skb)) + goto error; + + /* Check NTB header and retrieve first NDP offset */ + ndpoffset = mbim_rx_verify_nth16(skb); + if (ndpoffset < 0) { + net_err_ratelimited("%s: Incorrect NTB header\n", ndev->name); + goto error; + } + + /* Process each NDP */ + while (1) { + struct usb_cdc_ncm_ndp16 *ndp16; + struct usb_cdc_ncm_dpe16 *dpe16; + int nframes, n; + + /* Check NDP header and retrieve number of datagrams */ + nframes = mbim_rx_verify_ndp16(skb, ndpoffset); + if (nframes < 0) { + net_err_ratelimited("%s: Incorrect NDP16\n", ndev->name); + __mbim_length_errors_inc(mhi_netdev); + goto error; + } + + /* Only IP data type supported, no DSS in MHI context */ + ndp16 = (struct usb_cdc_ncm_ndp16 *)(skb->data + ndpoffset); + if ((ndp16->dwSignature & cpu_to_le32(MBIM_NDP16_SIGN_MASK)) + != cpu_to_le32(USB_CDC_MBIM_NDP16_IPS_SIGN)) { + net_err_ratelimited("%s: Unsupported NDP type\n", ndev->name); + __mbim_errors_inc(mhi_netdev); + goto next_ndp; + } + + /* Only primary IP session 0 (0x00) supported for now */ + if (ndp16->dwSignature & ~cpu_to_le32(MBIM_NDP16_SIGN_MASK)) { + net_err_ratelimited("%s: bad packet session\n", ndev->name); + __mbim_errors_inc(mhi_netdev); + goto next_ndp; + } + + /* de-aggregate and deliver IP packets */ + dpe16 = ndp16->dpe16; + for (n = 0; n < nframes; n++, dpe16++) { + u16 dgram_offset = le16_to_cpu(dpe16->wDatagramIndex); + u16 dgram_len = le16_to_cpu(dpe16->wDatagramLength); + struct sk_buff *skbn; + + if (!dgram_offset || !dgram_len) + break; /* null terminator */ + + skbn = netdev_alloc_skb(ndev, dgram_len); + if (!skbn) + continue; + + skb_put(skbn, dgram_len); + memcpy(skbn->data, skb->data + dgram_offset, dgram_len); + + switch (skbn->data[0] & 0xf0) { + case 0x40: + skbn->protocol = htons(ETH_P_IP); + break; + case 0x60: + skbn->protocol = htons(ETH_P_IPV6); + break; + default: + net_err_ratelimited("%s: unknown protocol\n", + ndev->name); + __mbim_errors_inc(mhi_netdev); + dev_kfree_skb_any(skbn); + continue; + } + + netif_rx(skbn); + } +next_ndp: + /* Other NDP to process? */ + ndpoffset = (int)le16_to_cpu(ndp16->wNextNdpIndex); + if (!ndpoffset) + break; + } + + /* free skb */ + dev_consume_skb_any(skb); + return; +error: + dev_kfree_skb_any(skb); +} + +struct mbim_tx_hdr { + struct usb_cdc_ncm_nth16 nth16; + struct usb_cdc_ncm_ndp16 ndp16; + struct usb_cdc_ncm_dpe16 dpe16[2]; +} __packed; + +static struct sk_buff *mbim_tx_fixup(struct mhi_net_dev *mhi_netdev, + struct sk_buff *skb) +{ + struct mbim_context *ctx = mhi_netdev->proto_data; + unsigned int dgram_size = skb->len; + struct usb_cdc_ncm_nth16 *nth16; + struct usb_cdc_ncm_ndp16 *ndp16; + struct mbim_tx_hdr *mbim_hdr; + + /* For now, this is a partial implementation of CDC MBIM, only one NDP + * is sent, containing the IP packet (no aggregation). + */ + + /* Ensure we have enough headroom for crafting MBIM header */ + if (skb_cow_head(skb, sizeof(struct mbim_tx_hdr))) { + dev_kfree_skb_any(skb); + return NULL; + } + + mbim_hdr = skb_push(skb, sizeof(struct mbim_tx_hdr)); + + /* Fill NTB header */ + nth16 = &mbim_hdr->nth16; + nth16->dwSignature = cpu_to_le32(USB_CDC_NCM_NTH16_SIGN); + nth16->wHeaderLength = cpu_to_le16(sizeof(struct usb_cdc_ncm_nth16)); + nth16->wSequence = cpu_to_le16(ctx->tx_seq++); + nth16->wBlockLength = cpu_to_le16(skb->len); + nth16->wNdpIndex = cpu_to_le16(sizeof(struct usb_cdc_ncm_nth16)); + + /* Fill the unique NDP */ + ndp16 = &mbim_hdr->ndp16; + ndp16->dwSignature = cpu_to_le32(USB_CDC_MBIM_NDP16_IPS_SIGN); + ndp16->wLength = cpu_to_le16(sizeof(struct usb_cdc_ncm_ndp16) + + sizeof(struct usb_cdc_ncm_dpe16) * 2); + ndp16->wNextNdpIndex = 0; + + /* Datagram follows the mbim header */ + ndp16->dpe16[0].wDatagramIndex = cpu_to_le16(sizeof(struct mbim_tx_hdr)); + ndp16->dpe16[0].wDatagramLength = cpu_to_le16(dgram_size); + + /* null termination */ + ndp16->dpe16[1].wDatagramIndex = 0; + ndp16->dpe16[1].wDatagramLength = 0; + + return skb; +} + +static int mbim_init(struct mhi_net_dev *mhi_netdev) +{ + struct net_device *ndev = mhi_netdev->ndev; + + mhi_netdev->proto_data = devm_kzalloc(&ndev->dev, + sizeof(struct mbim_context), + GFP_KERNEL); + if (!mhi_netdev->proto_data) + return -ENOMEM; + + ndev->needed_headroom = sizeof(struct mbim_tx_hdr); + + return 0; +} + +const struct mhi_net_proto proto_mbim = { + .init = mbim_init, + .rx = mbim_rx, + .tx_fixup = mbim_tx_fixup, +};