From patchwork Mon Feb 1 21:05:40 2021 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Loic Poulain X-Patchwork-Id: 374289 Delivered-To: patch@linaro.org Received: by 2002:a17:906:24d5:0:0:0:0 with SMTP id f21csp4245880ejb; Mon, 1 Feb 2021 13:00:05 -0800 (PST) X-Google-Smtp-Source: ABdhPJzjgMEblB+kP4kUUXH3AshtVXmMOa/bdCwom8PfRuUz5btOql/ul5H6JEFXbACak8fPGvAs X-Received: by 2002:a05:6402:268a:: with SMTP id w10mr20847771edd.331.1612213205793; Mon, 01 Feb 2021 13:00:05 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1612213205; cv=none; d=google.com; s=arc-20160816; b=zO9ELkm0WnbynE6ZBw0ruJLTH/isZuAcb1U+oOeS53vuUPvyONbOCVZjjkDzxAnWI9 CDfujW1p6O3xgoVPWxbAZnbrh0nRWbNDhASNrVC/ORhpnqbK1iypDCv2Fuo1ob4v81pb PGuc8oO4PhXMawdrW+D8s4XQX2XBEEHvy6sHg1/Qe/zCBJ+PdjVBfjHUdVpfJxKn2Cf9 oQKLyheou/IlQxErc38mS+eBpe3qOFdpTvR+CHmgU9E9xmZUKd1po6tZpOfbIdP019o/ +EjtsFWanuuClBDNn3K7zHLJeqVkMRUdslYXqBrNnlF7Abbewa8U0YEhWxKiV7BS7+Ag 510Q== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:to:message-id:date:subject:cc:from :dkim-signature; bh=bOVZWokpQ33DTK5ssj5sCeUQAOwoz/k0b935F4WFbOQ=; b=QFCVZlNEo7DtecRnW/HNSsaDC2lWeMMvwDZQGGjTTwpY8sT+hm4KHl/rZEE9GCCkNj A3LimawdP1WoODg+HZFlJecuuTvZZ382yzBiHuygy9ORLZrzfU45MZdQFp8HnfuWECcv OFEUwR+u+RdFXklu73utKpyKIfhZCAivKazmkkWuxGw2OKiEIgR3/8D3uhHqjbi8WjeW 2QnAJx2xtwGRCR6ikru84ta3LzPdn9NnqmUEtWJDA42MW1KKC8mUxZ//PTPuBNZlNWEx eQXEHKKgCFHlgPD9a+sOYwr0atrbRSCN01TvsFK3IuJIn40JwU6j7MoZbR626xOC5czp 6rPA== ARC-Authentication-Results: i=1; mx.google.com; dkim=fail header.i=@linaro.org header.s=google header.b=X3TeoSJK; spf=pass (google.com: domain of netdev-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=netdev-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=linaro.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id a18si11311916ejr.276.2021.02.01.13.00.05; Mon, 01 Feb 2021 13:00:05 -0800 (PST) Received-SPF: pass (google.com: domain of netdev-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=fail header.i=@linaro.org header.s=google header.b=X3TeoSJK; spf=pass (google.com: domain of netdev-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=netdev-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=linaro.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S232009AbhBAU6t (ORCPT + 7 others); Mon, 1 Feb 2021 15:58:49 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:56064 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S231406AbhBAU6n (ORCPT ); Mon, 1 Feb 2021 15:58:43 -0500 Received: from mail-wr1-x430.google.com (mail-wr1-x430.google.com [IPv6:2a00:1450:4864:20::430]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 6D9B2C061573 for ; Mon, 1 Feb 2021 12:58:03 -0800 (PST) Received: by mail-wr1-x430.google.com with SMTP id a1so18150346wrq.6 for ; Mon, 01 Feb 2021 12:58:03 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linaro.org; s=google; h=from:to:cc:subject:date:message-id; bh=bOVZWokpQ33DTK5ssj5sCeUQAOwoz/k0b935F4WFbOQ=; b=X3TeoSJKKUrNEmpZMmPHXsCZ5lE6SX4DabvrmMMoYRkPoG+91T3/NGX5LuWPG8pJsC mBIMDjuneenMxt8l5IHCgWCx6qyXj6+dznSYo6o5oNsk1NaOZlsfFg0Iks5AeuhYuWsZ qxX3iGOIlA2C06W4TyRs+F0kGIKsVOIp7gsY2CsppFCstzEO61aaqfxZkG/Zmf6i5JGS 8Zy6aGHLLs4sLGnWA6omkCFGC91kZG0a/ipbXa9U0g0HwtQNm7Tmg3MXp81DYDi2timA MVElSgUGOiD21/KyCVu1omNRlsuZWHCY3m9ET1HFiAl7gdP+st4rvHy/2LlLA4qgiEqQ f4tA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:to:cc:subject:date:message-id; bh=bOVZWokpQ33DTK5ssj5sCeUQAOwoz/k0b935F4WFbOQ=; b=Ro0+/Z2qE/7DUV6FNWJG7fjWazUOeKsSMeaRjrg7nry+z4sSrCf2mku+vTa/ArvEx1 0YtgQSk9XnxcEEMnmXg8ILOShyKBr1P8gefvshIjY+fCmc3JBT7taUj49XWPmjRDXMK4 ra8Dn6w4kR+9zHDd8/2zaC4a/Gj4B34Vt0lJTopwSlDViHD46ljZn+IFj0hgnTXL6m5a C1IrCk3ThKnyROxOf19j9JmYiNr8pjprFu1f4L8SsgJEErMQ8HjU5UBQtRG3pqe17JqD gzvNTojBLy7or8PT1GxhIyjmX4uJcbDXGtwa7jtJrx+Dlj3phZr2vdqKTiT05iPEUjpR UFxQ== X-Gm-Message-State: AOAM531eBmgbzU2qD2qBBEA6K/oeyTtxMMLMSdcnbp699BBKx+4AHhmo FcE+UFHRYUv+kcBBxSDVDlQk5A== X-Received: by 2002:a5d:4d8d:: with SMTP id b13mr19321345wru.178.1612213082108; Mon, 01 Feb 2021 12:58:02 -0800 (PST) Received: from localhost.localdomain ([88.122.66.28]) by smtp.gmail.com with ESMTPSA id b7sm27894226wru.33.2021.02.01.12.58.01 (version=TLS1_2 cipher=ECDHE-ECDSA-AES128-GCM-SHA256 bits=128/128); Mon, 01 Feb 2021 12:58:01 -0800 (PST) From: Loic Poulain Cc: bjorn@mork.no, dcbw@redhat.com, netdev@vger.kernel.org, carl.yin@quectel.com, Loic Poulain Subject: [PATCH net-next v2 1/3] net: mhi: Add RX/TX fixup callbacks Date: Mon, 1 Feb 2021 22:05:40 +0100 Message-Id: <1612213542-17257-1-git-send-email-loic.poulain@linaro.org> X-Mailer: git-send-email 2.7.4 To: unlisted-recipients:; (no To-header on input) Precedence: bulk List-ID: X-Mailing-List: netdev@vger.kernel.org MHI can transport different protocols, some are handled at upper level, like IP and QMAP(rmnet/netlink), but others will need to be inside MHI net driver, like mbim. This change adds support for protocol rx/tx fixup callbacks registration, that can be used to encode/decode the targeted protocol. Signed-off-by: Loic Poulain --- v2: mhi_net_dev as rx/tx_fixup parameter drivers/net/mhi_net.c | 70 ++++++++++++++++++++++++++++++++++++++++++--------- 1 file changed, 58 insertions(+), 12 deletions(-) -- 2.7.4 diff --git a/drivers/net/mhi_net.c b/drivers/net/mhi_net.c index 4f512531..34d4bcf 100644 --- a/drivers/net/mhi_net.c +++ b/drivers/net/mhi_net.c @@ -32,11 +32,24 @@ struct mhi_net_stats { struct mhi_net_dev { struct mhi_device *mdev; struct net_device *ndev; + const struct mhi_net_proto *proto; + void *proto_data; struct delayed_work rx_refill; struct mhi_net_stats stats; u32 rx_queue_sz; }; +struct mhi_net_proto { + int (*init)(struct mhi_net_dev *mhi_netdev); + struct sk_buff * (*tx_fixup)(struct mhi_net_dev *mhi_netdev, struct sk_buff *skb); + int (*rx_fixup)(struct mhi_net_dev *mhi_netdev, struct sk_buff *skb); +}; + +struct mhi_device_info { + const char *netname; + const struct mhi_net_proto *proto; +}; + static int mhi_ndo_open(struct net_device *ndev) { struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); @@ -66,26 +79,35 @@ static int mhi_ndo_stop(struct net_device *ndev) static int mhi_ndo_xmit(struct sk_buff *skb, struct net_device *ndev) { struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); + const struct mhi_net_proto *proto = mhi_netdev->proto; struct mhi_device *mdev = mhi_netdev->mdev; int err; + if (proto && proto->tx_fixup) { + skb = proto->tx_fixup(mhi_netdev, skb); + if (unlikely(!skb)) + goto exit_drop; + } + err = mhi_queue_skb(mdev, DMA_TO_DEVICE, skb, skb->len, MHI_EOT); if (unlikely(err)) { net_err_ratelimited("%s: Failed to queue TX buf (%d)\n", ndev->name, err); - - u64_stats_update_begin(&mhi_netdev->stats.tx_syncp); - u64_stats_inc(&mhi_netdev->stats.tx_dropped); - u64_stats_update_end(&mhi_netdev->stats.tx_syncp); - - /* drop the packet */ dev_kfree_skb_any(skb); + goto exit_drop; } if (mhi_queue_is_full(mdev, DMA_TO_DEVICE)) netif_stop_queue(ndev); return NETDEV_TX_OK; + +exit_drop: + u64_stats_update_begin(&mhi_netdev->stats.tx_syncp); + u64_stats_inc(&mhi_netdev->stats.tx_dropped); + u64_stats_update_end(&mhi_netdev->stats.tx_syncp); + + return NETDEV_TX_OK; } static void mhi_ndo_get_stats64(struct net_device *ndev, @@ -136,6 +158,7 @@ static void mhi_net_dl_callback(struct mhi_device *mhi_dev, struct mhi_result *mhi_res) { struct mhi_net_dev *mhi_netdev = dev_get_drvdata(&mhi_dev->dev); + const struct mhi_net_proto *proto = mhi_netdev->proto; struct sk_buff *skb = mhi_res->buf_addr; int free_desc_count; @@ -170,7 +193,11 @@ static void mhi_net_dl_callback(struct mhi_device *mhi_dev, } skb_put(skb, mhi_res->bytes_xferd); - netif_rx(skb); + + if (proto && proto->rx_fixup) + proto->rx_fixup(mhi_netdev, skb); + else + netif_rx(skb); } /* Refill if RX buffers queue becomes low */ @@ -252,14 +279,14 @@ static struct device_type wwan_type = { static int mhi_net_probe(struct mhi_device *mhi_dev, const struct mhi_device_id *id) { - const char *netname = (char *)id->driver_data; + const struct mhi_device_info *info = (struct mhi_device_info *)id->driver_data; struct device *dev = &mhi_dev->dev; struct mhi_net_dev *mhi_netdev; struct net_device *ndev; int err; - ndev = alloc_netdev(sizeof(*mhi_netdev), netname, NET_NAME_PREDICTABLE, - mhi_net_setup); + ndev = alloc_netdev(sizeof(*mhi_netdev), info->netname, + NET_NAME_PREDICTABLE, mhi_net_setup); if (!ndev) return -ENOMEM; @@ -267,6 +294,7 @@ static int mhi_net_probe(struct mhi_device *mhi_dev, dev_set_drvdata(dev, mhi_netdev); mhi_netdev->ndev = ndev; mhi_netdev->mdev = mhi_dev; + mhi_netdev->proto = info->proto; SET_NETDEV_DEV(ndev, &mhi_dev->dev); SET_NETDEV_DEVTYPE(ndev, &wwan_type); @@ -286,8 +314,16 @@ static int mhi_net_probe(struct mhi_device *mhi_dev, if (err) goto out_err; + if (mhi_netdev->proto) { + err = mhi_netdev->proto->init(mhi_netdev); + if (err) + goto out_err_proto; + } + return 0; +out_err_proto: + unregister_netdev(ndev); out_err: free_netdev(ndev); return err; @@ -304,9 +340,19 @@ static void mhi_net_remove(struct mhi_device *mhi_dev) free_netdev(mhi_netdev->ndev); } +static const struct mhi_device_info mhi_hwip0 = { + .netname = "mhi_hwip%d", +}; + +static const struct mhi_device_info mhi_swip0 = { + .netname = "mhi_swip%d", +}; + static const struct mhi_device_id mhi_net_id_table[] = { - { .chan = "IP_HW0", .driver_data = (kernel_ulong_t)"mhi_hwip%d" }, - { .chan = "IP_SW0", .driver_data = (kernel_ulong_t)"mhi_swip%d" }, + /* Hardware accelerated data PATH (to modem IPA), protocol agnostic */ + { .chan = "IP_HW0", .driver_data = (kernel_ulong_t)&mhi_hwip0 }, + /* Software data PATH (to modem CPU) */ + { .chan = "IP_SW0", .driver_data = (kernel_ulong_t)&mhi_swip0 }, {} }; MODULE_DEVICE_TABLE(mhi, mhi_net_id_table); From patchwork Mon Feb 1 21:05:41 2021 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Loic Poulain X-Patchwork-Id: 374291 Delivered-To: patch@linaro.org Received: by 2002:a17:906:24d5:0:0:0:0 with SMTP id f21csp4245905ejb; Mon, 1 Feb 2021 13:00:07 -0800 (PST) X-Google-Smtp-Source: ABdhPJx8zvG3PlNtY7NrPvtn/sPT9pCNkTpyimOZqjcax+05C2LhjnLzsxPA/0b99BBLllTyZMqg X-Received: by 2002:a17:906:149a:: with SMTP id x26mr3258443ejc.486.1612213207603; Mon, 01 Feb 2021 13:00:07 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1612213207; cv=none; d=google.com; s=arc-20160816; b=xWaE3P+CHDm8xgXctEGQe3I8n9is7kE4Me3qG8po5R5ojPbOUFKSpTe14Uv3z99zdv wNbj7tx4EZ6qGBqQkpesJ3f9OhoB84SOqC5O18D4TaxpqflDM/IruINXtncpXZDHdaaY 6LmIY3JDGNg2tPpyhOPprN9MabMMtSBiwjyc+iZIfO9A9ooznnss0EGUqWN3aJPpKrdB xpEsY5+yWi9UcEcsGe+R0VsiDfjgjDwEsQiJHu6SS2b1nWLZRLDGgS7OhnSbuRSX2ji3 EN8m7FC9xc52I6s0qSo38/Nec3z4ZPQtAk3qVehK73WR4GLVpKSaBpZFCfzik+JQYfnA Zn9g== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:to:references:in-reply-to:message-id:date :subject:cc:from:dkim-signature; bh=icnzBpu/KppaOHF5xFnZ9z1YfAwotjEDKuTEkBI4UeE=; b=Vez7L912cMYvkBm6pLuJnjbd1pxaoAdRlKYj/9zWPG6EeSxHVrpgFmd6F4Uf5BDkcn IpDVe7YjCVxRJEeSUzUiVIbB051Wgl2ed7oOP8hFBEK85aO8RZi69paLJNISbz/lyNiu QfMGUeKcI9AbMf5w4fdsbYfA+9r87/GqK5nw6/+mFWu8/lpuPWMKcdwt7nk13kzywTIq lda8WCh0Qs7Ud+6mbwOGcR893/RH92qxJ/dBPJz2Eay37j7w1CQG1BkbktNfPm/B7S7J BhQNwxX8VEuJS3gkY6JRiXRdYUoQPntAep4SJDETI2GUANb5+5NzaTA3n0D3LpRrLjNt tAgg== ARC-Authentication-Results: i=1; mx.google.com; dkim=fail header.i=@linaro.org header.s=google header.b=lDXvwxJm; spf=pass (google.com: domain of netdev-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=netdev-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=linaro.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id a18si11311916ejr.276.2021.02.01.13.00.07; Mon, 01 Feb 2021 13:00:07 -0800 (PST) Received-SPF: pass (google.com: domain of netdev-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=fail header.i=@linaro.org header.s=google header.b=lDXvwxJm; spf=pass (google.com: domain of netdev-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=netdev-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=linaro.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S232094AbhBAU6x (ORCPT + 7 others); Mon, 1 Feb 2021 15:58:53 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:56074 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S231864AbhBAU6p (ORCPT ); Mon, 1 Feb 2021 15:58:45 -0500 Received: from mail-wr1-x432.google.com (mail-wr1-x432.google.com [IPv6:2a00:1450:4864:20::432]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id C4D7BC061756 for ; Mon, 1 Feb 2021 12:58:04 -0800 (PST) Received: by mail-wr1-x432.google.com with SMTP id v15so18167162wrx.4 for ; Mon, 01 Feb 2021 12:58:04 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linaro.org; s=google; h=from:to:cc:subject:date:message-id:in-reply-to:references; bh=icnzBpu/KppaOHF5xFnZ9z1YfAwotjEDKuTEkBI4UeE=; b=lDXvwxJmRYsCxRfcGesUnUgw2xyVQFuOjUNcz3FcfHFYfY99srtqZrk2dpIY4NieIO fy2B/HT14cshVaJEYvzoKctDModLhXG9OCskerbaMoo1DzXqtUMb+asZ+Qus5VkGyHy+ n+qQ4rISbr9M+km2dAAZTN9LS9Mg7+bF9++Lp/WYi5vDvtVmSkBondfk3/ISerE2qIWB 69I2kncSaeGvPA+PtchGYgynNin6jAfk/J1/Lvwn/ga22yTXJC4g+KFTiBZOaQ1loWQA m9NVq7yH86rHR/2S2wS3WGjwoOR9SToQrpR4JXPJtG3c/SVZQ2DQjDkAOZg9yTQ4lilU ph2Q== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references; bh=icnzBpu/KppaOHF5xFnZ9z1YfAwotjEDKuTEkBI4UeE=; b=Bs1qFZiFvpDVcwVtEYH1jfx4HDCzLnxHRjXov6s+VfnohbFC4+/TAMTaZXC4pUoZBy AHWtzMz18nva1GFrGJyoppEwsqeZFX4K5/oWZ8Ow6hUIDEKphjPFRhIXzGxeVvaMwrTz 9AFCrxxo+8nQLpobFesLQCUEYle36qRzb5bpwmDtKIDOZuHzodlqkSUFKXW1bil8Dic2 f/l0+GxaMlsN6eYXjEOk/CrAzqCFco44Hbt8H2cEGJ/LF0buYcDzhqqFSbVSYwT4UqEA /+D2HGfjcK8cZfQUvGiZLPGfJKrLxe/6js+xmZVQjsApbu/8ctFh+RmLeeVwJ4AVTY/w En6A== X-Gm-Message-State: AOAM531R467P4dXBnPlW9cm5tA1ajzegAhpnFdmujmkXhAhhPw1Tksu+ SEDcF+/ntyJHV/k4hKyPqIop4A== X-Received: by 2002:a5d:6044:: with SMTP id j4mr19880549wrt.287.1612213083376; Mon, 01 Feb 2021 12:58:03 -0800 (PST) Received: from localhost.localdomain ([88.122.66.28]) by smtp.gmail.com with ESMTPSA id b7sm27894226wru.33.2021.02.01.12.58.02 (version=TLS1_2 cipher=ECDHE-ECDSA-AES128-GCM-SHA256 bits=128/128); Mon, 01 Feb 2021 12:58:03 -0800 (PST) From: Loic Poulain Cc: bjorn@mork.no, dcbw@redhat.com, netdev@vger.kernel.org, carl.yin@quectel.com, Loic Poulain Subject: [PATCH net-next v2 2/3] net: mhi: Add dedicated folder Date: Mon, 1 Feb 2021 22:05:41 +0100 Message-Id: <1612213542-17257-2-git-send-email-loic.poulain@linaro.org> X-Mailer: git-send-email 2.7.4 In-Reply-To: <1612213542-17257-1-git-send-email-loic.poulain@linaro.org> References: <1612213542-17257-1-git-send-email-loic.poulain@linaro.org> To: unlisted-recipients:; (no To-header on input) Precedence: bulk List-ID: X-Mailing-List: netdev@vger.kernel.org Create a dedicated mhi directory for mhi-net, mhi-net is going to be split into differente files (for additional protocol support). Signed-off-by: Loic Poulain --- v2: no change drivers/net/Makefile | 2 +- drivers/net/mhi/Makefile | 3 + drivers/net/mhi/net.c | 376 +++++++++++++++++++++++++++++++++++++++++++++++ drivers/net/mhi_net.c | 376 ----------------------------------------------- 4 files changed, 380 insertions(+), 377 deletions(-) create mode 100644 drivers/net/mhi/Makefile create mode 100644 drivers/net/mhi/net.c delete mode 100644 drivers/net/mhi_net.c -- 2.7.4 diff --git a/drivers/net/Makefile b/drivers/net/Makefile index 36e2e41..f4990ff 100644 --- a/drivers/net/Makefile +++ b/drivers/net/Makefile @@ -36,7 +36,7 @@ obj-$(CONFIG_GTP) += gtp.o obj-$(CONFIG_NLMON) += nlmon.o obj-$(CONFIG_NET_VRF) += vrf.o obj-$(CONFIG_VSOCKMON) += vsockmon.o -obj-$(CONFIG_MHI_NET) += mhi_net.o +obj-$(CONFIG_MHI_NET) += mhi/ # # Networking Drivers diff --git a/drivers/net/mhi/Makefile b/drivers/net/mhi/Makefile new file mode 100644 index 0000000..0acf989 --- /dev/null +++ b/drivers/net/mhi/Makefile @@ -0,0 +1,3 @@ +obj-$(CONFIG_MHI_NET) += mhi_net.o + +mhi_net-y := net.o diff --git a/drivers/net/mhi/net.c b/drivers/net/mhi/net.c new file mode 100644 index 0000000..34d4bcf --- /dev/null +++ b/drivers/net/mhi/net.c @@ -0,0 +1,376 @@ +// SPDX-License-Identifier: GPL-2.0-or-later +/* MHI Network driver - Network over MHI bus + * + * Copyright (C) 2020 Linaro Ltd + */ + +#include +#include +#include +#include +#include +#include +#include + +#define MHI_NET_MIN_MTU ETH_MIN_MTU +#define MHI_NET_MAX_MTU 0xffff +#define MHI_NET_DEFAULT_MTU 0x4000 + +struct mhi_net_stats { + u64_stats_t rx_packets; + u64_stats_t rx_bytes; + u64_stats_t rx_errors; + u64_stats_t rx_dropped; + u64_stats_t tx_packets; + u64_stats_t tx_bytes; + u64_stats_t tx_errors; + u64_stats_t tx_dropped; + struct u64_stats_sync tx_syncp; + struct u64_stats_sync rx_syncp; +}; + +struct mhi_net_dev { + struct mhi_device *mdev; + struct net_device *ndev; + const struct mhi_net_proto *proto; + void *proto_data; + struct delayed_work rx_refill; + struct mhi_net_stats stats; + u32 rx_queue_sz; +}; + +struct mhi_net_proto { + int (*init)(struct mhi_net_dev *mhi_netdev); + struct sk_buff * (*tx_fixup)(struct mhi_net_dev *mhi_netdev, struct sk_buff *skb); + int (*rx_fixup)(struct mhi_net_dev *mhi_netdev, struct sk_buff *skb); +}; + +struct mhi_device_info { + const char *netname; + const struct mhi_net_proto *proto; +}; + +static int mhi_ndo_open(struct net_device *ndev) +{ + struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); + + /* Feed the rx buffer pool */ + schedule_delayed_work(&mhi_netdev->rx_refill, 0); + + /* Carrier is established via out-of-band channel (e.g. qmi) */ + netif_carrier_on(ndev); + + netif_start_queue(ndev); + + return 0; +} + +static int mhi_ndo_stop(struct net_device *ndev) +{ + struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); + + netif_stop_queue(ndev); + netif_carrier_off(ndev); + cancel_delayed_work_sync(&mhi_netdev->rx_refill); + + return 0; +} + +static int mhi_ndo_xmit(struct sk_buff *skb, struct net_device *ndev) +{ + struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); + const struct mhi_net_proto *proto = mhi_netdev->proto; + struct mhi_device *mdev = mhi_netdev->mdev; + int err; + + if (proto && proto->tx_fixup) { + skb = proto->tx_fixup(mhi_netdev, skb); + if (unlikely(!skb)) + goto exit_drop; + } + + err = mhi_queue_skb(mdev, DMA_TO_DEVICE, skb, skb->len, MHI_EOT); + if (unlikely(err)) { + net_err_ratelimited("%s: Failed to queue TX buf (%d)\n", + ndev->name, err); + dev_kfree_skb_any(skb); + goto exit_drop; + } + + if (mhi_queue_is_full(mdev, DMA_TO_DEVICE)) + netif_stop_queue(ndev); + + return NETDEV_TX_OK; + +exit_drop: + u64_stats_update_begin(&mhi_netdev->stats.tx_syncp); + u64_stats_inc(&mhi_netdev->stats.tx_dropped); + u64_stats_update_end(&mhi_netdev->stats.tx_syncp); + + return NETDEV_TX_OK; +} + +static void mhi_ndo_get_stats64(struct net_device *ndev, + struct rtnl_link_stats64 *stats) +{ + struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); + unsigned int start; + + do { + start = u64_stats_fetch_begin_irq(&mhi_netdev->stats.rx_syncp); + stats->rx_packets = u64_stats_read(&mhi_netdev->stats.rx_packets); + stats->rx_bytes = u64_stats_read(&mhi_netdev->stats.rx_bytes); + stats->rx_errors = u64_stats_read(&mhi_netdev->stats.rx_errors); + stats->rx_dropped = u64_stats_read(&mhi_netdev->stats.rx_dropped); + } while (u64_stats_fetch_retry_irq(&mhi_netdev->stats.rx_syncp, start)); + + do { + start = u64_stats_fetch_begin_irq(&mhi_netdev->stats.tx_syncp); + stats->tx_packets = u64_stats_read(&mhi_netdev->stats.tx_packets); + stats->tx_bytes = u64_stats_read(&mhi_netdev->stats.tx_bytes); + stats->tx_errors = u64_stats_read(&mhi_netdev->stats.tx_errors); + stats->tx_dropped = u64_stats_read(&mhi_netdev->stats.tx_dropped); + } while (u64_stats_fetch_retry_irq(&mhi_netdev->stats.tx_syncp, start)); +} + +static const struct net_device_ops mhi_netdev_ops = { + .ndo_open = mhi_ndo_open, + .ndo_stop = mhi_ndo_stop, + .ndo_start_xmit = mhi_ndo_xmit, + .ndo_get_stats64 = mhi_ndo_get_stats64, +}; + +static void mhi_net_setup(struct net_device *ndev) +{ + ndev->header_ops = NULL; /* No header */ + ndev->type = ARPHRD_RAWIP; + ndev->hard_header_len = 0; + ndev->addr_len = 0; + ndev->flags = IFF_POINTOPOINT | IFF_NOARP; + ndev->netdev_ops = &mhi_netdev_ops; + ndev->mtu = MHI_NET_DEFAULT_MTU; + ndev->min_mtu = MHI_NET_MIN_MTU; + ndev->max_mtu = MHI_NET_MAX_MTU; + ndev->tx_queue_len = 1000; +} + +static void mhi_net_dl_callback(struct mhi_device *mhi_dev, + struct mhi_result *mhi_res) +{ + struct mhi_net_dev *mhi_netdev = dev_get_drvdata(&mhi_dev->dev); + const struct mhi_net_proto *proto = mhi_netdev->proto; + struct sk_buff *skb = mhi_res->buf_addr; + int free_desc_count; + + free_desc_count = mhi_get_free_desc_count(mhi_dev, DMA_FROM_DEVICE); + + if (unlikely(mhi_res->transaction_status)) { + dev_kfree_skb_any(skb); + + /* MHI layer stopping/resetting the DL channel */ + if (mhi_res->transaction_status == -ENOTCONN) + return; + + u64_stats_update_begin(&mhi_netdev->stats.rx_syncp); + u64_stats_inc(&mhi_netdev->stats.rx_errors); + u64_stats_update_end(&mhi_netdev->stats.rx_syncp); + } else { + u64_stats_update_begin(&mhi_netdev->stats.rx_syncp); + u64_stats_inc(&mhi_netdev->stats.rx_packets); + u64_stats_add(&mhi_netdev->stats.rx_bytes, mhi_res->bytes_xferd); + u64_stats_update_end(&mhi_netdev->stats.rx_syncp); + + switch (skb->data[0] & 0xf0) { + case 0x40: + skb->protocol = htons(ETH_P_IP); + break; + case 0x60: + skb->protocol = htons(ETH_P_IPV6); + break; + default: + skb->protocol = htons(ETH_P_MAP); + break; + } + + skb_put(skb, mhi_res->bytes_xferd); + + if (proto && proto->rx_fixup) + proto->rx_fixup(mhi_netdev, skb); + else + netif_rx(skb); + } + + /* Refill if RX buffers queue becomes low */ + if (free_desc_count >= mhi_netdev->rx_queue_sz / 2) + schedule_delayed_work(&mhi_netdev->rx_refill, 0); +} + +static void mhi_net_ul_callback(struct mhi_device *mhi_dev, + struct mhi_result *mhi_res) +{ + struct mhi_net_dev *mhi_netdev = dev_get_drvdata(&mhi_dev->dev); + struct net_device *ndev = mhi_netdev->ndev; + struct mhi_device *mdev = mhi_netdev->mdev; + struct sk_buff *skb = mhi_res->buf_addr; + + /* Hardware has consumed the buffer, so free the skb (which is not + * freed by the MHI stack) and perform accounting. + */ + dev_consume_skb_any(skb); + + u64_stats_update_begin(&mhi_netdev->stats.tx_syncp); + if (unlikely(mhi_res->transaction_status)) { + + /* MHI layer stopping/resetting the UL channel */ + if (mhi_res->transaction_status == -ENOTCONN) { + u64_stats_update_end(&mhi_netdev->stats.tx_syncp); + return; + } + + u64_stats_inc(&mhi_netdev->stats.tx_errors); + } else { + u64_stats_inc(&mhi_netdev->stats.tx_packets); + u64_stats_add(&mhi_netdev->stats.tx_bytes, mhi_res->bytes_xferd); + } + u64_stats_update_end(&mhi_netdev->stats.tx_syncp); + + if (netif_queue_stopped(ndev) && !mhi_queue_is_full(mdev, DMA_TO_DEVICE)) + netif_wake_queue(ndev); +} + +static void mhi_net_rx_refill_work(struct work_struct *work) +{ + struct mhi_net_dev *mhi_netdev = container_of(work, struct mhi_net_dev, + rx_refill.work); + struct net_device *ndev = mhi_netdev->ndev; + struct mhi_device *mdev = mhi_netdev->mdev; + int size = READ_ONCE(ndev->mtu); + struct sk_buff *skb; + int err; + + while (!mhi_queue_is_full(mdev, DMA_FROM_DEVICE)) { + skb = netdev_alloc_skb(ndev, size); + if (unlikely(!skb)) + break; + + err = mhi_queue_skb(mdev, DMA_FROM_DEVICE, skb, size, MHI_EOT); + if (unlikely(err)) { + net_err_ratelimited("%s: Failed to queue RX buf (%d)\n", + ndev->name, err); + kfree_skb(skb); + break; + } + + /* Do not hog the CPU if rx buffers are consumed faster than + * queued (unlikely). + */ + cond_resched(); + } + + /* If we're still starved of rx buffers, reschedule later */ + if (mhi_get_free_desc_count(mdev, DMA_FROM_DEVICE) == mhi_netdev->rx_queue_sz) + schedule_delayed_work(&mhi_netdev->rx_refill, HZ / 2); +} + +static struct device_type wwan_type = { + .name = "wwan", +}; + +static int mhi_net_probe(struct mhi_device *mhi_dev, + const struct mhi_device_id *id) +{ + const struct mhi_device_info *info = (struct mhi_device_info *)id->driver_data; + struct device *dev = &mhi_dev->dev; + struct mhi_net_dev *mhi_netdev; + struct net_device *ndev; + int err; + + ndev = alloc_netdev(sizeof(*mhi_netdev), info->netname, + NET_NAME_PREDICTABLE, mhi_net_setup); + if (!ndev) + return -ENOMEM; + + mhi_netdev = netdev_priv(ndev); + dev_set_drvdata(dev, mhi_netdev); + mhi_netdev->ndev = ndev; + mhi_netdev->mdev = mhi_dev; + mhi_netdev->proto = info->proto; + SET_NETDEV_DEV(ndev, &mhi_dev->dev); + SET_NETDEV_DEVTYPE(ndev, &wwan_type); + + INIT_DELAYED_WORK(&mhi_netdev->rx_refill, mhi_net_rx_refill_work); + u64_stats_init(&mhi_netdev->stats.rx_syncp); + u64_stats_init(&mhi_netdev->stats.tx_syncp); + + /* Start MHI channels */ + err = mhi_prepare_for_transfer(mhi_dev); + if (err) + goto out_err; + + /* Number of transfer descriptors determines size of the queue */ + mhi_netdev->rx_queue_sz = mhi_get_free_desc_count(mhi_dev, DMA_FROM_DEVICE); + + err = register_netdev(ndev); + if (err) + goto out_err; + + if (mhi_netdev->proto) { + err = mhi_netdev->proto->init(mhi_netdev); + if (err) + goto out_err_proto; + } + + return 0; + +out_err_proto: + unregister_netdev(ndev); +out_err: + free_netdev(ndev); + return err; +} + +static void mhi_net_remove(struct mhi_device *mhi_dev) +{ + struct mhi_net_dev *mhi_netdev = dev_get_drvdata(&mhi_dev->dev); + + unregister_netdev(mhi_netdev->ndev); + + mhi_unprepare_from_transfer(mhi_netdev->mdev); + + free_netdev(mhi_netdev->ndev); +} + +static const struct mhi_device_info mhi_hwip0 = { + .netname = "mhi_hwip%d", +}; + +static const struct mhi_device_info mhi_swip0 = { + .netname = "mhi_swip%d", +}; + +static const struct mhi_device_id mhi_net_id_table[] = { + /* Hardware accelerated data PATH (to modem IPA), protocol agnostic */ + { .chan = "IP_HW0", .driver_data = (kernel_ulong_t)&mhi_hwip0 }, + /* Software data PATH (to modem CPU) */ + { .chan = "IP_SW0", .driver_data = (kernel_ulong_t)&mhi_swip0 }, + {} +}; +MODULE_DEVICE_TABLE(mhi, mhi_net_id_table); + +static struct mhi_driver mhi_net_driver = { + .probe = mhi_net_probe, + .remove = mhi_net_remove, + .dl_xfer_cb = mhi_net_dl_callback, + .ul_xfer_cb = mhi_net_ul_callback, + .id_table = mhi_net_id_table, + .driver = { + .name = "mhi_net", + .owner = THIS_MODULE, + }, +}; + +module_mhi_driver(mhi_net_driver); + +MODULE_AUTHOR("Loic Poulain "); +MODULE_DESCRIPTION("Network over MHI"); +MODULE_LICENSE("GPL v2"); diff --git a/drivers/net/mhi_net.c b/drivers/net/mhi_net.c deleted file mode 100644 index 34d4bcf..0000000 --- a/drivers/net/mhi_net.c +++ /dev/null @@ -1,376 +0,0 @@ -// SPDX-License-Identifier: GPL-2.0-or-later -/* MHI Network driver - Network over MHI bus - * - * Copyright (C) 2020 Linaro Ltd - */ - -#include -#include -#include -#include -#include -#include -#include - -#define MHI_NET_MIN_MTU ETH_MIN_MTU -#define MHI_NET_MAX_MTU 0xffff -#define MHI_NET_DEFAULT_MTU 0x4000 - -struct mhi_net_stats { - u64_stats_t rx_packets; - u64_stats_t rx_bytes; - u64_stats_t rx_errors; - u64_stats_t rx_dropped; - u64_stats_t tx_packets; - u64_stats_t tx_bytes; - u64_stats_t tx_errors; - u64_stats_t tx_dropped; - struct u64_stats_sync tx_syncp; - struct u64_stats_sync rx_syncp; -}; - -struct mhi_net_dev { - struct mhi_device *mdev; - struct net_device *ndev; - const struct mhi_net_proto *proto; - void *proto_data; - struct delayed_work rx_refill; - struct mhi_net_stats stats; - u32 rx_queue_sz; -}; - -struct mhi_net_proto { - int (*init)(struct mhi_net_dev *mhi_netdev); - struct sk_buff * (*tx_fixup)(struct mhi_net_dev *mhi_netdev, struct sk_buff *skb); - int (*rx_fixup)(struct mhi_net_dev *mhi_netdev, struct sk_buff *skb); -}; - -struct mhi_device_info { - const char *netname; - const struct mhi_net_proto *proto; -}; - -static int mhi_ndo_open(struct net_device *ndev) -{ - struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); - - /* Feed the rx buffer pool */ - schedule_delayed_work(&mhi_netdev->rx_refill, 0); - - /* Carrier is established via out-of-band channel (e.g. qmi) */ - netif_carrier_on(ndev); - - netif_start_queue(ndev); - - return 0; -} - -static int mhi_ndo_stop(struct net_device *ndev) -{ - struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); - - netif_stop_queue(ndev); - netif_carrier_off(ndev); - cancel_delayed_work_sync(&mhi_netdev->rx_refill); - - return 0; -} - -static int mhi_ndo_xmit(struct sk_buff *skb, struct net_device *ndev) -{ - struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); - const struct mhi_net_proto *proto = mhi_netdev->proto; - struct mhi_device *mdev = mhi_netdev->mdev; - int err; - - if (proto && proto->tx_fixup) { - skb = proto->tx_fixup(mhi_netdev, skb); - if (unlikely(!skb)) - goto exit_drop; - } - - err = mhi_queue_skb(mdev, DMA_TO_DEVICE, skb, skb->len, MHI_EOT); - if (unlikely(err)) { - net_err_ratelimited("%s: Failed to queue TX buf (%d)\n", - ndev->name, err); - dev_kfree_skb_any(skb); - goto exit_drop; - } - - if (mhi_queue_is_full(mdev, DMA_TO_DEVICE)) - netif_stop_queue(ndev); - - return NETDEV_TX_OK; - -exit_drop: - u64_stats_update_begin(&mhi_netdev->stats.tx_syncp); - u64_stats_inc(&mhi_netdev->stats.tx_dropped); - u64_stats_update_end(&mhi_netdev->stats.tx_syncp); - - return NETDEV_TX_OK; -} - -static void mhi_ndo_get_stats64(struct net_device *ndev, - struct rtnl_link_stats64 *stats) -{ - struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); - unsigned int start; - - do { - start = u64_stats_fetch_begin_irq(&mhi_netdev->stats.rx_syncp); - stats->rx_packets = u64_stats_read(&mhi_netdev->stats.rx_packets); - stats->rx_bytes = u64_stats_read(&mhi_netdev->stats.rx_bytes); - stats->rx_errors = u64_stats_read(&mhi_netdev->stats.rx_errors); - stats->rx_dropped = u64_stats_read(&mhi_netdev->stats.rx_dropped); - } while (u64_stats_fetch_retry_irq(&mhi_netdev->stats.rx_syncp, start)); - - do { - start = u64_stats_fetch_begin_irq(&mhi_netdev->stats.tx_syncp); - stats->tx_packets = u64_stats_read(&mhi_netdev->stats.tx_packets); - stats->tx_bytes = u64_stats_read(&mhi_netdev->stats.tx_bytes); - stats->tx_errors = u64_stats_read(&mhi_netdev->stats.tx_errors); - stats->tx_dropped = u64_stats_read(&mhi_netdev->stats.tx_dropped); - } while (u64_stats_fetch_retry_irq(&mhi_netdev->stats.tx_syncp, start)); -} - -static const struct net_device_ops mhi_netdev_ops = { - .ndo_open = mhi_ndo_open, - .ndo_stop = mhi_ndo_stop, - .ndo_start_xmit = mhi_ndo_xmit, - .ndo_get_stats64 = mhi_ndo_get_stats64, -}; - -static void mhi_net_setup(struct net_device *ndev) -{ - ndev->header_ops = NULL; /* No header */ - ndev->type = ARPHRD_RAWIP; - ndev->hard_header_len = 0; - ndev->addr_len = 0; - ndev->flags = IFF_POINTOPOINT | IFF_NOARP; - ndev->netdev_ops = &mhi_netdev_ops; - ndev->mtu = MHI_NET_DEFAULT_MTU; - ndev->min_mtu = MHI_NET_MIN_MTU; - ndev->max_mtu = MHI_NET_MAX_MTU; - ndev->tx_queue_len = 1000; -} - -static void mhi_net_dl_callback(struct mhi_device *mhi_dev, - struct mhi_result *mhi_res) -{ - struct mhi_net_dev *mhi_netdev = dev_get_drvdata(&mhi_dev->dev); - const struct mhi_net_proto *proto = mhi_netdev->proto; - struct sk_buff *skb = mhi_res->buf_addr; - int free_desc_count; - - free_desc_count = mhi_get_free_desc_count(mhi_dev, DMA_FROM_DEVICE); - - if (unlikely(mhi_res->transaction_status)) { - dev_kfree_skb_any(skb); - - /* MHI layer stopping/resetting the DL channel */ - if (mhi_res->transaction_status == -ENOTCONN) - return; - - u64_stats_update_begin(&mhi_netdev->stats.rx_syncp); - u64_stats_inc(&mhi_netdev->stats.rx_errors); - u64_stats_update_end(&mhi_netdev->stats.rx_syncp); - } else { - u64_stats_update_begin(&mhi_netdev->stats.rx_syncp); - u64_stats_inc(&mhi_netdev->stats.rx_packets); - u64_stats_add(&mhi_netdev->stats.rx_bytes, mhi_res->bytes_xferd); - u64_stats_update_end(&mhi_netdev->stats.rx_syncp); - - switch (skb->data[0] & 0xf0) { - case 0x40: - skb->protocol = htons(ETH_P_IP); - break; - case 0x60: - skb->protocol = htons(ETH_P_IPV6); - break; - default: - skb->protocol = htons(ETH_P_MAP); - break; - } - - skb_put(skb, mhi_res->bytes_xferd); - - if (proto && proto->rx_fixup) - proto->rx_fixup(mhi_netdev, skb); - else - netif_rx(skb); - } - - /* Refill if RX buffers queue becomes low */ - if (free_desc_count >= mhi_netdev->rx_queue_sz / 2) - schedule_delayed_work(&mhi_netdev->rx_refill, 0); -} - -static void mhi_net_ul_callback(struct mhi_device *mhi_dev, - struct mhi_result *mhi_res) -{ - struct mhi_net_dev *mhi_netdev = dev_get_drvdata(&mhi_dev->dev); - struct net_device *ndev = mhi_netdev->ndev; - struct mhi_device *mdev = mhi_netdev->mdev; - struct sk_buff *skb = mhi_res->buf_addr; - - /* Hardware has consumed the buffer, so free the skb (which is not - * freed by the MHI stack) and perform accounting. - */ - dev_consume_skb_any(skb); - - u64_stats_update_begin(&mhi_netdev->stats.tx_syncp); - if (unlikely(mhi_res->transaction_status)) { - - /* MHI layer stopping/resetting the UL channel */ - if (mhi_res->transaction_status == -ENOTCONN) { - u64_stats_update_end(&mhi_netdev->stats.tx_syncp); - return; - } - - u64_stats_inc(&mhi_netdev->stats.tx_errors); - } else { - u64_stats_inc(&mhi_netdev->stats.tx_packets); - u64_stats_add(&mhi_netdev->stats.tx_bytes, mhi_res->bytes_xferd); - } - u64_stats_update_end(&mhi_netdev->stats.tx_syncp); - - if (netif_queue_stopped(ndev) && !mhi_queue_is_full(mdev, DMA_TO_DEVICE)) - netif_wake_queue(ndev); -} - -static void mhi_net_rx_refill_work(struct work_struct *work) -{ - struct mhi_net_dev *mhi_netdev = container_of(work, struct mhi_net_dev, - rx_refill.work); - struct net_device *ndev = mhi_netdev->ndev; - struct mhi_device *mdev = mhi_netdev->mdev; - int size = READ_ONCE(ndev->mtu); - struct sk_buff *skb; - int err; - - while (!mhi_queue_is_full(mdev, DMA_FROM_DEVICE)) { - skb = netdev_alloc_skb(ndev, size); - if (unlikely(!skb)) - break; - - err = mhi_queue_skb(mdev, DMA_FROM_DEVICE, skb, size, MHI_EOT); - if (unlikely(err)) { - net_err_ratelimited("%s: Failed to queue RX buf (%d)\n", - ndev->name, err); - kfree_skb(skb); - break; - } - - /* Do not hog the CPU if rx buffers are consumed faster than - * queued (unlikely). - */ - cond_resched(); - } - - /* If we're still starved of rx buffers, reschedule later */ - if (mhi_get_free_desc_count(mdev, DMA_FROM_DEVICE) == mhi_netdev->rx_queue_sz) - schedule_delayed_work(&mhi_netdev->rx_refill, HZ / 2); -} - -static struct device_type wwan_type = { - .name = "wwan", -}; - -static int mhi_net_probe(struct mhi_device *mhi_dev, - const struct mhi_device_id *id) -{ - const struct mhi_device_info *info = (struct mhi_device_info *)id->driver_data; - struct device *dev = &mhi_dev->dev; - struct mhi_net_dev *mhi_netdev; - struct net_device *ndev; - int err; - - ndev = alloc_netdev(sizeof(*mhi_netdev), info->netname, - NET_NAME_PREDICTABLE, mhi_net_setup); - if (!ndev) - return -ENOMEM; - - mhi_netdev = netdev_priv(ndev); - dev_set_drvdata(dev, mhi_netdev); - mhi_netdev->ndev = ndev; - mhi_netdev->mdev = mhi_dev; - mhi_netdev->proto = info->proto; - SET_NETDEV_DEV(ndev, &mhi_dev->dev); - SET_NETDEV_DEVTYPE(ndev, &wwan_type); - - INIT_DELAYED_WORK(&mhi_netdev->rx_refill, mhi_net_rx_refill_work); - u64_stats_init(&mhi_netdev->stats.rx_syncp); - u64_stats_init(&mhi_netdev->stats.tx_syncp); - - /* Start MHI channels */ - err = mhi_prepare_for_transfer(mhi_dev); - if (err) - goto out_err; - - /* Number of transfer descriptors determines size of the queue */ - mhi_netdev->rx_queue_sz = mhi_get_free_desc_count(mhi_dev, DMA_FROM_DEVICE); - - err = register_netdev(ndev); - if (err) - goto out_err; - - if (mhi_netdev->proto) { - err = mhi_netdev->proto->init(mhi_netdev); - if (err) - goto out_err_proto; - } - - return 0; - -out_err_proto: - unregister_netdev(ndev); -out_err: - free_netdev(ndev); - return err; -} - -static void mhi_net_remove(struct mhi_device *mhi_dev) -{ - struct mhi_net_dev *mhi_netdev = dev_get_drvdata(&mhi_dev->dev); - - unregister_netdev(mhi_netdev->ndev); - - mhi_unprepare_from_transfer(mhi_netdev->mdev); - - free_netdev(mhi_netdev->ndev); -} - -static const struct mhi_device_info mhi_hwip0 = { - .netname = "mhi_hwip%d", -}; - -static const struct mhi_device_info mhi_swip0 = { - .netname = "mhi_swip%d", -}; - -static const struct mhi_device_id mhi_net_id_table[] = { - /* Hardware accelerated data PATH (to modem IPA), protocol agnostic */ - { .chan = "IP_HW0", .driver_data = (kernel_ulong_t)&mhi_hwip0 }, - /* Software data PATH (to modem CPU) */ - { .chan = "IP_SW0", .driver_data = (kernel_ulong_t)&mhi_swip0 }, - {} -}; -MODULE_DEVICE_TABLE(mhi, mhi_net_id_table); - -static struct mhi_driver mhi_net_driver = { - .probe = mhi_net_probe, - .remove = mhi_net_remove, - .dl_xfer_cb = mhi_net_dl_callback, - .ul_xfer_cb = mhi_net_ul_callback, - .id_table = mhi_net_id_table, - .driver = { - .name = "mhi_net", - .owner = THIS_MODULE, - }, -}; - -module_mhi_driver(mhi_net_driver); - -MODULE_AUTHOR("Loic Poulain "); -MODULE_DESCRIPTION("Network over MHI"); -MODULE_LICENSE("GPL v2"); From patchwork Mon Feb 1 21:05:42 2021 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 8bit X-Patchwork-Submitter: Loic Poulain X-Patchwork-Id: 374290 Delivered-To: patch@linaro.org Received: by 2002:a17:906:24d5:0:0:0:0 with SMTP id f21csp4245896ejb; Mon, 1 Feb 2021 13:00:07 -0800 (PST) X-Google-Smtp-Source: ABdhPJyswyzFi/n9a+arH8/Vf3Xtsjflj8kSrBXlXMumaW1ite1eWwOL08FvsLqS1I34uQ0lvOos X-Received: by 2002:a17:906:2a42:: with SMTP id k2mr19444759eje.118.1612213207075; Mon, 01 Feb 2021 13:00:07 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1612213207; cv=none; d=google.com; s=arc-20160816; b=CQkC0/mGCu+nJeKwkhcUgM6MlOPZXh8LozDloccmfNZQezrgzrgGGZ5bRcNdaNha0E PvQkR4Ik7wlDy51awEvr30ViKkIIfzoSd9zToX2WGDePd5wUvC4i3OSV7xpd07iS27+U GGFd3imAXEWaVcq2i91Dg1hJXdQYOVluk8OyWkSSl3ULxgOvEVb8Ea3kcSxFGpoBfaWB PymwWIHXFpCE2O5KbP+ivX6h3d1RHHaOLxa6EbXQ7ljVLqSlZ6IXiziwWjdfqfrePX9A iOFpGQNyTAF6UWM682q89G/Ptvgc+HZZNrUZrWm0vEecLq7sKFBvJ8orqAQFeIlcBTJv DWJg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:to:content-transfer-encoding:mime-version :references:in-reply-to:message-id:date:subject:cc:from :dkim-signature; bh=htzqG19itchr0FC+l5UAplmrUGBTQ6p8V+c9XdRZqME=; b=uUpfjyQH+X2oksW6t50q0IytVPl210nMlpXxl7UporSQYTSrSNqPRlB+Mh2xgl7Stv 2X1RItBzu6bhZlX2aLR19ZeznDv3GRXQ/3lfq19XARrNcGBDg7o2jt/TETgbGumjI2Yg J1NElv8bMFu9mawH8kttSJGkgO+Ag0wpep+MLQSxd/VIYOy7d/9cIkkXVbJ6x15oi7Jl DjhRS7tb0EXavoeXSz6grj6iJQTfMf0VaH0REH8FMyzkm0W6QdumdLJpR/f9BsIIs3Lz eOfSOkaNseb7gXfycNmC3ItxhBTKlcGnvQzVW/K90v84XiDBMbOR7wL45iOhUeSthn70 Sokg== ARC-Authentication-Results: i=1; mx.google.com; dkim=fail header.i=@linaro.org header.s=google header.b="mJgdG6a/"; spf=pass (google.com: domain of netdev-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=netdev-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=linaro.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id a18si11311916ejr.276.2021.02.01.13.00.06; Mon, 01 Feb 2021 13:00:07 -0800 (PST) Received-SPF: pass (google.com: domain of netdev-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=fail header.i=@linaro.org header.s=google header.b="mJgdG6a/"; spf=pass (google.com: domain of netdev-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=netdev-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=linaro.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S232021AbhBAU6w (ORCPT + 7 others); Mon, 1 Feb 2021 15:58:52 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:56080 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S231948AbhBAU6q (ORCPT ); Mon, 1 Feb 2021 15:58:46 -0500 Received: from mail-wm1-x335.google.com (mail-wm1-x335.google.com [IPv6:2a00:1450:4864:20::335]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id D2C85C0613D6 for ; Mon, 1 Feb 2021 12:58:05 -0800 (PST) Received: by mail-wm1-x335.google.com with SMTP id m1so471951wml.2 for ; Mon, 01 Feb 2021 12:58:05 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linaro.org; s=google; h=from:to:cc:subject:date:message-id:in-reply-to:references :mime-version:content-transfer-encoding; bh=htzqG19itchr0FC+l5UAplmrUGBTQ6p8V+c9XdRZqME=; b=mJgdG6a/T1rI5s53QYkbeBYy1ERT7fUHP8dkofXv+ookyfeLPAED9hlD4GaR0qyLLf VVz6smXyTIpxmpiDkL30Dm2/HXSaKDdePJxzm1XKLXHZg1zrgzAttFUBiU8sFPAOReXx Aof3Qt1K2E69FpRHgCoM3F2JlEgcKKYPst795KnZAsknH93QRgZacy/CjSt9Dmya6VZr InItjYPUpiqRObVT3cQ1hSsUK3mlopoItNd2irwvVYGFK9aGcsLH6XXcwIn9WEGA5luv i1Lvt2W4szCzmK6ocJ/kQ2g0hu35VrNGwqjPqpTW76APaqmQOA4ndz/S2A/9cLo0fk3M 9DxA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references:mime-version:content-transfer-encoding; bh=htzqG19itchr0FC+l5UAplmrUGBTQ6p8V+c9XdRZqME=; b=rz1cwVTCCQs0if1N6Do6aP9j6mCPSrQ/a8ZojKsp88OsGrAwslsZB2mQ7uYP0gKkmv orh2Cgh33ofYx7teC54HCNNLb9ANd8H9WqDU7xZWKBw7uSq1j5L7tAls6Y9sVIdmzV7T XewrOEBrOcLrecqRMqDOmL8D8kjZZ19aGY8DJRQv/UTc/Y38UefxBSPp/BrMZlHP1yGk +0gjxpPQtf5qZQwGwSjnlDmVz01P8fTOl8vC72ePus6Zem1m/yztNMfrSaSK9HI5mGFp zSB7ySgIfnsv8UxujY8rtBucgsNDccKwYxX45b6CGF/scVZpLoN56XWwom1yjO+GXsic 3H3w== X-Gm-Message-State: AOAM532+UOI5atQX0FEdcnrYAgDLJtJY/maWi9rUj5zZE0AWaMFR3FOi nTxMF1PQEQKkwB7iFPbQSN+RWw== X-Received: by 2002:a1c:4c07:: with SMTP id z7mr591327wmf.129.1612213084493; Mon, 01 Feb 2021 12:58:04 -0800 (PST) Received: from localhost.localdomain ([88.122.66.28]) by smtp.gmail.com with ESMTPSA id b7sm27894226wru.33.2021.02.01.12.58.03 (version=TLS1_2 cipher=ECDHE-ECDSA-AES128-GCM-SHA256 bits=128/128); Mon, 01 Feb 2021 12:58:04 -0800 (PST) From: Loic Poulain Cc: bjorn@mork.no, dcbw@redhat.com, netdev@vger.kernel.org, carl.yin@quectel.com, Loic Poulain Subject: [PATCH net-next v2 3/3] net: mhi: Add mbim proto Date: Mon, 1 Feb 2021 22:05:42 +0100 Message-Id: <1612213542-17257-3-git-send-email-loic.poulain@linaro.org> X-Mailer: git-send-email 2.7.4 In-Reply-To: <1612213542-17257-1-git-send-email-loic.poulain@linaro.org> References: <1612213542-17257-1-git-send-email-loic.poulain@linaro.org> MIME-Version: 1.0 To: unlisted-recipients:; (no To-header on input) Precedence: bulk List-ID: X-Mailing-List: netdev@vger.kernel.org MBIM has initially been specified by USB-IF for transporting data (IP) between a modem and a host over USB. However some modern modems also support MBIM over PCIe (via MHI). In the same way as QMAP(rmnet), it allows to aggregate IP packets and to perform context multiplexing. This change adds minimal MBIM data transport support to MHI, allowing to support MBIM only modems. MBIM being based on USB NCM, it reuses and copy some helpers/functions from the USB stack (cdc-ncm, cdc-mbim). Note that is a subset of the CDC-MBIM specification, supporting only transport of network data (IP), there is no support for DSS. Moreover the multi-session (for multi-pdn) is not supported in this initial version, but will be added latter, and aligned with the cdc-mbim solution (VLAN tags). This code has been inspired from the mhi_mbim downstream implementation (Carl Yin ). Signed-off-by: Loic Poulain --- v2: - Check nth size/sequence in nth16_verify - Add netif_dbg message for verbose error - Add inline comment for MHI MBIM limitation (no DSS) - Fix copyright issue - Reword commit message drivers/net/mhi/Makefile | 2 +- drivers/net/mhi/mhi.h | 39 +++++++ drivers/net/mhi/net.c | 40 ++----- drivers/net/mhi/proto_mbim.c | 270 +++++++++++++++++++++++++++++++++++++++++++ 4 files changed, 320 insertions(+), 31 deletions(-) create mode 100644 drivers/net/mhi/mhi.h create mode 100644 drivers/net/mhi/proto_mbim.c -- 2.7.4 diff --git a/drivers/net/mhi/Makefile b/drivers/net/mhi/Makefile index 0acf989..f71b9f8 100644 --- a/drivers/net/mhi/Makefile +++ b/drivers/net/mhi/Makefile @@ -1,3 +1,3 @@ obj-$(CONFIG_MHI_NET) += mhi_net.o -mhi_net-y := net.o +mhi_net-y := net.o proto_mbim.o diff --git a/drivers/net/mhi/mhi.h b/drivers/net/mhi/mhi.h new file mode 100644 index 0000000..94797b5 --- /dev/null +++ b/drivers/net/mhi/mhi.h @@ -0,0 +1,39 @@ +/* SPDX-License-Identifier: GPL-2.0-or-later */ +/* MHI Network driver - Network over MHI bus + * + * Copyright (C) 2021 Linaro Ltd + */ + +struct mhi_net_stats { + u64_stats_t rx_packets; + u64_stats_t rx_bytes; + u64_stats_t rx_errors; + u64_stats_t rx_dropped; + u64_stats_t tx_packets; + u64_stats_t tx_bytes; + u64_stats_t tx_errors; + u64_stats_t tx_dropped; + atomic_t rx_queued; + struct u64_stats_sync tx_syncp; + struct u64_stats_sync rx_syncp; +}; + +struct mhi_net_dev { + struct mhi_device *mdev; + struct net_device *ndev; + const struct mhi_device_info *info; + const struct mhi_net_proto *proto; + void *proto_data; + int msg_enable; + struct delayed_work rx_refill; + struct mhi_net_stats stats; + u32 rx_queue_sz; +}; + +struct mhi_net_proto { + int (*init)(struct mhi_net_dev *mhi_netdev); + struct sk_buff * (*tx_fixup)(struct mhi_net_dev *mhi_netdev, struct sk_buff *skb); + int (*rx_fixup)(struct mhi_net_dev *mhi_netdev, struct sk_buff *skb); +}; + +extern const struct mhi_net_proto proto_mbim; diff --git a/drivers/net/mhi/net.c b/drivers/net/mhi/net.c index 34d4bcf..33643572 100644 --- a/drivers/net/mhi/net.c +++ b/drivers/net/mhi/net.c @@ -12,39 +12,12 @@ #include #include +#include "mhi.h" + #define MHI_NET_MIN_MTU ETH_MIN_MTU #define MHI_NET_MAX_MTU 0xffff #define MHI_NET_DEFAULT_MTU 0x4000 -struct mhi_net_stats { - u64_stats_t rx_packets; - u64_stats_t rx_bytes; - u64_stats_t rx_errors; - u64_stats_t rx_dropped; - u64_stats_t tx_packets; - u64_stats_t tx_bytes; - u64_stats_t tx_errors; - u64_stats_t tx_dropped; - struct u64_stats_sync tx_syncp; - struct u64_stats_sync rx_syncp; -}; - -struct mhi_net_dev { - struct mhi_device *mdev; - struct net_device *ndev; - const struct mhi_net_proto *proto; - void *proto_data; - struct delayed_work rx_refill; - struct mhi_net_stats stats; - u32 rx_queue_sz; -}; - -struct mhi_net_proto { - int (*init)(struct mhi_net_dev *mhi_netdev); - struct sk_buff * (*tx_fixup)(struct mhi_net_dev *mhi_netdev, struct sk_buff *skb); - int (*rx_fixup)(struct mhi_net_dev *mhi_netdev, struct sk_buff *skb); -}; - struct mhi_device_info { const char *netname; const struct mhi_net_proto *proto; @@ -348,12 +321,19 @@ static const struct mhi_device_info mhi_swip0 = { .netname = "mhi_swip%d", }; +static const struct mhi_device_info mhi_hwip0_mbim = { + .netname = "mhi_mbim%d", + .proto = &proto_mbim, +}; + static const struct mhi_device_id mhi_net_id_table[] = { /* Hardware accelerated data PATH (to modem IPA), protocol agnostic */ { .chan = "IP_HW0", .driver_data = (kernel_ulong_t)&mhi_hwip0 }, /* Software data PATH (to modem CPU) */ { .chan = "IP_SW0", .driver_data = (kernel_ulong_t)&mhi_swip0 }, - {} + /* Hardware accelerated data PATH (to modem IPA), MBIM protocol */ + { .chan = "IP_HW0_MBIM", .driver_data = (kernel_ulong_t)&mhi_hwip0_mbim }, + { } }; MODULE_DEVICE_TABLE(mhi, mhi_net_id_table); diff --git a/drivers/net/mhi/proto_mbim.c b/drivers/net/mhi/proto_mbim.c new file mode 100644 index 0000000..a8499aa --- /dev/null +++ b/drivers/net/mhi/proto_mbim.c @@ -0,0 +1,270 @@ +// SPDX-License-Identifier: GPL-2.0-or-later +/* MHI Network driver - Network over MHI bus + * + * Copyright (C) 2021 Linaro Ltd + * + * This driver copy some code from cdc_ncm, which is: + * Copyright (C) ST-Ericsson 2010-2012 + * and cdc_mbim, which is: + * Copyright (c) 2012 Smith Micro Software, Inc. + * Copyright (c) 2012 Bjørn Mork + * + */ + +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include + +#include "mhi.h" + +#define MHI_MBIM_MAX_BLOCK_SZ (31 * 1024) + +struct mbim_context { + u16 rx_seq; + u16 tx_seq; +}; + +static int mbim_rx_verify_nth16(struct sk_buff *skb) +{ + struct mhi_net_dev *dev = netdev_priv(skb->dev); + struct mbim_context *ctx = dev->proto_data; + struct usb_cdc_ncm_nth16 *nth16; + int ret = -EINVAL; + int len; + + if (skb->len < (sizeof(struct usb_cdc_ncm_nth16) + + sizeof(struct usb_cdc_ncm_ndp16))) { + netif_dbg(dev, rx_err, dev->ndev, "frame too short\n"); + goto error; + } + + nth16 = (struct usb_cdc_ncm_nth16 *)skb->data; + + if (nth16->dwSignature != cpu_to_le32(USB_CDC_NCM_NTH16_SIGN)) { + netif_dbg(dev, rx_err, dev->ndev, + "invalid NTH16 signature <%#010x>\n", + le32_to_cpu(nth16->dwSignature)); + goto error; + } + + /* No limit on the block length, except the size of the data pkt */ + len = le16_to_cpu(nth16->wBlockLength); + if (len > skb->len) { + netif_dbg(dev, rx_err, dev->ndev, + "NTB does not fit into the skb %u/%u\n", len, + skb->len); + goto error; + } + + if ((ctx->rx_seq + 1) != le16_to_cpu(nth16->wSequence) && + (ctx->rx_seq || le16_to_cpu(nth16->wSequence)) && + !((ctx->rx_seq == 0xffff) && !le16_to_cpu(nth16->wSequence))) { + netif_dbg(dev, rx_err, dev->ndev, + "sequence number glitch prev=%d curr=%d\n", + ctx->rx_seq, le16_to_cpu(nth16->wSequence)); + } + ctx->rx_seq = le16_to_cpu(nth16->wSequence); + + ret = le16_to_cpu(nth16->wNdpIndex); +error: + return ret; +} + +static int mbim_rx_verify_ndp16(struct sk_buff *skb, int ndpoffset) +{ + struct mhi_net_dev *dev = netdev_priv(skb->dev); + struct usb_cdc_ncm_ndp16 *ndp16; + int ret = -EINVAL; + + if ((ndpoffset + sizeof(struct usb_cdc_ncm_ndp16)) > skb->len) { + netif_dbg(dev, rx_err, dev->ndev, "invalid NDP offset <%u>\n", + ndpoffset); + goto error; + } + + ndp16 = (struct usb_cdc_ncm_ndp16 *)(skb->data + ndpoffset); + + if (le16_to_cpu(ndp16->wLength) < USB_CDC_NCM_NDP16_LENGTH_MIN) { + netif_dbg(dev, rx_err, dev->ndev, "invalid DPT16 length <%u>\n", + le16_to_cpu(ndp16->wLength)); + goto error; + } + + ret = ((le16_to_cpu(ndp16->wLength) - + sizeof(struct usb_cdc_ncm_ndp16)) / + sizeof(struct usb_cdc_ncm_dpe16)); + ret--; /* Last entry is always a NULL terminator */ + + if ((sizeof(struct usb_cdc_ncm_ndp16) + + ret * (sizeof(struct usb_cdc_ncm_dpe16))) > skb->len) { + netif_dbg(dev, rx_err, dev->ndev, + "Invalid nframes = %d\n", ret); + ret = -EINVAL; + } + +error: + return ret; +} + +static int mbim_rx_fixup(struct mhi_net_dev *mhi_netdev, struct sk_buff *skb) +{ + struct net_device *ndev = mhi_netdev->ndev; + int ndpoffset; + + /* Check NTB header and retrieve first NDP offset */ + ndpoffset = mbim_rx_verify_nth16(skb); + if (ndpoffset < 0) { + netdev_err(ndev, "MBIM: Incorrect NTB header\n"); + goto error; + } + + /* Process each NDP */ + while (1) { + struct usb_cdc_ncm_ndp16 *ndp16; + struct usb_cdc_ncm_dpe16 *dpe16; + int nframes, n; + + /* Check NDP header and retrieve number of datagrams */ + nframes = mbim_rx_verify_ndp16(skb, ndpoffset); + if (nframes < 0) { + netdev_err(ndev, "MBIM: Incorrect NDP16\n"); + goto error; + } + + /* Only support the IPS session 0 for now (only one PDN context) + * There is no Data Service Stream (DSS) in MHI context. + */ + ndp16 = (struct usb_cdc_ncm_ndp16 *)(skb->data + ndpoffset); + switch (ndp16->dwSignature & cpu_to_le32(0x00ffffff)) { + case cpu_to_le32(USB_CDC_MBIM_NDP16_IPS_SIGN): + break; + default: + netdev_err(ndev, "MBIM: Unsupported NDP type\n"); + goto next_ndp; + } + + /* de-aggregate and deliver IP packets */ + dpe16 = ndp16->dpe16; + for (n = 0; n < nframes; n++, dpe16++) { + u16 dgram_offset = le16_to_cpu(dpe16->wDatagramIndex); + u16 dgram_len = le16_to_cpu(dpe16->wDatagramLength); + struct sk_buff *skbn; + + if (!dgram_offset || !dgram_len) + break; /* null terminator */ + + skbn = netdev_alloc_skb(ndev, dgram_len); + if (!skbn) + continue; + + skb_put(skbn, dgram_len); + memcpy(skbn->data, skb->data + dgram_offset, dgram_len); + + switch (skbn->data[0] & 0xf0) { + case 0x40: + skbn->protocol = htons(ETH_P_IP); + break; + case 0x60: + skbn->protocol = htons(ETH_P_IPV6); + break; + default: + netdev_err(ndev, "MBIM: unknown protocol\n"); + continue; + } + + netif_rx(skbn); + } +next_ndp: + /* Other NDP to process? */ + ndpoffset = le16_to_cpu(ndp16->wNextNdpIndex); + if (!ndpoffset) + break; + } + + /* free skb */ + dev_consume_skb_any(skb); + return 0; +error: + dev_kfree_skb_any(skb); + return -EIO; +} + +struct mbim_tx_hdr { + struct usb_cdc_ncm_nth16 nth16; + struct usb_cdc_ncm_ndp16 ndp16; + struct usb_cdc_ncm_dpe16 dpe16[2]; +} __packed; + +static struct sk_buff *mbim_tx_fixup(struct mhi_net_dev *mhi_netdev, + struct sk_buff *skb) +{ + struct mbim_context *ctx = mhi_netdev->proto_data; + unsigned int dgram_size = skb->len; + struct usb_cdc_ncm_nth16 *nth16; + struct usb_cdc_ncm_ndp16 *ndp16; + struct mbim_tx_hdr *mbim_hdr; + + /* For now, this is a partial implementation of CDC MBIM, only one NDP + * is sent, containing the IP packet (no aggregation). + */ + + if (skb_headroom(skb) < sizeof(struct mbim_tx_hdr)) { + dev_kfree_skb_any(skb); + return NULL; + } + + mbim_hdr = skb_push(skb, sizeof(struct mbim_tx_hdr)); + + /* Fill NTB header */ + nth16 = &mbim_hdr->nth16; + nth16->dwSignature = cpu_to_le32(USB_CDC_NCM_NTH16_SIGN); + nth16->wHeaderLength = cpu_to_le16(sizeof(struct usb_cdc_ncm_nth16)); + nth16->wSequence = cpu_to_le16(ctx->tx_seq++); + nth16->wBlockLength = cpu_to_le16(skb->len); + nth16->wNdpIndex = cpu_to_le16(sizeof(struct usb_cdc_ncm_nth16)); + + /* Fill the unique NDP */ + ndp16 = &mbim_hdr->ndp16; + ndp16->dwSignature = cpu_to_le32(USB_CDC_MBIM_NDP16_IPS_SIGN); + ndp16->wLength = cpu_to_le16(sizeof(struct usb_cdc_ncm_ndp16) + + sizeof(struct usb_cdc_ncm_dpe16) * 2); + ndp16->wNextNdpIndex = 0; + + /* Datagram follows the mbim header */ + ndp16->dpe16[0].wDatagramIndex = cpu_to_le16(sizeof(struct mbim_tx_hdr)); + ndp16->dpe16[0].wDatagramLength = cpu_to_le16(dgram_size); + + /* null termination */ + ndp16->dpe16[1].wDatagramIndex = 0; + ndp16->dpe16[1].wDatagramLength = 0; + + return skb; +} + +static int mbim_init(struct mhi_net_dev *mhi_netdev) +{ + struct net_device *ndev = mhi_netdev->ndev; + + mhi_netdev->proto_data = devm_kzalloc(&ndev->dev, + sizeof(struct mbim_context), + GFP_KERNEL); + if (!mhi_netdev->proto_data) + return -ENOMEM; + + ndev->needed_headroom = sizeof(struct mbim_tx_hdr); + + return 0; +} + +const struct mhi_net_proto proto_mbim = { + .init = mbim_init, + .rx_fixup = mbim_rx_fixup, + .tx_fixup = mbim_tx_fixup, +};