Message ID | 1602275611-7440-1-git-send-email-loic.poulain@linaro.org |
---|---|
State | New |
Headers | show |
Series | net: Add mhi-net driver | expand |
On Fri, 9 Oct 2020 22:33:31 +0200 Loic Poulain wrote: > This patch adds a new network driver implementing MHI transport for > network packets. Packets can be in any format, though QMAP (rmnet) > is the usual protocol (flow control + PDN mux). > > It support two MHI devices, IP_HW0 which is, the path to the IPA > (IP accelerator) on qcom modem, And IP_SW0 which is the software > driven IP path (to modem CPU). > > Signed-off-by: Loic Poulain <loic.poulain@linaro.org> Does not appear to build in net-next - presumably you have some dependency in another tree? make: *** [__sub-make] Error 2 ../drivers/net/mhi_net.c: In function ‘mhi_net_probe’: ../drivers/net/mhi_net.c:234:8: error: too many arguments to function ‘mhi_prepare_for_transfer’ 234 | err = mhi_prepare_for_transfer(mhi_dev, 0); | ^~~~~~~~~~~~~~~~~~~~~~~~ In file included from ../drivers/net/mhi_net.c:8: ../include/linux/mhi.h:669:5: note: declared here 669 | int mhi_prepare_for_transfer(struct mhi_device *mhi_dev); | ^~~~~~~~~~~~~~~~~~~~~~~~ make[3]: *** [drivers/net/mhi_net.o] Error 1 make[2]: *** [drivers/net] Error 2 make[2]: *** Waiting for unfinished jobs.... make[1]: *** [drivers] Error 2 make[1]: *** Waiting for unfinished jobs.... make: *** [__sub-make] Error 2 New errors added
On Fri, 9 Oct 2020 22:33:31 +0200 Loic Poulain wrote: > This patch adds a new network driver implementing MHI transport for > network packets. Packets can be in any format, though QMAP (rmnet) > is the usual protocol (flow control + PDN mux). > > It support two MHI devices, IP_HW0 which is, the path to the IPA > (IP accelerator) on qcom modem, And IP_SW0 which is the software > driven IP path (to modem CPU). > > Signed-off-by: Loic Poulain <loic.poulain@linaro.org> > +static int mhi_ndo_xmit(struct sk_buff *skb, struct net_device *ndev) > +{ > + struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); > + struct mhi_device *mdev = mhi_netdev->mdev; > + int err; > + > + /* Only support for single buffer transfer for now */ > + err = skb_linearize(skb); Since you don't advertise NETIF_F_SG you shouldn't have to call this, no? > + if (unlikely(err)) { > + kfree_skb(skb); > + mhi_netdev->stats.tx_dropped++; > + return NETDEV_TX_OK; > + } > + > + skb_tx_timestamp(skb); > + > + /* mhi_queue_skb is not thread-safe, but xmit is serialized by the > + * network core. Once MHI core will be thread save, migrate to > + * NETIF_F_LLTX support. > + */ > + err = mhi_queue_skb(mdev, DMA_TO_DEVICE, skb, skb->len, MHI_EOT); > + if (err) { > + netdev_err(ndev, "mhi_queue_skb err %d\n", err); This needs to be at least rate limited. > + netif_stop_queue(ndev); What's going to start the queue if it's a transient errors and not NETDEV_TX_BUSY? > + return (err == -ENOMEM) ? NETDEV_TX_BUSY : err; You should drop the packet if it's not NETDEV_TX_BUSY, and return NETDEV_TX_OK. Don't return negative errors from ndo_xmit. > + } > + > + return NETDEV_TX_OK; > +} > + > +static void mhi_ndo_get_stats64(struct net_device *ndev, > + struct rtnl_link_stats64 *stats) > +{ > + struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); > + > + stats->rx_packets = mhi_netdev->stats.rx_packets; > + stats->rx_bytes = mhi_netdev->stats.rx_bytes; > + stats->rx_errors = mhi_netdev->stats.rx_errors; > + stats->rx_dropped = mhi_netdev->stats.rx_dropped; > + stats->tx_packets = mhi_netdev->stats.tx_packets; > + stats->tx_bytes = mhi_netdev->stats.tx_bytes; > + stats->tx_errors = mhi_netdev->stats.tx_errors; > + stats->tx_dropped = mhi_netdev->stats.tx_dropped; > +} Can you use > +static void mhi_net_dl_callback(struct mhi_device *mhi_dev, > + struct mhi_result *mhi_res) > +{ > + struct mhi_net_dev *mhi_netdev = dev_get_drvdata(&mhi_dev->dev); > + struct sk_buff *skb = mhi_res->buf_addr; > + > + atomic_dec(&mhi_netdev->stats.rx_queued); > + > + if (mhi_res->transaction_status) { > + mhi_netdev->stats.rx_errors++; > + kfree_skb(skb); > + } else { > + mhi_netdev->stats.rx_packets++; > + mhi_netdev->stats.rx_bytes += mhi_res->bytes_xferd; > + > + skb->protocol = htons(ETH_P_MAP); > + skb_put(skb, mhi_res->bytes_xferd); > + netif_rx(skb); > + } > + > + schedule_delayed_work(&mhi_netdev->rx_refill, 0); Scheduling a work to replace every single RX buffer looks quite inefficient. Any chance you can do batching? I assume mhi_queue_skb() has to be able to sleep? > +static void mhi_net_rx_refill_work(struct work_struct *work) > +{ > + struct mhi_net_dev *mhi_netdev = container_of(work, struct mhi_net_dev, > + rx_refill.work); > + struct net_device *ndev = mhi_netdev->ndev; > + struct mhi_device *mdev = mhi_netdev->mdev; > + struct sk_buff *skb; > + int err; > + > + if (!netif_running(ndev)) > + return; How can this happen? You cancel the work from ndo_stop. > + do { > + skb = netdev_alloc_skb(ndev, READ_ONCE(ndev->mtu)); > + if (unlikely(!skb)) { > + /* If we are starved of RX buffers, retry later */ > + if (!atomic_read(&mhi_netdev->stats.rx_queued)) > + schedule_delayed_work(&mhi_netdev->rx_refill, HZ / 2); > + break; > + } > + > + err = mhi_queue_skb(mdev, DMA_FROM_DEVICE, skb, ndev->mtu, > + MHI_EOT); > + if (err) { > + atomic_dec(&mhi_netdev->stats.rx_queued); This can never fail with an empty ring? No need to potentially reschedule the work here? > + kfree_skb(skb); > + break; > + } > + > + atomic_inc(&mhi_netdev->stats.rx_queued); > + > + } while (1); > +} > + > +static int mhi_net_probe(struct mhi_device *mhi_dev, > + const struct mhi_device_id *id) > +{ > + const char *netname = (char *)id->driver_data; > + struct mhi_net_dev *mhi_netdev; > + struct net_device *ndev; > + struct device *dev = &mhi_dev->dev; > + int err; > + > + ndev = alloc_netdev(sizeof(*mhi_netdev), netname, NET_NAME_PREDICTABLE, > + mhi_net_setup); > + if (!ndev) { > + err = -ENOMEM; > + return err; return -ENOMEM; > + } > + > + mhi_netdev = netdev_priv(ndev); > + dev_set_drvdata(dev, mhi_netdev); > + mhi_netdev->ndev = ndev; > + mhi_netdev->mdev = mhi_dev; SET_NETDEV_DEV() ? > + INIT_DELAYED_WORK(&mhi_netdev->rx_refill, mhi_net_rx_refill_work); > + > + /* Start MHI channels */ > + err = mhi_prepare_for_transfer(mhi_dev, 0); > + if (err) { > + free_netdev(ndev); > + return err; > + } > + > + err = register_netdev(ndev); > + if (err) { > + dev_err(dev, "mhi_net: registering device failed\n"); > + free_netdev(ndev); > + return -EINVAL; Why not propagate the error? > + } > + > + return 0; > +} > + > +static void mhi_net_remove(struct mhi_device *mhi_dev) > +{ > + struct mhi_net_dev *mhi_netdev = dev_get_drvdata(&mhi_dev->dev); > + > + mhi_unprepare_from_transfer(mhi_netdev->mdev); > + unregister_netdev(mhi_netdev->ndev); Isn't this the wrong way around? Should you not unregister the netdev before you stop transfers? > + /* netdev released from unregister */ > +}
Hi Jakub, and thanks for your review. On Sun, 11 Oct 2020 at 20:59, Jakub Kicinski <kuba@kernel.org> wrote: > > On Fri, 9 Oct 2020 22:33:31 +0200 Loic Poulain wrote: > > This patch adds a new network driver implementing MHI transport for > > network packets. Packets can be in any format, though QMAP (rmnet) > > is the usual protocol (flow control + PDN mux). > > > > It support two MHI devices, IP_HW0 which is, the path to the IPA > > (IP accelerator) on qcom modem, And IP_SW0 which is the software > > driven IP path (to modem CPU). > > > > Signed-off-by: Loic Poulain <loic.poulain@linaro.org> > > > +static int mhi_ndo_xmit(struct sk_buff *skb, struct net_device *ndev) > > +{ > > + struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); > > + struct mhi_device *mdev = mhi_netdev->mdev; > > + int err; > > + > > + /* Only support for single buffer transfer for now */ > > + err = skb_linearize(skb); > > Since you don't advertise NETIF_F_SG you shouldn't have to call this, > no? Right, good to know. > > > + if (unlikely(err)) { > > + kfree_skb(skb); > > + mhi_netdev->stats.tx_dropped++; > > + return NETDEV_TX_OK; > > + } > > + > > + skb_tx_timestamp(skb); > > + > > + /* mhi_queue_skb is not thread-safe, but xmit is serialized by the > > + * network core. Once MHI core will be thread save, migrate to > > + * NETIF_F_LLTX support. > > + */ > > + err = mhi_queue_skb(mdev, DMA_TO_DEVICE, skb, skb->len, MHI_EOT); > > + if (err) { > > + netdev_err(ndev, "mhi_queue_skb err %d\n", err); > > This needs to be at least rate limited. Yes, I've missed removing that message, going to keep rate-limited version for non-busy error case. > > > + netif_stop_queue(ndev); > > What's going to start the queue if it's a transient errors and not > NETDEV_TX_BUSY? > > > + return (err == -ENOMEM) ? NETDEV_TX_BUSY : err; > > You should drop the packet if it's not NETDEV_TX_BUSY, and return > NETDEV_TX_OK. Don't return negative errors from ndo_xmit. Ok, going to determine which error it is and act accordingly. > > > + } > > + > > + return NETDEV_TX_OK; > > +} > > + > > +static void mhi_ndo_get_stats64(struct net_device *ndev, > > + struct rtnl_link_stats64 *stats) > > +{ > > + struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); > > + > > + stats->rx_packets = mhi_netdev->stats.rx_packets; > > + stats->rx_bytes = mhi_netdev->stats.rx_bytes; > > + stats->rx_errors = mhi_netdev->stats.rx_errors; > > + stats->rx_dropped = mhi_netdev->stats.rx_dropped; > > + stats->tx_packets = mhi_netdev->stats.tx_packets; > > + stats->tx_bytes = mhi_netdev->stats.tx_bytes; > > + stats->tx_errors = mhi_netdev->stats.tx_errors; > > + stats->tx_dropped = mhi_netdev->stats.tx_dropped; > > +} > > Can you use ? > > > +static void mhi_net_dl_callback(struct mhi_device *mhi_dev, > > + struct mhi_result *mhi_res) > > +{ > > + struct mhi_net_dev *mhi_netdev = dev_get_drvdata(&mhi_dev->dev); > > + struct sk_buff *skb = mhi_res->buf_addr; > > + > > + atomic_dec(&mhi_netdev->stats.rx_queued); > > + > > + if (mhi_res->transaction_status) { > > + mhi_netdev->stats.rx_errors++; > > + kfree_skb(skb); > > + } else { > > + mhi_netdev->stats.rx_packets++; > > + mhi_netdev->stats.rx_bytes += mhi_res->bytes_xferd; > > + > > + skb->protocol = htons(ETH_P_MAP); > > + skb_put(skb, mhi_res->bytes_xferd); > > + netif_rx(skb); > > + } > > + > > + schedule_delayed_work(&mhi_netdev->rx_refill, 0); > > Scheduling a work to replace every single RX buffer looks quite > inefficient. Any chance you can do batching? I assume mhi_queue_skb() > has to be able to sleep? There is already some kind of batching, the hardware can perform interrupt coalescing (called interrupt mitigation in case of MHI), which, in high traffic case will cause the dl_callback to be called in row for multiple packets, and so the work to be scheduled only once. mhi_queue_skb does not sleep. However we could implement some additional mitigation by e.g. only scheduling the work if the current rx-queue fall under a certain limit (e.g. queue size / 2), like it is done in virtio-net. What do you think? > > > +static void mhi_net_rx_refill_work(struct work_struct *work) > > +{ > > + struct mhi_net_dev *mhi_netdev = container_of(work, struct mhi_net_dev, > > + rx_refill.work); > > + struct net_device *ndev = mhi_netdev->ndev; > > + struct mhi_device *mdev = mhi_netdev->mdev; > > + struct sk_buff *skb; > > + int err; > > + > > + if (!netif_running(ndev)) > > + return; > > How can this happen? You cancel the work from ndo_stop. Right, If the work is executed while we are currently canceling it, I wanted to prevent the work to fully run and reschedule. But yes it's not strictly necessary since cancel_work_sync will 1. wait for the job to exit and 2. dequeue the work. Will remove that. > > > + do { > > + skb = netdev_alloc_skb(ndev, READ_ONCE(ndev->mtu)); > > + if (unlikely(!skb)) { > > + /* If we are starved of RX buffers, retry later */ > > + if (!atomic_read(&mhi_netdev->stats.rx_queued)) > > + schedule_delayed_work(&mhi_netdev->rx_refill, HZ / 2); > > + break; > > + } > > + > > + err = mhi_queue_skb(mdev, DMA_FROM_DEVICE, skb, ndev->mtu, > > + MHI_EOT); > > + if (err) { > > + atomic_dec(&mhi_netdev->stats.rx_queued); > > This can never fail with an empty ring? No need to potentially > reschedule the work here? Well, it can fail because IO/transient error, so yes, I need to check that case and reschedule if necessary. > > > + kfree_skb(skb); > > + break; > > + } > > + > > + atomic_inc(&mhi_netdev->stats.rx_queued); > > + > > + } while (1); > > +} > > + > > +static int mhi_net_probe(struct mhi_device *mhi_dev, > > + const struct mhi_device_id *id) > > +{ > > + const char *netname = (char *)id->driver_data; > > + struct mhi_net_dev *mhi_netdev; > > + struct net_device *ndev; > > + struct device *dev = &mhi_dev->dev; > > + int err; > > + > > + ndev = alloc_netdev(sizeof(*mhi_netdev), netname, NET_NAME_PREDICTABLE, > > + mhi_net_setup); > > + if (!ndev) { > > + err = -ENOMEM; > > + return err; > > return -ENOMEM; > > > + } > > + > > + mhi_netdev = netdev_priv(ndev); > > + dev_set_drvdata(dev, mhi_netdev); > > + mhi_netdev->ndev = ndev; > > + mhi_netdev->mdev = mhi_dev; > > SET_NETDEV_DEV() ? Ok. > > > + INIT_DELAYED_WORK(&mhi_netdev->rx_refill, mhi_net_rx_refill_work); > > + > > + /* Start MHI channels */ > > + err = mhi_prepare_for_transfer(mhi_dev, 0); > > + if (err) { > > + free_netdev(ndev); > > + return err; > > + } > > + > > + err = register_netdev(ndev); > > + if (err) { > > + dev_err(dev, "mhi_net: registering device failed\n"); > > + free_netdev(ndev); > > + return -EINVAL; > > Why not propagate the error? Will do. > > > + } > > + > > + return 0; > > +} > > + > > +static void mhi_net_remove(struct mhi_device *mhi_dev) > > +{ > > + struct mhi_net_dev *mhi_netdev = dev_get_drvdata(&mhi_dev->dev); > > + > > + mhi_unprepare_from_transfer(mhi_netdev->mdev); > > + unregister_netdev(mhi_netdev->ndev); > > Isn't this the wrong way around? > > Should you not unregister the netdev before you stop transfers? That can be done that way, but wanted to be sure that no transfer callback is called after netdev has been released (freed in unregister), though the MHI core takes care of that in its remove procedure. > > > + /* netdev released from unregister */ > > > +} Thanks, Loic
On Fri, 2020-10-09 at 22:33 +0200, Loic Poulain wrote: > This patch adds a new network driver implementing MHI transport for > network packets. Packets can be in any format, though QMAP (rmnet) > is the usual protocol (flow control + PDN mux). > > It support two MHI devices, IP_HW0 which is, the path to the IPA > (IP accelerator) on qcom modem, And IP_SW0 which is the software > driven IP path (to modem CPU). > > Signed-off-by: Loic Poulain <loic.poulain@linaro.org> > --- > drivers/net/Kconfig | 7 ++ > drivers/net/Makefile | 1 + > drivers/net/mhi_net.c | 281 > ++++++++++++++++++++++++++++++++++++++++++++++++++ > 3 files changed, 289 insertions(+) > create mode 100644 drivers/net/mhi_net.c > > diff --git a/drivers/net/Kconfig b/drivers/net/Kconfig > index 1368d1d..11a6357 100644 > --- a/drivers/net/Kconfig > +++ b/drivers/net/Kconfig > @@ -426,6 +426,13 @@ config VSOCKMON > mostly intended for developers or support to debug vsock > issues. If > unsure, say N. > > +config MHI_NET > + tristate "MHI network driver" > + depends on MHI_BUS > + help > + This is the network driver for MHI. It can be used with > + QCOM based WWAN modems (like SDX55). Say Y or M. > + > endif # NET_CORE > > config SUNGEM_PHY > diff --git a/drivers/net/Makefile b/drivers/net/Makefile > index 94b6080..8312037 100644 > --- a/drivers/net/Makefile > +++ b/drivers/net/Makefile > @@ -34,6 +34,7 @@ obj-$(CONFIG_GTP) += gtp.o > obj-$(CONFIG_NLMON) += nlmon.o > obj-$(CONFIG_NET_VRF) += vrf.o > obj-$(CONFIG_VSOCKMON) += vsockmon.o > +obj-$(CONFIG_MHI_NET) += mhi_net.o > > # > # Networking Drivers > diff --git a/drivers/net/mhi_net.c b/drivers/net/mhi_net.c > new file mode 100644 > index 0000000..f66248e > --- /dev/null > +++ b/drivers/net/mhi_net.c > @@ -0,0 +1,281 @@ > +// SPDX-License-Identifier: GPL-2.0-or-later > +/* MHI Network driver - Network over MHI > + * > + * Copyright (C) 2020 Linaro Ltd <loic.poulain@linaro.org> > + */ > + > +#include <linux/if_arp.h> > +#include <linux/mhi.h> > +#include <linux/mod_devicetable.h> > +#include <linux/module.h> > +#include <linux/netdevice.h> > +#include <linux/skbuff.h> > + > +#define MIN_MTU ETH_MIN_MTU > +#define MAX_MTU 0xffff > +#define DEFAULT_MTU 8192 > + > +struct mhi_net_stats { > + u64 rx_packets; > + u64 rx_bytes; > + u64 rx_errors; > + u64 rx_dropped; > + u64 tx_packets; > + u64 tx_bytes; > + u64 tx_errors; > + u64 tx_dropped; > + atomic_t rx_queued; > +}; > + > +struct mhi_net_dev { > + struct mhi_device *mdev; > + struct net_device *ndev; > + struct delayed_work rx_refill; > + struct mhi_net_stats stats; > +}; > + > +static int mhi_ndo_open(struct net_device *ndev) > +{ > + struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); > + > + /* Feed the rx buffer pool */ > + schedule_delayed_work(&mhi_netdev->rx_refill, 0); > + > + /* Carrier is established via out-of-band channel (e.g. qmi) */ > + netif_carrier_on(ndev); > + > + netif_start_queue(ndev); > + > + return 0; > +} > + > +static int mhi_ndo_stop(struct net_device *ndev) > +{ > + struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); > + > + netif_stop_queue(ndev); > + netif_carrier_off(ndev); > + cancel_delayed_work_sync(&mhi_netdev->rx_refill); > + > + return 0; > +} > + > +static int mhi_ndo_xmit(struct sk_buff *skb, struct net_device > *ndev) > +{ > + struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); > + struct mhi_device *mdev = mhi_netdev->mdev; > + int err; > + > + /* Only support for single buffer transfer for now */ > + err = skb_linearize(skb); > + if (unlikely(err)) { > + kfree_skb(skb); > + mhi_netdev->stats.tx_dropped++; > + return NETDEV_TX_OK; > + } > + > + skb_tx_timestamp(skb); > + > + /* mhi_queue_skb is not thread-safe, but xmit is serialized by > the > + * network core. Once MHI core will be thread save, migrate to > + * NETIF_F_LLTX support. > + */ > + err = mhi_queue_skb(mdev, DMA_TO_DEVICE, skb, skb->len, > MHI_EOT); > + if (err) { > + netdev_err(ndev, "mhi_queue_skb err %d\n", err); > + netif_stop_queue(ndev); > + return (err == -ENOMEM) ? NETDEV_TX_BUSY : err; > + } > + > + return NETDEV_TX_OK; > +} > + > +static void mhi_ndo_get_stats64(struct net_device *ndev, > + struct rtnl_link_stats64 *stats) > +{ > + struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); > + > + stats->rx_packets = mhi_netdev->stats.rx_packets; > + stats->rx_bytes = mhi_netdev->stats.rx_bytes; > + stats->rx_errors = mhi_netdev->stats.rx_errors; > + stats->rx_dropped = mhi_netdev->stats.rx_dropped; > + stats->tx_packets = mhi_netdev->stats.tx_packets; > + stats->tx_bytes = mhi_netdev->stats.tx_bytes; > + stats->tx_errors = mhi_netdev->stats.tx_errors; > + stats->tx_dropped = mhi_netdev->stats.tx_dropped; > +} > + > +static const struct net_device_ops mhi_netdev_ops = { > + .ndo_open = mhi_ndo_open, > + .ndo_stop = mhi_ndo_stop, > + .ndo_start_xmit = mhi_ndo_xmit, > + .ndo_get_stats64 = mhi_ndo_get_stats64, > +}; > + > +static void mhi_net_setup(struct net_device *ndev) > +{ > + ndev->header_ops = NULL; /* No header */ > + ndev->type = ARPHRD_NONE; /* QMAP... */ > + ndev->hard_header_len = 0; > + ndev->addr_len = 0; > + ndev->flags &= ~(IFF_BROADCAST | IFF_MULTICAST); > + ndev->netdev_ops = &mhi_netdev_ops; > + ndev->mtu = DEFAULT_MTU; > + ndev->min_mtu = MIN_MTU; > + ndev->max_mtu = MAX_MTU; > + ndev->needs_free_netdev = true; > + ndev->tx_queue_len = 1000; > +} > + > +static void mhi_net_dl_callback(struct mhi_device *mhi_dev, > + struct mhi_result *mhi_res) > +{ > + struct mhi_net_dev *mhi_netdev = dev_get_drvdata(&mhi_dev- > >dev); > + struct sk_buff *skb = mhi_res->buf_addr; > + > + atomic_dec(&mhi_netdev->stats.rx_queued); > + > + if (mhi_res->transaction_status) { > + mhi_netdev->stats.rx_errors++; > + kfree_skb(skb); > + } else { > + mhi_netdev->stats.rx_packets++; > + mhi_netdev->stats.rx_bytes += mhi_res->bytes_xferd; > + > + skb->protocol = htons(ETH_P_MAP); > + skb_put(skb, mhi_res->bytes_xferd); > + netif_rx(skb); > + } > + > + schedule_delayed_work(&mhi_netdev->rx_refill, 0); > +} > + > +static void mhi_net_ul_callback(struct mhi_device *mhi_dev, > + struct mhi_result *mhi_res) > +{ > + struct mhi_net_dev *mhi_netdev = dev_get_drvdata(&mhi_dev- > >dev); > + struct net_device *ndev = mhi_netdev->ndev; > + struct sk_buff *skb = mhi_res->buf_addr; > + > + /* Hardware has consumed the buffer, so free the skb (which is > not > + * freed by the MHI stack) and perform accounting. > + */ > + consume_skb(skb); > + > + if (unlikely(mhi_res->transaction_status)) { > + mhi_netdev->stats.tx_errors++; > + } else { > + mhi_netdev->stats.tx_packets++; > + mhi_netdev->stats.tx_bytes += mhi_res->bytes_xferd; > + } > + > + if (netif_queue_stopped(ndev)) > + netif_wake_queue(ndev); > +} > + > +static void mhi_net_rx_refill_work(struct work_struct *work) > +{ > + struct mhi_net_dev *mhi_netdev = container_of(work, struct > mhi_net_dev, > + rx_refill.work); > + struct net_device *ndev = mhi_netdev->ndev; > + struct mhi_device *mdev = mhi_netdev->mdev; > + struct sk_buff *skb; > + int err; > + > + if (!netif_running(ndev)) > + return; > + > + do { > + skb = netdev_alloc_skb(ndev, READ_ONCE(ndev->mtu)); > + if (unlikely(!skb)) { > + /* If we are starved of RX buffers, retry later > */ > + if (!atomic_read(&mhi_netdev->stats.rx_queued)) > + schedule_delayed_work(&mhi_netdev- > >rx_refill, HZ / 2); > + break; > + } > + > + err = mhi_queue_skb(mdev, DMA_FROM_DEVICE, skb, ndev- > >mtu, > + MHI_EOT); > + if (err) { > + atomic_dec(&mhi_netdev->stats.rx_queued); > + kfree_skb(skb); > + break; > + } > + > + atomic_inc(&mhi_netdev->stats.rx_queued); > + > + } while (1); > +} > + > +static int mhi_net_probe(struct mhi_device *mhi_dev, > + const struct mhi_device_id *id) > +{ > + const char *netname = (char *)id->driver_data; > + struct mhi_net_dev *mhi_netdev; > + struct net_device *ndev; > + struct device *dev = &mhi_dev->dev; > + int err; > + > + ndev = alloc_netdev(sizeof(*mhi_netdev), netname, > NET_NAME_PREDICTABLE, > + mhi_net_setup); > + if (!ndev) { > + err = -ENOMEM; > + return err; > + } > + > + mhi_netdev = netdev_priv(ndev); > + dev_set_drvdata(dev, mhi_netdev); > + mhi_netdev->ndev = ndev; > + mhi_netdev->mdev = mhi_dev; SET_NETDEV_DEV(netdev, &mhi_dev->dev); ? Let's at least set the sysfs links up right... > + > + INIT_DELAYED_WORK(&mhi_netdev->rx_refill, > mhi_net_rx_refill_work); > + > + /* Start MHI channels */ > + err = mhi_prepare_for_transfer(mhi_dev, 0); > + if (err) { > + free_netdev(ndev); > + return err; > + } > + > + err = register_netdev(ndev); > + if (err) { > + dev_err(dev, "mhi_net: registering device failed\n"); > + free_netdev(ndev); > + return -EINVAL; > + } > + > + return 0; > +} > + > +static void mhi_net_remove(struct mhi_device *mhi_dev) > +{ > + struct mhi_net_dev *mhi_netdev = dev_get_drvdata(&mhi_dev- > >dev); > + > + mhi_unprepare_from_transfer(mhi_netdev->mdev); > + unregister_netdev(mhi_netdev->ndev); > + /* netdev released from unregister */ > +} > + > +static const struct mhi_device_id mhi_net_id_table[] = { > + { .chan = "IP_HW0", .driver_data = (kernel_ulong_t)"mhi_hwip%d" > }, > + { .chan = "IP_SW0", .driver_data = (kernel_ulong_t)"mhi_swip%d" > }, > + {} > +}; > +MODULE_DEVICE_TABLE(mhi, mhi_net_id_table); > + > +static struct mhi_driver mhi_net_driver = { > + .probe = mhi_net_probe, > + .remove = mhi_net_remove, > + .dl_xfer_cb = mhi_net_dl_callback, > + .ul_xfer_cb = mhi_net_ul_callback, > + .id_table = mhi_net_id_table, > + .driver = { > + .name = "mhi_net", > + }, > +}; > + > +module_mhi_driver(mhi_net_driver); > + > +MODULE_AUTHOR("Loic Poulain <loic.poulain@linaro.org>"); > +MODULE_DESCRIPTION("Network over MHI"); > +MODULE_LICENSE("GPL v2");
diff --git a/drivers/net/Kconfig b/drivers/net/Kconfig index 1368d1d..11a6357 100644 --- a/drivers/net/Kconfig +++ b/drivers/net/Kconfig @@ -426,6 +426,13 @@ config VSOCKMON mostly intended for developers or support to debug vsock issues. If unsure, say N. +config MHI_NET + tristate "MHI network driver" + depends on MHI_BUS + help + This is the network driver for MHI. It can be used with + QCOM based WWAN modems (like SDX55). Say Y or M. + endif # NET_CORE config SUNGEM_PHY diff --git a/drivers/net/Makefile b/drivers/net/Makefile index 94b6080..8312037 100644 --- a/drivers/net/Makefile +++ b/drivers/net/Makefile @@ -34,6 +34,7 @@ obj-$(CONFIG_GTP) += gtp.o obj-$(CONFIG_NLMON) += nlmon.o obj-$(CONFIG_NET_VRF) += vrf.o obj-$(CONFIG_VSOCKMON) += vsockmon.o +obj-$(CONFIG_MHI_NET) += mhi_net.o # # Networking Drivers diff --git a/drivers/net/mhi_net.c b/drivers/net/mhi_net.c new file mode 100644 index 0000000..f66248e --- /dev/null +++ b/drivers/net/mhi_net.c @@ -0,0 +1,281 @@ +// SPDX-License-Identifier: GPL-2.0-or-later +/* MHI Network driver - Network over MHI + * + * Copyright (C) 2020 Linaro Ltd <loic.poulain@linaro.org> + */ + +#include <linux/if_arp.h> +#include <linux/mhi.h> +#include <linux/mod_devicetable.h> +#include <linux/module.h> +#include <linux/netdevice.h> +#include <linux/skbuff.h> + +#define MIN_MTU ETH_MIN_MTU +#define MAX_MTU 0xffff +#define DEFAULT_MTU 8192 + +struct mhi_net_stats { + u64 rx_packets; + u64 rx_bytes; + u64 rx_errors; + u64 rx_dropped; + u64 tx_packets; + u64 tx_bytes; + u64 tx_errors; + u64 tx_dropped; + atomic_t rx_queued; +}; + +struct mhi_net_dev { + struct mhi_device *mdev; + struct net_device *ndev; + struct delayed_work rx_refill; + struct mhi_net_stats stats; +}; + +static int mhi_ndo_open(struct net_device *ndev) +{ + struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); + + /* Feed the rx buffer pool */ + schedule_delayed_work(&mhi_netdev->rx_refill, 0); + + /* Carrier is established via out-of-band channel (e.g. qmi) */ + netif_carrier_on(ndev); + + netif_start_queue(ndev); + + return 0; +} + +static int mhi_ndo_stop(struct net_device *ndev) +{ + struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); + + netif_stop_queue(ndev); + netif_carrier_off(ndev); + cancel_delayed_work_sync(&mhi_netdev->rx_refill); + + return 0; +} + +static int mhi_ndo_xmit(struct sk_buff *skb, struct net_device *ndev) +{ + struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); + struct mhi_device *mdev = mhi_netdev->mdev; + int err; + + /* Only support for single buffer transfer for now */ + err = skb_linearize(skb); + if (unlikely(err)) { + kfree_skb(skb); + mhi_netdev->stats.tx_dropped++; + return NETDEV_TX_OK; + } + + skb_tx_timestamp(skb); + + /* mhi_queue_skb is not thread-safe, but xmit is serialized by the + * network core. Once MHI core will be thread save, migrate to + * NETIF_F_LLTX support. + */ + err = mhi_queue_skb(mdev, DMA_TO_DEVICE, skb, skb->len, MHI_EOT); + if (err) { + netdev_err(ndev, "mhi_queue_skb err %d\n", err); + netif_stop_queue(ndev); + return (err == -ENOMEM) ? NETDEV_TX_BUSY : err; + } + + return NETDEV_TX_OK; +} + +static void mhi_ndo_get_stats64(struct net_device *ndev, + struct rtnl_link_stats64 *stats) +{ + struct mhi_net_dev *mhi_netdev = netdev_priv(ndev); + + stats->rx_packets = mhi_netdev->stats.rx_packets; + stats->rx_bytes = mhi_netdev->stats.rx_bytes; + stats->rx_errors = mhi_netdev->stats.rx_errors; + stats->rx_dropped = mhi_netdev->stats.rx_dropped; + stats->tx_packets = mhi_netdev->stats.tx_packets; + stats->tx_bytes = mhi_netdev->stats.tx_bytes; + stats->tx_errors = mhi_netdev->stats.tx_errors; + stats->tx_dropped = mhi_netdev->stats.tx_dropped; +} + +static const struct net_device_ops mhi_netdev_ops = { + .ndo_open = mhi_ndo_open, + .ndo_stop = mhi_ndo_stop, + .ndo_start_xmit = mhi_ndo_xmit, + .ndo_get_stats64 = mhi_ndo_get_stats64, +}; + +static void mhi_net_setup(struct net_device *ndev) +{ + ndev->header_ops = NULL; /* No header */ + ndev->type = ARPHRD_NONE; /* QMAP... */ + ndev->hard_header_len = 0; + ndev->addr_len = 0; + ndev->flags &= ~(IFF_BROADCAST | IFF_MULTICAST); + ndev->netdev_ops = &mhi_netdev_ops; + ndev->mtu = DEFAULT_MTU; + ndev->min_mtu = MIN_MTU; + ndev->max_mtu = MAX_MTU; + ndev->needs_free_netdev = true; + ndev->tx_queue_len = 1000; +} + +static void mhi_net_dl_callback(struct mhi_device *mhi_dev, + struct mhi_result *mhi_res) +{ + struct mhi_net_dev *mhi_netdev = dev_get_drvdata(&mhi_dev->dev); + struct sk_buff *skb = mhi_res->buf_addr; + + atomic_dec(&mhi_netdev->stats.rx_queued); + + if (mhi_res->transaction_status) { + mhi_netdev->stats.rx_errors++; + kfree_skb(skb); + } else { + mhi_netdev->stats.rx_packets++; + mhi_netdev->stats.rx_bytes += mhi_res->bytes_xferd; + + skb->protocol = htons(ETH_P_MAP); + skb_put(skb, mhi_res->bytes_xferd); + netif_rx(skb); + } + + schedule_delayed_work(&mhi_netdev->rx_refill, 0); +} + +static void mhi_net_ul_callback(struct mhi_device *mhi_dev, + struct mhi_result *mhi_res) +{ + struct mhi_net_dev *mhi_netdev = dev_get_drvdata(&mhi_dev->dev); + struct net_device *ndev = mhi_netdev->ndev; + struct sk_buff *skb = mhi_res->buf_addr; + + /* Hardware has consumed the buffer, so free the skb (which is not + * freed by the MHI stack) and perform accounting. + */ + consume_skb(skb); + + if (unlikely(mhi_res->transaction_status)) { + mhi_netdev->stats.tx_errors++; + } else { + mhi_netdev->stats.tx_packets++; + mhi_netdev->stats.tx_bytes += mhi_res->bytes_xferd; + } + + if (netif_queue_stopped(ndev)) + netif_wake_queue(ndev); +} + +static void mhi_net_rx_refill_work(struct work_struct *work) +{ + struct mhi_net_dev *mhi_netdev = container_of(work, struct mhi_net_dev, + rx_refill.work); + struct net_device *ndev = mhi_netdev->ndev; + struct mhi_device *mdev = mhi_netdev->mdev; + struct sk_buff *skb; + int err; + + if (!netif_running(ndev)) + return; + + do { + skb = netdev_alloc_skb(ndev, READ_ONCE(ndev->mtu)); + if (unlikely(!skb)) { + /* If we are starved of RX buffers, retry later */ + if (!atomic_read(&mhi_netdev->stats.rx_queued)) + schedule_delayed_work(&mhi_netdev->rx_refill, HZ / 2); + break; + } + + err = mhi_queue_skb(mdev, DMA_FROM_DEVICE, skb, ndev->mtu, + MHI_EOT); + if (err) { + atomic_dec(&mhi_netdev->stats.rx_queued); + kfree_skb(skb); + break; + } + + atomic_inc(&mhi_netdev->stats.rx_queued); + + } while (1); +} + +static int mhi_net_probe(struct mhi_device *mhi_dev, + const struct mhi_device_id *id) +{ + const char *netname = (char *)id->driver_data; + struct mhi_net_dev *mhi_netdev; + struct net_device *ndev; + struct device *dev = &mhi_dev->dev; + int err; + + ndev = alloc_netdev(sizeof(*mhi_netdev), netname, NET_NAME_PREDICTABLE, + mhi_net_setup); + if (!ndev) { + err = -ENOMEM; + return err; + } + + mhi_netdev = netdev_priv(ndev); + dev_set_drvdata(dev, mhi_netdev); + mhi_netdev->ndev = ndev; + mhi_netdev->mdev = mhi_dev; + + INIT_DELAYED_WORK(&mhi_netdev->rx_refill, mhi_net_rx_refill_work); + + /* Start MHI channels */ + err = mhi_prepare_for_transfer(mhi_dev, 0); + if (err) { + free_netdev(ndev); + return err; + } + + err = register_netdev(ndev); + if (err) { + dev_err(dev, "mhi_net: registering device failed\n"); + free_netdev(ndev); + return -EINVAL; + } + + return 0; +} + +static void mhi_net_remove(struct mhi_device *mhi_dev) +{ + struct mhi_net_dev *mhi_netdev = dev_get_drvdata(&mhi_dev->dev); + + mhi_unprepare_from_transfer(mhi_netdev->mdev); + unregister_netdev(mhi_netdev->ndev); + /* netdev released from unregister */ +} + +static const struct mhi_device_id mhi_net_id_table[] = { + { .chan = "IP_HW0", .driver_data = (kernel_ulong_t)"mhi_hwip%d" }, + { .chan = "IP_SW0", .driver_data = (kernel_ulong_t)"mhi_swip%d" }, + {} +}; +MODULE_DEVICE_TABLE(mhi, mhi_net_id_table); + +static struct mhi_driver mhi_net_driver = { + .probe = mhi_net_probe, + .remove = mhi_net_remove, + .dl_xfer_cb = mhi_net_dl_callback, + .ul_xfer_cb = mhi_net_ul_callback, + .id_table = mhi_net_id_table, + .driver = { + .name = "mhi_net", + }, +}; + +module_mhi_driver(mhi_net_driver); + +MODULE_AUTHOR("Loic Poulain <loic.poulain@linaro.org>"); +MODULE_DESCRIPTION("Network over MHI"); +MODULE_LICENSE("GPL v2");
This patch adds a new network driver implementing MHI transport for network packets. Packets can be in any format, though QMAP (rmnet) is the usual protocol (flow control + PDN mux). It support two MHI devices, IP_HW0 which is, the path to the IPA (IP accelerator) on qcom modem, And IP_SW0 which is the software driven IP path (to modem CPU). Signed-off-by: Loic Poulain <loic.poulain@linaro.org> --- drivers/net/Kconfig | 7 ++ drivers/net/Makefile | 1 + drivers/net/mhi_net.c | 281 ++++++++++++++++++++++++++++++++++++++++++++++++++ 3 files changed, 289 insertions(+) create mode 100644 drivers/net/mhi_net.c