From: Stanislav Fomichev <sdf@google.com>
To: bpf@vger.kernel.org
Cc: ast@kernel.org, daniel@iogearbox.net, andrii@kernel.org,
martin.lau@linux.dev, song@kernel.org, yhs@fb.com,
john.fastabend@gmail.com, kpsingh@kernel.org, sdf@google.com,
haoluo@google.com, jolsa@kernel.org, netdev@vger.kernel.org
Subject: [RFC bpf-next v2 06/11] net: veth: Implement devtx timestamp kfuncs
Date: Wed, 21 Jun 2023 10:02:39 -0700 [thread overview]
Message-ID: <20230621170244.1283336-7-sdf@google.com> (raw)
In-Reply-To: <20230621170244.1283336-1-sdf@google.com>
Have a software-based example for kfuncs to showcase how it
can be used in the real devices and to have something to
test against in the selftests.
Both path (skb & xdp) are covered. Only the skb path is really
tested though.
Cc: netdev@vger.kernel.org
Signed-off-by: Stanislav Fomichev <sdf@google.com>
---
drivers/net/veth.c | 116 +++++++++++++++++++++++++++++++++++++++++++--
1 file changed, 112 insertions(+), 4 deletions(-)
diff --git a/drivers/net/veth.c b/drivers/net/veth.c
index 614f3e3efab0..632f0f3771e4 100644
--- a/drivers/net/veth.c
+++ b/drivers/net/veth.c
@@ -27,6 +27,7 @@
#include <linux/bpf_trace.h>
#include <linux/net_tstamp.h>
#include <net/page_pool.h>
+#include <net/devtx.h>
#define DRV_NAME "veth"
#define DRV_VERSION "1.0"
@@ -123,6 +124,13 @@ struct veth_xdp_buff {
struct sk_buff *skb;
};
+struct veth_devtx_frame {
+ struct devtx_frame frame;
+ bool request_timestamp;
+ ktime_t xdp_tx_timestamp;
+ struct sk_buff *skb;
+};
+
static int veth_get_link_ksettings(struct net_device *dev,
struct ethtool_link_ksettings *cmd)
{
@@ -313,10 +321,43 @@ static int veth_xdp_rx(struct veth_rq *rq, struct sk_buff *skb)
return NET_RX_SUCCESS;
}
+__weak noinline void veth_devtx_submit(struct devtx_frame *ctx)
+{
+}
+
+__weak noinline void veth_devtx_complete(struct devtx_frame *ctx)
+{
+}
+
+BTF_SET8_START(veth_devtx_hook_ids)
+BTF_ID_FLAGS(func, veth_devtx_submit)
+BTF_ID_FLAGS(func, veth_devtx_complete)
+BTF_SET8_END(veth_devtx_hook_ids)
+
static int veth_forward_skb(struct net_device *dev, struct sk_buff *skb,
- struct veth_rq *rq, bool xdp)
+ struct veth_rq *rq, bool xdp, bool request_timestamp)
{
- return __dev_forward_skb(dev, skb) ?: xdp ?
+ struct net_device *orig_dev = skb->dev;
+ int ret;
+
+ ret = __dev_forward_skb(dev, skb);
+ if (ret)
+ return ret;
+
+ if (devtx_enabled()) {
+ struct veth_devtx_frame ctx;
+
+ if (unlikely(request_timestamp))
+ __net_timestamp(skb);
+
+ devtx_frame_from_skb(&ctx.frame, skb, orig_dev);
+ ctx.frame.data -= ETH_HLEN; /* undo eth_type_trans pull */
+ ctx.frame.len += ETH_HLEN;
+ ctx.skb = skb;
+ veth_devtx_complete(&ctx.frame);
+ }
+
+ return xdp ?
veth_xdp_rx(rq, skb) :
__netif_rx(skb);
}
@@ -343,6 +384,7 @@ static bool veth_skb_is_eligible_for_gro(const struct net_device *dev,
static netdev_tx_t veth_xmit(struct sk_buff *skb, struct net_device *dev)
{
struct veth_priv *rcv_priv, *priv = netdev_priv(dev);
+ bool request_timestamp = false;
struct veth_rq *rq = NULL;
struct net_device *rcv;
int length = skb->len;
@@ -356,6 +398,15 @@ static netdev_tx_t veth_xmit(struct sk_buff *skb, struct net_device *dev)
goto drop;
}
+ if (devtx_enabled()) {
+ struct veth_devtx_frame ctx;
+
+ devtx_frame_from_skb(&ctx.frame, skb, dev);
+ ctx.request_timestamp = false;
+ veth_devtx_submit(&ctx.frame);
+ request_timestamp = ctx.request_timestamp;
+ }
+
rcv_priv = netdev_priv(rcv);
rxq = skb_get_queue_mapping(skb);
if (rxq < rcv->real_num_rx_queues) {
@@ -370,7 +421,7 @@ static netdev_tx_t veth_xmit(struct sk_buff *skb, struct net_device *dev)
}
skb_tx_timestamp(skb);
- if (likely(veth_forward_skb(rcv, skb, rq, use_napi) == NET_RX_SUCCESS)) {
+ if (likely(veth_forward_skb(rcv, skb, rq, use_napi, request_timestamp) == NET_RX_SUCCESS)) {
if (!use_napi)
dev_lstats_add(dev, length);
} else {
@@ -483,6 +534,7 @@ static int veth_xdp_xmit(struct net_device *dev, int n,
{
struct veth_priv *rcv_priv, *priv = netdev_priv(dev);
int i, ret = -ENXIO, nxmit = 0;
+ ktime_t tx_timestamp = 0;
struct net_device *rcv;
unsigned int max_len;
struct veth_rq *rq;
@@ -511,9 +563,32 @@ static int veth_xdp_xmit(struct net_device *dev, int n,
void *ptr = veth_xdp_to_ptr(frame);
if (unlikely(xdp_get_frame_len(frame) > max_len ||
- __ptr_ring_produce(&rq->xdp_ring, ptr)))
+ __ptr_ring_full(&rq->xdp_ring)))
+ break;
+
+ if (devtx_enabled()) {
+ struct veth_devtx_frame ctx;
+
+ devtx_frame_from_xdp(&ctx.frame, frame, dev);
+ ctx.request_timestamp = false;
+ veth_devtx_submit(&ctx.frame);
+
+ if (unlikely(ctx.request_timestamp))
+ tx_timestamp = ktime_get_real();
+ }
+
+ if (unlikely(__ptr_ring_produce(&rq->xdp_ring, ptr)))
break;
nxmit++;
+
+ if (devtx_enabled()) {
+ struct veth_devtx_frame ctx;
+
+ devtx_frame_from_xdp(&ctx.frame, frame, dev);
+ ctx.xdp_tx_timestamp = tx_timestamp;
+ ctx.skb = NULL;
+ veth_devtx_complete(&ctx.frame);
+ }
}
spin_unlock(&rq->xdp_ring.producer_lock);
@@ -1732,6 +1807,28 @@ static int veth_xdp_rx_hash(const struct xdp_md *ctx, u32 *hash,
return 0;
}
+static int veth_devtx_sb_request_timestamp(const struct devtx_frame *_ctx)
+{
+ struct veth_devtx_frame *ctx = (struct veth_devtx_frame *)_ctx;
+
+ ctx->request_timestamp = true;
+
+ return 0;
+}
+
+static int veth_devtx_cp_timestamp(const struct devtx_frame *_ctx, u64 *timestamp)
+{
+ struct veth_devtx_frame *ctx = (struct veth_devtx_frame *)_ctx;
+
+ if (ctx->skb) {
+ *timestamp = ctx->skb->tstamp;
+ return 0;
+ }
+
+ *timestamp = ctx->xdp_tx_timestamp;
+ return 0;
+}
+
static const struct net_device_ops veth_netdev_ops = {
.ndo_init = veth_dev_init,
.ndo_open = veth_open,
@@ -1756,6 +1853,8 @@ static const struct net_device_ops veth_netdev_ops = {
static const struct xdp_metadata_ops veth_xdp_metadata_ops = {
.xmo_rx_timestamp = veth_xdp_rx_timestamp,
.xmo_rx_hash = veth_xdp_rx_hash,
+ .xmo_sb_request_timestamp = veth_devtx_sb_request_timestamp,
+ .xmo_cp_timestamp = veth_devtx_cp_timestamp,
};
#define VETH_FEATURES (NETIF_F_SG | NETIF_F_FRAGLIST | NETIF_F_HW_CSUM | \
@@ -2041,11 +2140,20 @@ static struct rtnl_link_ops veth_link_ops = {
static __init int veth_init(void)
{
+ int ret;
+
+ ret = devtx_hooks_register(&veth_devtx_hook_ids, &veth_xdp_metadata_ops);
+ if (ret) {
+ pr_warn("failed to register devtx hooks: %d", ret);
+ return ret;
+ }
+
return rtnl_link_register(&veth_link_ops);
}
static __exit void veth_exit(void)
{
+ devtx_hooks_unregister(&veth_devtx_hook_ids);
rtnl_link_unregister(&veth_link_ops);
}
--
2.41.0.162.gfafddb0af9-goog
next prev parent reply other threads:[~2023-06-21 17:02 UTC|newest]
Thread overview: 51+ messages / expand[flat|nested] mbox.gz Atom feed top
2023-06-21 17:02 [RFC bpf-next v2 00/11] bpf: Netdev TX metadata Stanislav Fomichev
2023-06-21 17:02 ` [RFC bpf-next v2 01/11] bpf: Rename some xdp-metadata functions into dev-bound Stanislav Fomichev
2023-06-21 17:02 ` [RFC bpf-next v2 02/11] bpf: Resolve single typedef when walking structs Stanislav Fomichev
2023-06-22 5:17 ` Alexei Starovoitov
2023-06-22 17:55 ` Stanislav Fomichev
2023-06-21 17:02 ` [RFC bpf-next v2 04/11] bpf: Implement devtx hook points Stanislav Fomichev
2023-06-21 17:02 ` [RFC bpf-next v2 05/11] bpf: Implement devtx timestamp kfunc Stanislav Fomichev
2023-06-22 12:07 ` Jesper D. Brouer
2023-06-22 17:55 ` Stanislav Fomichev
2023-06-21 17:02 ` Stanislav Fomichev [this message]
2023-06-23 23:29 ` [RFC bpf-next v2 06/11] net: veth: Implement devtx timestamp kfuncs Vinicius Costa Gomes
2023-06-26 17:00 ` Stanislav Fomichev
2023-06-26 22:00 ` Vinicius Costa Gomes
2023-06-26 23:29 ` Stanislav Fomichev
2023-06-27 1:38 ` Vinicius Costa Gomes
2023-06-21 17:02 ` [RFC bpf-next v2 09/11] selftests/bpf: Extend xdp_metadata with devtx kfuncs Stanislav Fomichev
2023-06-23 11:12 ` Jesper D. Brouer
2023-06-23 17:40 ` Stanislav Fomichev
2023-06-21 17:02 ` [RFC bpf-next v2 10/11] selftests/bpf: Extend xdp_hw_metadata " Stanislav Fomichev
2023-06-21 17:02 ` [RFC bpf-next v2 11/11] net/mlx5e: Support TX timestamp metadata Stanislav Fomichev
2023-06-22 19:57 ` Alexei Starovoitov
2023-06-22 20:13 ` Stanislav Fomichev
2023-06-22 21:47 ` Alexei Starovoitov
2023-06-22 22:13 ` Stanislav Fomichev
2023-06-23 2:35 ` Alexei Starovoitov
2023-06-23 10:16 ` Maryam Tahhan
2023-06-23 16:32 ` Alexei Starovoitov
2023-06-23 17:47 ` Maryam Tahhan
2023-06-23 17:24 ` Stanislav Fomichev
2023-06-23 18:57 ` Donald Hunter
2023-06-24 0:25 ` John Fastabend
2023-06-24 2:52 ` Alexei Starovoitov
2023-06-24 21:38 ` Jakub Kicinski
2023-06-25 1:12 ` Stanislav Fomichev
2023-06-26 21:36 ` Stanislav Fomichev
2023-06-26 22:37 ` Alexei Starovoitov
2023-06-26 23:29 ` Stanislav Fomichev
2023-06-27 13:35 ` Toke Høiland-Jørgensen
2023-06-27 21:43 ` John Fastabend
2023-06-27 22:56 ` Stanislav Fomichev
2023-06-27 23:33 ` John Fastabend
2023-06-27 23:50 ` Alexei Starovoitov
2023-06-28 18:52 ` Jakub Kicinski
2023-06-29 11:43 ` Toke Høiland-Jørgensen
2023-06-30 18:54 ` Stanislav Fomichev
2023-07-01 0:52 ` John Fastabend
2023-07-01 3:11 ` Jakub Kicinski
2023-07-03 18:30 ` John Fastabend
2023-07-03 19:33 ` Jakub Kicinski
2023-06-22 8:41 ` [RFC bpf-next v2 00/11] bpf: Netdev TX metadata Jesper Dangaard Brouer
2023-06-22 17:55 ` Stanislav Fomichev
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20230621170244.1283336-7-sdf@google.com \
--to=sdf@google.com \
--cc=andrii@kernel.org \
--cc=ast@kernel.org \
--cc=bpf@vger.kernel.org \
--cc=daniel@iogearbox.net \
--cc=haoluo@google.com \
--cc=john.fastabend@gmail.com \
--cc=jolsa@kernel.org \
--cc=kpsingh@kernel.org \
--cc=martin.lau@linux.dev \
--cc=netdev@vger.kernel.org \
--cc=song@kernel.org \
--cc=yhs@fb.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).