1 // SPDX-License-Identifier: GPL-2.0
2 /* XDP sockets monitoring support
4 * Copyright(c) 2019 Intel Corporation.
6 * Author: Björn Töpel <bjorn.topel@intel.com>
9 #include <linux/module.h>
10 #include <net/xdp_sock.h>
11 #include <linux/xdp_diag.h>
12 #include <linux/sock_diag.h>
14 #include "xsk_queue.h"
17 static int xsk_diag_put_info(const struct xdp_sock *xs, struct sk_buff *nlskb)
19 struct xdp_diag_info di = {};
21 di.ifindex = xs->dev ? xs->dev->ifindex : 0;
22 di.queue_id = xs->queue_id;
23 return nla_put(nlskb, XDP_DIAG_INFO, sizeof(di), &di);
26 static int xsk_diag_put_ring(const struct xsk_queue *queue, int nl_type,
27 struct sk_buff *nlskb)
29 struct xdp_diag_ring dr = {};
31 dr.entries = queue->nentries;
32 return nla_put(nlskb, nl_type, sizeof(dr), &dr);
35 static int xsk_diag_put_rings_cfg(const struct xdp_sock *xs,
36 struct sk_buff *nlskb)
41 err = xsk_diag_put_ring(xs->rx, XDP_DIAG_RX_RING, nlskb);
43 err = xsk_diag_put_ring(xs->tx, XDP_DIAG_TX_RING, nlskb);
47 static int xsk_diag_put_umem(const struct xdp_sock *xs, struct sk_buff *nlskb)
49 struct xdp_umem *umem = xs->umem;
50 struct xdp_diag_umem du = {};
58 du.num_pages = umem->npgs;
59 du.chunk_size = (__u32)(~umem->chunk_mask + 1);
60 du.headroom = umem->headroom;
61 du.ifindex = umem->dev ? umem->dev->ifindex : 0;
62 du.queue_id = umem->queue_id;
65 du.flags |= XDP_DU_F_ZEROCOPY;
66 du.refs = refcount_read(&umem->users);
68 err = nla_put(nlskb, XDP_DIAG_UMEM, sizeof(du), &du);
71 err = xsk_diag_put_ring(xs->tx, XDP_DIAG_UMEM_FILL_RING, nlskb);
72 if (!err && umem->cq) {
73 err = xsk_diag_put_ring(xs->tx, XDP_DIAG_UMEM_COMPLETION_RING,
79 static int xsk_diag_fill(struct sock *sk, struct sk_buff *nlskb,
80 struct xdp_diag_req *req,
81 struct user_namespace *user_ns,
82 u32 portid, u32 seq, u32 flags, int sk_ino)
84 struct xdp_sock *xs = xdp_sk(sk);
85 struct xdp_diag_msg *msg;
88 nlh = nlmsg_put(nlskb, portid, seq, SOCK_DIAG_BY_FAMILY, sizeof(*msg),
93 msg = nlmsg_data(nlh);
94 memset(msg, 0, sizeof(*msg));
95 msg->xdiag_family = AF_XDP;
96 msg->xdiag_type = sk->sk_type;
97 msg->xdiag_ino = sk_ino;
98 sock_diag_save_cookie(sk, msg->xdiag_cookie);
100 if ((req->xdiag_show & XDP_SHOW_INFO) && xsk_diag_put_info(xs, nlskb))
103 if ((req->xdiag_show & XDP_SHOW_INFO) &&
104 nla_put_u32(nlskb, XDP_DIAG_UID,
105 from_kuid_munged(user_ns, sock_i_uid(sk))))
108 if ((req->xdiag_show & XDP_SHOW_RING_CFG) &&
109 xsk_diag_put_rings_cfg(xs, nlskb))
112 if ((req->xdiag_show & XDP_SHOW_UMEM) &&
113 xsk_diag_put_umem(xs, nlskb))
116 if ((req->xdiag_show & XDP_SHOW_MEMINFO) &&
117 sock_diag_put_meminfo(sk, nlskb, XDP_DIAG_MEMINFO))
120 nlmsg_end(nlskb, nlh);
124 nlmsg_cancel(nlskb, nlh);
128 static int xsk_diag_dump(struct sk_buff *nlskb, struct netlink_callback *cb)
130 struct xdp_diag_req *req = nlmsg_data(cb->nlh);
131 struct net *net = sock_net(nlskb->sk);
132 int num = 0, s_num = cb->args[0];
135 mutex_lock(&net->xdp.lock);
137 sk_for_each(sk, &net->xdp.list) {
138 if (!net_eq(sock_net(sk), net))
143 if (xsk_diag_fill(sk, nlskb, req,
144 sk_user_ns(NETLINK_CB(cb->skb).sk),
145 NETLINK_CB(cb->skb).portid,
146 cb->nlh->nlmsg_seq, NLM_F_MULTI,
147 sock_i_ino(sk)) < 0) {
153 mutex_unlock(&net->xdp.lock);
158 static int xsk_diag_handler_dump(struct sk_buff *nlskb, struct nlmsghdr *hdr)
160 struct netlink_dump_control c = { .dump = xsk_diag_dump };
161 int hdrlen = sizeof(struct xdp_diag_req);
162 struct net *net = sock_net(nlskb->sk);
164 if (nlmsg_len(hdr) < hdrlen)
167 if (!(hdr->nlmsg_flags & NLM_F_DUMP))
170 return netlink_dump_start(net->diag_nlsk, nlskb, hdr, &c);
173 static const struct sock_diag_handler xsk_diag_handler = {
175 .dump = xsk_diag_handler_dump,
178 static int __init xsk_diag_init(void)
180 return sock_diag_register(&xsk_diag_handler);
183 static void __exit xsk_diag_exit(void)
185 sock_diag_unregister(&xsk_diag_handler);
188 module_init(xsk_diag_init);
189 module_exit(xsk_diag_exit);
190 MODULE_LICENSE("GPL");
191 MODULE_ALIAS_NET_PF_PROTO_TYPE(PF_NETLINK, NETLINK_SOCK_DIAG, AF_XDP);