1 // SPDX-License-Identifier: GPL-2.0-or-later
5 * 2004+ Copyright (c) Evgeniy Polyakov <zbr@ioremap.net>
9 #include <linux/compiler.h>
10 #include <linux/kernel.h>
11 #include <linux/module.h>
12 #include <linux/list.h>
13 #include <linux/skbuff.h>
14 #include <net/netlink.h>
15 #include <linux/moduleparam.h>
16 #include <linux/connector.h>
17 #include <linux/slab.h>
18 #include <linux/mutex.h>
19 #include <linux/proc_fs.h>
20 #include <linux/spinlock.h>
24 MODULE_LICENSE("GPL");
25 MODULE_AUTHOR("Evgeniy Polyakov <zbr@ioremap.net>");
26 MODULE_DESCRIPTION("Generic userspace <-> kernelspace connector.");
27 MODULE_ALIAS_NET_PF_PROTO(PF_NETLINK, NETLINK_CONNECTOR);
29 static struct cn_dev cdev;
31 static int cn_already_initialized;
34 * Sends mult (multiple) cn_msg at a time.
36 * msg->seq and msg->ack are used to determine message genealogy.
37 * When someone sends message it puts there locally unique sequence
38 * and random acknowledge numbers. Sequence number may be copied into
39 * nlmsghdr->nlmsg_seq too.
41 * Sequence number is incremented with each message to be sent.
43 * If we expect a reply to our message then the sequence number in
44 * received message MUST be the same as in original message, and
45 * acknowledge number MUST be the same + 1.
47 * If we receive a message and its sequence number is not equal to the
48 * one we are expecting then it is a new message.
50 * If we receive a message and its sequence number is the same as one
51 * we are expecting but it's acknowledgement number is not equal to
52 * the acknowledgement number in the original message + 1, then it is
55 * If msg->len != len, then additional cn_msg messages are expected following
58 * The message is sent to, the portid if given, the group if given, both if
59 * both, or if both are zero then the group is looked up and sent there.
61 int cn_netlink_send_mult(struct cn_msg *msg, u16 len, u32 portid, u32 __group,
64 struct cn_callback_entry *__cbq;
69 struct cn_dev *dev = &cdev;
73 if (portid || __group) {
76 spin_lock_bh(&dev->cbdev->queue_lock);
77 list_for_each_entry(__cbq, &dev->cbdev->queue_list,
79 if (cn_cb_equal(&__cbq->id.id, &msg->id)) {
85 spin_unlock_bh(&dev->cbdev->queue_lock);
91 if (!portid && !netlink_has_listeners(dev->nls, group))
94 size = sizeof(*msg) + len;
96 skb = nlmsg_new(size, gfp_mask);
100 nlh = nlmsg_put(skb, 0, msg->seq, NLMSG_DONE, size, 0);
106 data = nlmsg_data(nlh);
108 memcpy(data, msg, size);
110 NETLINK_CB(skb).dst_group = group;
113 return netlink_broadcast(dev->nls, skb, portid, group,
115 return netlink_unicast(dev->nls, skb, portid,
116 !gfpflags_allow_blocking(gfp_mask));
118 EXPORT_SYMBOL_GPL(cn_netlink_send_mult);
120 /* same as cn_netlink_send_mult except msg->len is used for len */
121 int cn_netlink_send(struct cn_msg *msg, u32 portid, u32 __group,
124 return cn_netlink_send_mult(msg, msg->len, portid, __group, gfp_mask);
126 EXPORT_SYMBOL_GPL(cn_netlink_send);
129 * Callback helper - queues work and setup destructor for given data.
131 static int cn_call_callback(struct sk_buff *skb)
133 struct nlmsghdr *nlh;
134 struct cn_callback_entry *i, *cbq = NULL;
135 struct cn_dev *dev = &cdev;
136 struct cn_msg *msg = nlmsg_data(nlmsg_hdr(skb));
137 struct netlink_skb_parms *nsp = &NETLINK_CB(skb);
140 /* verify msg->len is within skb */
141 nlh = nlmsg_hdr(skb);
142 if (nlh->nlmsg_len < NLMSG_HDRLEN + sizeof(struct cn_msg) + msg->len)
145 spin_lock_bh(&dev->cbdev->queue_lock);
146 list_for_each_entry(i, &dev->cbdev->queue_list, callback_entry) {
147 if (cn_cb_equal(&i->id.id, &msg->id)) {
148 refcount_inc(&i->refcnt);
153 spin_unlock_bh(&dev->cbdev->queue_lock);
156 cbq->callback(msg, nsp);
158 cn_queue_release_callback(cbq);
166 * Main netlink receiving function.
168 * It checks skb, netlink header and msg sizes, and calls callback helper.
170 static void cn_rx_skb(struct sk_buff *skb)
172 struct nlmsghdr *nlh;
175 if (skb->len >= NLMSG_HDRLEN) {
176 nlh = nlmsg_hdr(skb);
177 len = nlmsg_len(nlh);
179 if (len < (int)sizeof(struct cn_msg) ||
180 skb->len < nlh->nlmsg_len ||
181 len > CONNECTOR_MAX_MSG_SIZE)
184 err = cn_call_callback(skb_get(skb));
191 * Callback add routing - adds callback with given ID and name.
192 * If there is registered callback with the same ID it will not be added.
196 int cn_add_callback(const struct cb_id *id, const char *name,
197 void (*callback)(struct cn_msg *,
198 struct netlink_skb_parms *))
200 struct cn_dev *dev = &cdev;
202 if (!cn_already_initialized)
205 return cn_queue_add_callback(dev->cbdev, name, id, callback);
207 EXPORT_SYMBOL_GPL(cn_add_callback);
210 * Callback remove routing - removes callback
212 * If there is no registered callback with given
213 * ID nothing happens.
215 * May sleep while waiting for reference counter to become zero.
217 void cn_del_callback(const struct cb_id *id)
219 struct cn_dev *dev = &cdev;
221 cn_queue_del_callback(dev->cbdev, id);
223 EXPORT_SYMBOL_GPL(cn_del_callback);
225 static int __maybe_unused cn_proc_show(struct seq_file *m, void *v)
227 struct cn_queue_dev *dev = cdev.cbdev;
228 struct cn_callback_entry *cbq;
230 seq_printf(m, "Name ID\n");
232 spin_lock_bh(&dev->queue_lock);
234 list_for_each_entry(cbq, &dev->queue_list, callback_entry) {
235 seq_printf(m, "%-15s %u:%u\n",
241 spin_unlock_bh(&dev->queue_lock);
246 static int cn_init(void)
248 struct cn_dev *dev = &cdev;
249 struct netlink_kernel_cfg cfg = {
250 .groups = CN_NETLINK_USERS + 0xf,
254 dev->nls = netlink_kernel_create(&init_net, NETLINK_CONNECTOR, &cfg);
258 dev->cbdev = cn_queue_alloc_dev("cqueue", dev->nls);
260 netlink_kernel_release(dev->nls);
264 cn_already_initialized = 1;
266 proc_create_single("connector", S_IRUGO, init_net.proc_net, cn_proc_show);
271 static void cn_fini(void)
273 struct cn_dev *dev = &cdev;
275 cn_already_initialized = 0;
277 remove_proc_entry("connector", init_net.proc_net);
279 cn_queue_free_dev(dev->cbdev);
280 netlink_kernel_release(dev->nls);
283 subsys_initcall(cn_init);
284 module_exit(cn_fini);