1 // SPDX-License-Identifier: GPL-2.0-only
5 * Copyright (c) 2020, Red Hat. All rights reserved.
6 * Author: Jason Wang <jasowang@redhat.com>
10 #include <linux/module.h>
11 #include <linux/idr.h>
12 #include <linux/slab.h>
13 #include <linux/vdpa.h>
14 #include <uapi/linux/vdpa.h>
15 #include <net/genetlink.h>
16 #include <linux/mod_devicetable.h>
17 #include <linux/virtio_ids.h>
19 static LIST_HEAD(mdev_head);
20 /* A global mutex that protects vdpa management device and device level operations. */
21 static DECLARE_RWSEM(vdpa_dev_lock);
22 static DEFINE_IDA(vdpa_index_ida);
24 void vdpa_set_status(struct vdpa_device *vdev, u8 status)
26 down_write(&vdev->cf_lock);
27 vdev->config->set_status(vdev, status);
28 up_write(&vdev->cf_lock);
30 EXPORT_SYMBOL(vdpa_set_status);
32 static struct genl_family vdpa_nl_family;
34 static int vdpa_dev_probe(struct device *d)
36 struct vdpa_device *vdev = dev_to_vdpa(d);
37 struct vdpa_driver *drv = drv_to_vdpa(vdev->dev.driver);
38 const struct vdpa_config_ops *ops = vdev->config;
39 u32 max_num, min_num = 1;
42 d->dma_mask = &d->coherent_dma_mask;
43 ret = dma_set_mask_and_coherent(d, DMA_BIT_MASK(64));
47 max_num = ops->get_vq_num_max(vdev);
48 if (ops->get_vq_num_min)
49 min_num = ops->get_vq_num_min(vdev);
50 if (max_num < min_num)
53 if (drv && drv->probe)
54 ret = drv->probe(vdev);
59 static void vdpa_dev_remove(struct device *d)
61 struct vdpa_device *vdev = dev_to_vdpa(d);
62 struct vdpa_driver *drv = drv_to_vdpa(vdev->dev.driver);
64 if (drv && drv->remove)
68 static int vdpa_dev_match(struct device *dev, struct device_driver *drv)
70 struct vdpa_device *vdev = dev_to_vdpa(dev);
72 /* Check override first, and if set, only use the named driver */
73 if (vdev->driver_override)
74 return strcmp(vdev->driver_override, drv->name) == 0;
76 /* Currently devices must be supported by all vDPA bus drivers */
80 static ssize_t driver_override_store(struct device *dev,
81 struct device_attribute *attr,
82 const char *buf, size_t count)
84 struct vdpa_device *vdev = dev_to_vdpa(dev);
87 ret = driver_set_override(dev, &vdev->driver_override, buf, count);
94 static ssize_t driver_override_show(struct device *dev,
95 struct device_attribute *attr, char *buf)
97 struct vdpa_device *vdev = dev_to_vdpa(dev);
101 len = snprintf(buf, PAGE_SIZE, "%s\n", vdev->driver_override);
106 static DEVICE_ATTR_RW(driver_override);
108 static struct attribute *vdpa_dev_attrs[] = {
109 &dev_attr_driver_override.attr,
113 static const struct attribute_group vdpa_dev_group = {
114 .attrs = vdpa_dev_attrs,
116 __ATTRIBUTE_GROUPS(vdpa_dev);
118 static struct bus_type vdpa_bus = {
120 .dev_groups = vdpa_dev_groups,
121 .match = vdpa_dev_match,
122 .probe = vdpa_dev_probe,
123 .remove = vdpa_dev_remove,
126 static void vdpa_release_dev(struct device *d)
128 struct vdpa_device *vdev = dev_to_vdpa(d);
129 const struct vdpa_config_ops *ops = vdev->config;
134 ida_simple_remove(&vdpa_index_ida, vdev->index);
135 kfree(vdev->driver_override);
140 * __vdpa_alloc_device - allocate and initilaize a vDPA device
141 * This allows driver to some prepartion after device is
142 * initialized but before registered.
143 * @parent: the parent device
144 * @config: the bus operations that is supported by this device
145 * @ngroups: number of groups supported by this device
146 * @nas: number of address spaces supported by this device
147 * @size: size of the parent structure that contains private data
148 * @name: name of the vdpa device; optional.
149 * @use_va: indicate whether virtual address must be used by this device
151 * Driver should use vdpa_alloc_device() wrapper macro instead of
152 * using this directly.
154 * Return: Returns an error when parent/config/dma_dev is not set or fail to get
157 struct vdpa_device *__vdpa_alloc_device(struct device *parent,
158 const struct vdpa_config_ops *config,
159 unsigned int ngroups, unsigned int nas,
160 size_t size, const char *name,
163 struct vdpa_device *vdev;
169 if (!!config->dma_map != !!config->dma_unmap)
172 /* It should only work for the device that use on-chip IOMMU */
173 if (use_va && !(config->dma_map || config->set_map))
177 vdev = kzalloc(size, GFP_KERNEL);
181 err = ida_alloc(&vdpa_index_ida, GFP_KERNEL);
185 vdev->dev.bus = &vdpa_bus;
186 vdev->dev.parent = parent;
187 vdev->dev.release = vdpa_release_dev;
189 vdev->config = config;
190 vdev->features_valid = false;
191 vdev->use_va = use_va;
192 vdev->ngroups = ngroups;
196 err = dev_set_name(&vdev->dev, "%s", name);
198 err = dev_set_name(&vdev->dev, "vdpa%u", vdev->index);
202 init_rwsem(&vdev->cf_lock);
203 device_initialize(&vdev->dev);
208 ida_simple_remove(&vdpa_index_ida, vdev->index);
214 EXPORT_SYMBOL_GPL(__vdpa_alloc_device);
216 static int vdpa_name_match(struct device *dev, const void *data)
218 struct vdpa_device *vdev = container_of(dev, struct vdpa_device, dev);
220 return (strcmp(dev_name(&vdev->dev), data) == 0);
223 static int __vdpa_register_device(struct vdpa_device *vdev, u32 nvqs)
229 lockdep_assert_held(&vdpa_dev_lock);
230 dev = bus_find_device(&vdpa_bus, NULL, dev_name(&vdev->dev), vdpa_name_match);
235 return device_add(&vdev->dev);
239 * _vdpa_register_device - register a vDPA device with vdpa lock held
240 * Caller must have a succeed call of vdpa_alloc_device() before.
241 * Caller must invoke this routine in the management device dev_add()
242 * callback after setting up valid mgmtdev for this vdpa device.
243 * @vdev: the vdpa device to be registered to vDPA bus
244 * @nvqs: number of virtqueues supported by this device
246 * Return: Returns an error when fail to add device to vDPA bus
248 int _vdpa_register_device(struct vdpa_device *vdev, u32 nvqs)
253 return __vdpa_register_device(vdev, nvqs);
255 EXPORT_SYMBOL_GPL(_vdpa_register_device);
258 * vdpa_register_device - register a vDPA device
259 * Callers must have a succeed call of vdpa_alloc_device() before.
260 * @vdev: the vdpa device to be registered to vDPA bus
261 * @nvqs: number of virtqueues supported by this device
263 * Return: Returns an error when fail to add to vDPA bus
265 int vdpa_register_device(struct vdpa_device *vdev, u32 nvqs)
269 down_write(&vdpa_dev_lock);
270 err = __vdpa_register_device(vdev, nvqs);
271 up_write(&vdpa_dev_lock);
274 EXPORT_SYMBOL_GPL(vdpa_register_device);
277 * _vdpa_unregister_device - unregister a vDPA device
278 * Caller must invoke this routine as part of management device dev_del()
280 * @vdev: the vdpa device to be unregisted from vDPA bus
282 void _vdpa_unregister_device(struct vdpa_device *vdev)
284 lockdep_assert_held(&vdpa_dev_lock);
285 WARN_ON(!vdev->mdev);
286 device_unregister(&vdev->dev);
288 EXPORT_SYMBOL_GPL(_vdpa_unregister_device);
291 * vdpa_unregister_device - unregister a vDPA device
292 * @vdev: the vdpa device to be unregisted from vDPA bus
294 void vdpa_unregister_device(struct vdpa_device *vdev)
296 down_write(&vdpa_dev_lock);
297 device_unregister(&vdev->dev);
298 up_write(&vdpa_dev_lock);
300 EXPORT_SYMBOL_GPL(vdpa_unregister_device);
303 * __vdpa_register_driver - register a vDPA device driver
304 * @drv: the vdpa device driver to be registered
305 * @owner: module owner of the driver
307 * Return: Returns an err when fail to do the registration
309 int __vdpa_register_driver(struct vdpa_driver *drv, struct module *owner)
311 drv->driver.bus = &vdpa_bus;
312 drv->driver.owner = owner;
314 return driver_register(&drv->driver);
316 EXPORT_SYMBOL_GPL(__vdpa_register_driver);
319 * vdpa_unregister_driver - unregister a vDPA device driver
320 * @drv: the vdpa device driver to be unregistered
322 void vdpa_unregister_driver(struct vdpa_driver *drv)
324 driver_unregister(&drv->driver);
326 EXPORT_SYMBOL_GPL(vdpa_unregister_driver);
329 * vdpa_mgmtdev_register - register a vdpa management device
331 * @mdev: Pointer to vdpa management device
332 * vdpa_mgmtdev_register() register a vdpa management device which supports
333 * vdpa device management.
334 * Return: Returns 0 on success or failure when required callback ops are not
337 int vdpa_mgmtdev_register(struct vdpa_mgmt_dev *mdev)
339 if (!mdev->device || !mdev->ops || !mdev->ops->dev_add || !mdev->ops->dev_del)
342 INIT_LIST_HEAD(&mdev->list);
343 down_write(&vdpa_dev_lock);
344 list_add_tail(&mdev->list, &mdev_head);
345 up_write(&vdpa_dev_lock);
348 EXPORT_SYMBOL_GPL(vdpa_mgmtdev_register);
350 static int vdpa_match_remove(struct device *dev, void *data)
352 struct vdpa_device *vdev = container_of(dev, struct vdpa_device, dev);
353 struct vdpa_mgmt_dev *mdev = vdev->mdev;
356 mdev->ops->dev_del(mdev, vdev);
360 void vdpa_mgmtdev_unregister(struct vdpa_mgmt_dev *mdev)
362 down_write(&vdpa_dev_lock);
364 list_del(&mdev->list);
366 /* Filter out all the entries belong to this management device and delete it. */
367 bus_for_each_dev(&vdpa_bus, NULL, mdev, vdpa_match_remove);
369 up_write(&vdpa_dev_lock);
371 EXPORT_SYMBOL_GPL(vdpa_mgmtdev_unregister);
373 static void vdpa_get_config_unlocked(struct vdpa_device *vdev,
375 void *buf, unsigned int len)
377 const struct vdpa_config_ops *ops = vdev->config;
380 * Config accesses aren't supposed to trigger before features are set.
381 * If it does happen we assume a legacy guest.
383 if (!vdev->features_valid)
384 vdpa_set_features_unlocked(vdev, 0);
385 ops->get_config(vdev, offset, buf, len);
389 * vdpa_get_config - Get one or more device configuration fields.
390 * @vdev: vdpa device to operate on
391 * @offset: starting byte offset of the field
392 * @buf: buffer pointer to read to
393 * @len: length of the configuration fields in bytes
395 void vdpa_get_config(struct vdpa_device *vdev, unsigned int offset,
396 void *buf, unsigned int len)
398 down_read(&vdev->cf_lock);
399 vdpa_get_config_unlocked(vdev, offset, buf, len);
400 up_read(&vdev->cf_lock);
402 EXPORT_SYMBOL_GPL(vdpa_get_config);
405 * vdpa_set_config - Set one or more device configuration fields.
406 * @vdev: vdpa device to operate on
407 * @offset: starting byte offset of the field
408 * @buf: buffer pointer to read from
409 * @length: length of the configuration fields in bytes
411 void vdpa_set_config(struct vdpa_device *vdev, unsigned int offset,
412 const void *buf, unsigned int length)
414 down_write(&vdev->cf_lock);
415 vdev->config->set_config(vdev, offset, buf, length);
416 up_write(&vdev->cf_lock);
418 EXPORT_SYMBOL_GPL(vdpa_set_config);
420 static bool mgmtdev_handle_match(const struct vdpa_mgmt_dev *mdev,
421 const char *busname, const char *devname)
423 /* Bus name is optional for simulated management device, so ignore the
424 * device with bus if bus attribute is provided.
426 if ((busname && !mdev->device->bus) || (!busname && mdev->device->bus))
429 if (!busname && strcmp(dev_name(mdev->device), devname) == 0)
432 if (busname && (strcmp(mdev->device->bus->name, busname) == 0) &&
433 (strcmp(dev_name(mdev->device), devname) == 0))
439 static struct vdpa_mgmt_dev *vdpa_mgmtdev_get_from_attr(struct nlattr **attrs)
441 struct vdpa_mgmt_dev *mdev;
442 const char *busname = NULL;
445 if (!attrs[VDPA_ATTR_MGMTDEV_DEV_NAME])
446 return ERR_PTR(-EINVAL);
447 devname = nla_data(attrs[VDPA_ATTR_MGMTDEV_DEV_NAME]);
448 if (attrs[VDPA_ATTR_MGMTDEV_BUS_NAME])
449 busname = nla_data(attrs[VDPA_ATTR_MGMTDEV_BUS_NAME]);
451 list_for_each_entry(mdev, &mdev_head, list) {
452 if (mgmtdev_handle_match(mdev, busname, devname))
455 return ERR_PTR(-ENODEV);
458 static int vdpa_nl_mgmtdev_handle_fill(struct sk_buff *msg, const struct vdpa_mgmt_dev *mdev)
460 if (mdev->device->bus &&
461 nla_put_string(msg, VDPA_ATTR_MGMTDEV_BUS_NAME, mdev->device->bus->name))
463 if (nla_put_string(msg, VDPA_ATTR_MGMTDEV_DEV_NAME, dev_name(mdev->device)))
468 static u64 vdpa_mgmtdev_get_classes(const struct vdpa_mgmt_dev *mdev,
469 unsigned int *nclasses)
471 u64 supported_classes = 0;
474 for (int i = 0; mdev->id_table[i].device; i++) {
475 if (mdev->id_table[i].device > 63)
477 supported_classes |= BIT_ULL(mdev->id_table[i].device);
483 return supported_classes;
486 static int vdpa_mgmtdev_fill(const struct vdpa_mgmt_dev *mdev, struct sk_buff *msg,
487 u32 portid, u32 seq, int flags)
492 hdr = genlmsg_put(msg, portid, seq, &vdpa_nl_family, flags, VDPA_CMD_MGMTDEV_NEW);
495 err = vdpa_nl_mgmtdev_handle_fill(msg, mdev);
499 if (nla_put_u64_64bit(msg, VDPA_ATTR_MGMTDEV_SUPPORTED_CLASSES,
500 vdpa_mgmtdev_get_classes(mdev, NULL),
505 if (nla_put_u32(msg, VDPA_ATTR_DEV_MGMTDEV_MAX_VQS,
506 mdev->max_supported_vqs)) {
510 if (nla_put_u64_64bit(msg, VDPA_ATTR_DEV_SUPPORTED_FEATURES,
511 mdev->supported_features, VDPA_ATTR_PAD)) {
516 genlmsg_end(msg, hdr);
520 genlmsg_cancel(msg, hdr);
524 static int vdpa_nl_cmd_mgmtdev_get_doit(struct sk_buff *skb, struct genl_info *info)
526 struct vdpa_mgmt_dev *mdev;
530 msg = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL);
534 down_read(&vdpa_dev_lock);
535 mdev = vdpa_mgmtdev_get_from_attr(info->attrs);
537 up_read(&vdpa_dev_lock);
538 NL_SET_ERR_MSG_MOD(info->extack, "Fail to find the specified mgmt device");
543 err = vdpa_mgmtdev_fill(mdev, msg, info->snd_portid, info->snd_seq, 0);
544 up_read(&vdpa_dev_lock);
547 err = genlmsg_reply(msg, info);
556 vdpa_nl_cmd_mgmtdev_get_dumpit(struct sk_buff *msg, struct netlink_callback *cb)
558 struct vdpa_mgmt_dev *mdev;
559 int start = cb->args[0];
563 down_read(&vdpa_dev_lock);
564 list_for_each_entry(mdev, &mdev_head, list) {
569 err = vdpa_mgmtdev_fill(mdev, msg, NETLINK_CB(cb->skb).portid,
570 cb->nlh->nlmsg_seq, NLM_F_MULTI);
576 up_read(&vdpa_dev_lock);
581 #define VDPA_DEV_NET_ATTRS_MASK (BIT_ULL(VDPA_ATTR_DEV_NET_CFG_MACADDR) | \
582 BIT_ULL(VDPA_ATTR_DEV_NET_CFG_MTU) | \
583 BIT_ULL(VDPA_ATTR_DEV_NET_CFG_MAX_VQP))
586 * Bitmask for all per-device features: feature bits VIRTIO_TRANSPORT_F_START
587 * through VIRTIO_TRANSPORT_F_END are unset, i.e. 0xfffffc000fffffff for
588 * all 64bit features. If the features are extended beyond 64 bits, or new
589 * "holes" are reserved for other type of features than per-device, this
590 * macro would have to be updated.
592 #define VIRTIO_DEVICE_F_MASK (~0ULL << (VIRTIO_TRANSPORT_F_END + 1) | \
593 ((1ULL << VIRTIO_TRANSPORT_F_START) - 1))
595 static int vdpa_nl_cmd_dev_add_set_doit(struct sk_buff *skb, struct genl_info *info)
597 struct vdpa_dev_set_config config = {};
598 struct nlattr **nl_attrs = info->attrs;
599 struct vdpa_mgmt_dev *mdev;
600 unsigned int ncls = 0;
606 if (!info->attrs[VDPA_ATTR_DEV_NAME])
609 name = nla_data(info->attrs[VDPA_ATTR_DEV_NAME]);
611 if (nl_attrs[VDPA_ATTR_DEV_NET_CFG_MACADDR]) {
612 macaddr = nla_data(nl_attrs[VDPA_ATTR_DEV_NET_CFG_MACADDR]);
613 memcpy(config.net.mac, macaddr, sizeof(config.net.mac));
614 config.mask |= BIT_ULL(VDPA_ATTR_DEV_NET_CFG_MACADDR);
616 if (nl_attrs[VDPA_ATTR_DEV_NET_CFG_MTU]) {
618 nla_get_u16(nl_attrs[VDPA_ATTR_DEV_NET_CFG_MTU]);
619 config.mask |= BIT_ULL(VDPA_ATTR_DEV_NET_CFG_MTU);
621 if (nl_attrs[VDPA_ATTR_DEV_NET_CFG_MAX_VQP]) {
622 config.net.max_vq_pairs =
623 nla_get_u16(nl_attrs[VDPA_ATTR_DEV_NET_CFG_MAX_VQP]);
624 if (!config.net.max_vq_pairs) {
625 NL_SET_ERR_MSG_MOD(info->extack,
626 "At least one pair of VQs is required");
629 config.mask |= BIT_ULL(VDPA_ATTR_DEV_NET_CFG_MAX_VQP);
631 if (nl_attrs[VDPA_ATTR_DEV_FEATURES]) {
632 u64 missing = 0x0ULL;
634 config.device_features =
635 nla_get_u64(nl_attrs[VDPA_ATTR_DEV_FEATURES]);
636 if (nl_attrs[VDPA_ATTR_DEV_NET_CFG_MACADDR] &&
637 !(config.device_features & BIT_ULL(VIRTIO_NET_F_MAC)))
638 missing |= BIT_ULL(VIRTIO_NET_F_MAC);
639 if (nl_attrs[VDPA_ATTR_DEV_NET_CFG_MTU] &&
640 !(config.device_features & BIT_ULL(VIRTIO_NET_F_MTU)))
641 missing |= BIT_ULL(VIRTIO_NET_F_MTU);
642 if (nl_attrs[VDPA_ATTR_DEV_NET_CFG_MAX_VQP] &&
643 config.net.max_vq_pairs > 1 &&
644 !(config.device_features & BIT_ULL(VIRTIO_NET_F_MQ)))
645 missing |= BIT_ULL(VIRTIO_NET_F_MQ);
647 NL_SET_ERR_MSG_FMT_MOD(info->extack,
648 "Missing features 0x%llx for provided attributes",
652 config.mask |= BIT_ULL(VDPA_ATTR_DEV_FEATURES);
655 /* Skip checking capability if user didn't prefer to configure any
656 * device networking attributes. It is likely that user might have used
657 * a device specific method to configure such attributes or using device
658 * default attributes.
660 if ((config.mask & VDPA_DEV_NET_ATTRS_MASK) &&
661 !netlink_capable(skb, CAP_NET_ADMIN))
664 down_write(&vdpa_dev_lock);
665 mdev = vdpa_mgmtdev_get_from_attr(info->attrs);
667 NL_SET_ERR_MSG_MOD(info->extack, "Fail to find the specified management device");
672 if ((config.mask & mdev->config_attr_mask) != config.mask) {
673 NL_SET_ERR_MSG_FMT_MOD(info->extack,
674 "Some provided attributes are not supported: 0x%llx",
675 config.mask & ~mdev->config_attr_mask);
680 classes = vdpa_mgmtdev_get_classes(mdev, &ncls);
681 if (config.mask & VDPA_DEV_NET_ATTRS_MASK &&
682 !(classes & BIT_ULL(VIRTIO_ID_NET))) {
683 NL_SET_ERR_MSG_MOD(info->extack,
684 "Network class attributes provided on unsupported management device");
688 if (!(config.mask & VDPA_DEV_NET_ATTRS_MASK) &&
689 config.mask & BIT_ULL(VDPA_ATTR_DEV_FEATURES) &&
690 classes & BIT_ULL(VIRTIO_ID_NET) && ncls > 1 &&
691 config.device_features & VIRTIO_DEVICE_F_MASK) {
692 NL_SET_ERR_MSG_MOD(info->extack,
693 "Management device supports multi-class while device features specified are ambiguous");
698 err = mdev->ops->dev_add(mdev, name, &config);
700 up_write(&vdpa_dev_lock);
704 static int vdpa_nl_cmd_dev_del_set_doit(struct sk_buff *skb, struct genl_info *info)
706 struct vdpa_mgmt_dev *mdev;
707 struct vdpa_device *vdev;
712 if (!info->attrs[VDPA_ATTR_DEV_NAME])
714 name = nla_data(info->attrs[VDPA_ATTR_DEV_NAME]);
716 down_write(&vdpa_dev_lock);
717 dev = bus_find_device(&vdpa_bus, NULL, name, vdpa_name_match);
719 NL_SET_ERR_MSG_MOD(info->extack, "device not found");
723 vdev = container_of(dev, struct vdpa_device, dev);
725 NL_SET_ERR_MSG_MOD(info->extack, "Only user created device can be deleted by user");
730 mdev->ops->dev_del(mdev, vdev);
734 up_write(&vdpa_dev_lock);
739 vdpa_dev_fill(struct vdpa_device *vdev, struct sk_buff *msg, u32 portid, u32 seq,
740 int flags, struct netlink_ext_ack *extack)
749 hdr = genlmsg_put(msg, portid, seq, &vdpa_nl_family, flags, VDPA_CMD_DEV_NEW);
753 err = vdpa_nl_mgmtdev_handle_fill(msg, vdev->mdev);
757 device_id = vdev->config->get_device_id(vdev);
758 vendor_id = vdev->config->get_vendor_id(vdev);
759 max_vq_size = vdev->config->get_vq_num_max(vdev);
760 if (vdev->config->get_vq_num_min)
761 min_vq_size = vdev->config->get_vq_num_min(vdev);
764 if (nla_put_string(msg, VDPA_ATTR_DEV_NAME, dev_name(&vdev->dev)))
766 if (nla_put_u32(msg, VDPA_ATTR_DEV_ID, device_id))
768 if (nla_put_u32(msg, VDPA_ATTR_DEV_VENDOR_ID, vendor_id))
770 if (nla_put_u32(msg, VDPA_ATTR_DEV_MAX_VQS, vdev->nvqs))
772 if (nla_put_u16(msg, VDPA_ATTR_DEV_MAX_VQ_SIZE, max_vq_size))
774 if (nla_put_u16(msg, VDPA_ATTR_DEV_MIN_VQ_SIZE, min_vq_size))
777 genlmsg_end(msg, hdr);
781 genlmsg_cancel(msg, hdr);
785 static int vdpa_nl_cmd_dev_get_doit(struct sk_buff *skb, struct genl_info *info)
787 struct vdpa_device *vdev;
793 if (!info->attrs[VDPA_ATTR_DEV_NAME])
795 devname = nla_data(info->attrs[VDPA_ATTR_DEV_NAME]);
796 msg = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL);
800 down_read(&vdpa_dev_lock);
801 dev = bus_find_device(&vdpa_bus, NULL, devname, vdpa_name_match);
803 NL_SET_ERR_MSG_MOD(info->extack, "device not found");
807 vdev = container_of(dev, struct vdpa_device, dev);
812 err = vdpa_dev_fill(vdev, msg, info->snd_portid, info->snd_seq, 0, info->extack);
816 err = genlmsg_reply(msg, info);
818 up_read(&vdpa_dev_lock);
824 up_read(&vdpa_dev_lock);
829 struct vdpa_dev_dump_info {
831 struct netlink_callback *cb;
836 static int vdpa_dev_dump(struct device *dev, void *data)
838 struct vdpa_device *vdev = container_of(dev, struct vdpa_device, dev);
839 struct vdpa_dev_dump_info *info = data;
844 if (info->idx < info->start_idx) {
848 err = vdpa_dev_fill(vdev, info->msg, NETLINK_CB(info->cb->skb).portid,
849 info->cb->nlh->nlmsg_seq, NLM_F_MULTI, info->cb->extack);
857 static int vdpa_nl_cmd_dev_get_dumpit(struct sk_buff *msg, struct netlink_callback *cb)
859 struct vdpa_dev_dump_info info;
863 info.start_idx = cb->args[0];
866 down_read(&vdpa_dev_lock);
867 bus_for_each_dev(&vdpa_bus, NULL, &info, vdpa_dev_dump);
868 up_read(&vdpa_dev_lock);
869 cb->args[0] = info.idx;
873 static int vdpa_dev_net_mq_config_fill(struct sk_buff *msg, u64 features,
874 const struct virtio_net_config *config)
878 if ((features & BIT_ULL(VIRTIO_NET_F_MQ)) == 0 &&
879 (features & BIT_ULL(VIRTIO_NET_F_RSS)) == 0)
882 val_u16 = __virtio16_to_cpu(true, config->max_virtqueue_pairs);
884 return nla_put_u16(msg, VDPA_ATTR_DEV_NET_CFG_MAX_VQP, val_u16);
887 static int vdpa_dev_net_mtu_config_fill(struct sk_buff *msg, u64 features,
888 const struct virtio_net_config *config)
892 if ((features & BIT_ULL(VIRTIO_NET_F_MTU)) == 0)
895 val_u16 = __virtio16_to_cpu(true, config->mtu);
897 return nla_put_u16(msg, VDPA_ATTR_DEV_NET_CFG_MTU, val_u16);
900 static int vdpa_dev_net_mac_config_fill(struct sk_buff *msg, u64 features,
901 const struct virtio_net_config *config)
903 if ((features & BIT_ULL(VIRTIO_NET_F_MAC)) == 0)
906 return nla_put(msg, VDPA_ATTR_DEV_NET_CFG_MACADDR,
907 sizeof(config->mac), config->mac);
910 static int vdpa_dev_net_status_config_fill(struct sk_buff *msg, u64 features,
911 const struct virtio_net_config *config)
915 if ((features & BIT_ULL(VIRTIO_NET_F_STATUS)) == 0)
918 val_u16 = __virtio16_to_cpu(true, config->status);
919 return nla_put_u16(msg, VDPA_ATTR_DEV_NET_STATUS, val_u16);
922 static int vdpa_dev_net_config_fill(struct vdpa_device *vdev, struct sk_buff *msg)
924 struct virtio_net_config config = {};
927 vdev->config->get_config(vdev, 0, &config, sizeof(config));
929 features_device = vdev->config->get_device_features(vdev);
931 if (nla_put_u64_64bit(msg, VDPA_ATTR_DEV_FEATURES, features_device,
935 if (vdpa_dev_net_mtu_config_fill(msg, features_device, &config))
938 if (vdpa_dev_net_mac_config_fill(msg, features_device, &config))
941 if (vdpa_dev_net_status_config_fill(msg, features_device, &config))
944 return vdpa_dev_net_mq_config_fill(msg, features_device, &config);
948 vdpa_dev_config_fill(struct vdpa_device *vdev, struct sk_buff *msg, u32 portid, u32 seq,
949 int flags, struct netlink_ext_ack *extack)
957 down_read(&vdev->cf_lock);
958 hdr = genlmsg_put(msg, portid, seq, &vdpa_nl_family, flags,
959 VDPA_CMD_DEV_CONFIG_GET);
965 if (nla_put_string(msg, VDPA_ATTR_DEV_NAME, dev_name(&vdev->dev))) {
970 device_id = vdev->config->get_device_id(vdev);
971 if (nla_put_u32(msg, VDPA_ATTR_DEV_ID, device_id)) {
976 /* only read driver features after the feature negotiation is done */
977 status = vdev->config->get_status(vdev);
978 if (status & VIRTIO_CONFIG_S_FEATURES_OK) {
979 features_driver = vdev->config->get_driver_features(vdev);
980 if (nla_put_u64_64bit(msg, VDPA_ATTR_DEV_NEGOTIATED_FEATURES, features_driver,
989 err = vdpa_dev_net_config_fill(vdev, msg);
998 up_read(&vdev->cf_lock);
999 genlmsg_end(msg, hdr);
1003 genlmsg_cancel(msg, hdr);
1005 up_read(&vdev->cf_lock);
1009 static int vdpa_fill_stats_rec(struct vdpa_device *vdev, struct sk_buff *msg,
1010 struct genl_info *info, u32 index)
1012 struct virtio_net_config config = {};
1017 status = vdev->config->get_status(vdev);
1018 if (!(status & VIRTIO_CONFIG_S_FEATURES_OK)) {
1019 NL_SET_ERR_MSG_MOD(info->extack, "feature negotiation not complete");
1022 vdpa_get_config_unlocked(vdev, 0, &config, sizeof(config));
1024 features = vdev->config->get_driver_features(vdev);
1025 if (nla_put_u64_64bit(msg, VDPA_ATTR_DEV_NEGOTIATED_FEATURES,
1026 features, VDPA_ATTR_PAD))
1029 err = vdpa_dev_net_mq_config_fill(msg, features, &config);
1033 if (nla_put_u32(msg, VDPA_ATTR_DEV_QUEUE_INDEX, index))
1036 err = vdev->config->get_vendor_vq_stats(vdev, index, msg, info->extack);
1043 static int vendor_stats_fill(struct vdpa_device *vdev, struct sk_buff *msg,
1044 struct genl_info *info, u32 index)
1048 down_read(&vdev->cf_lock);
1049 if (!vdev->config->get_vendor_vq_stats) {
1054 err = vdpa_fill_stats_rec(vdev, msg, info, index);
1056 up_read(&vdev->cf_lock);
1060 static int vdpa_dev_vendor_stats_fill(struct vdpa_device *vdev,
1061 struct sk_buff *msg,
1062 struct genl_info *info, u32 index)
1067 u32 portid = info->snd_portid;
1068 u32 seq = info->snd_seq;
1071 hdr = genlmsg_put(msg, portid, seq, &vdpa_nl_family, flags,
1072 VDPA_CMD_DEV_VSTATS_GET);
1076 if (nla_put_string(msg, VDPA_ATTR_DEV_NAME, dev_name(&vdev->dev))) {
1081 device_id = vdev->config->get_device_id(vdev);
1082 if (nla_put_u32(msg, VDPA_ATTR_DEV_ID, device_id)) {
1087 switch (device_id) {
1089 if (index > VIRTIO_NET_CTRL_MQ_VQ_PAIRS_MAX) {
1090 NL_SET_ERR_MSG_MOD(info->extack, "queue index exceeds max value");
1095 err = vendor_stats_fill(vdev, msg, info, index);
1101 genlmsg_end(msg, hdr);
1106 genlmsg_cancel(msg, hdr);
1110 static int vdpa_nl_cmd_dev_config_get_doit(struct sk_buff *skb, struct genl_info *info)
1112 struct vdpa_device *vdev;
1113 struct sk_buff *msg;
1114 const char *devname;
1118 if (!info->attrs[VDPA_ATTR_DEV_NAME])
1120 devname = nla_data(info->attrs[VDPA_ATTR_DEV_NAME]);
1121 msg = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL);
1125 down_read(&vdpa_dev_lock);
1126 dev = bus_find_device(&vdpa_bus, NULL, devname, vdpa_name_match);
1128 NL_SET_ERR_MSG_MOD(info->extack, "device not found");
1132 vdev = container_of(dev, struct vdpa_device, dev);
1134 NL_SET_ERR_MSG_MOD(info->extack, "unmanaged vdpa device");
1138 err = vdpa_dev_config_fill(vdev, msg, info->snd_portid, info->snd_seq,
1141 err = genlmsg_reply(msg, info);
1146 up_read(&vdpa_dev_lock);
1152 static int vdpa_dev_config_dump(struct device *dev, void *data)
1154 struct vdpa_device *vdev = container_of(dev, struct vdpa_device, dev);
1155 struct vdpa_dev_dump_info *info = data;
1160 if (info->idx < info->start_idx) {
1164 err = vdpa_dev_config_fill(vdev, info->msg, NETLINK_CB(info->cb->skb).portid,
1165 info->cb->nlh->nlmsg_seq, NLM_F_MULTI,
1175 vdpa_nl_cmd_dev_config_get_dumpit(struct sk_buff *msg, struct netlink_callback *cb)
1177 struct vdpa_dev_dump_info info;
1181 info.start_idx = cb->args[0];
1184 down_read(&vdpa_dev_lock);
1185 bus_for_each_dev(&vdpa_bus, NULL, &info, vdpa_dev_config_dump);
1186 up_read(&vdpa_dev_lock);
1187 cb->args[0] = info.idx;
1191 static int vdpa_nl_cmd_dev_stats_get_doit(struct sk_buff *skb,
1192 struct genl_info *info)
1194 struct vdpa_device *vdev;
1195 struct sk_buff *msg;
1196 const char *devname;
1201 if (!info->attrs[VDPA_ATTR_DEV_NAME])
1204 if (!info->attrs[VDPA_ATTR_DEV_QUEUE_INDEX])
1207 devname = nla_data(info->attrs[VDPA_ATTR_DEV_NAME]);
1208 msg = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL);
1212 index = nla_get_u32(info->attrs[VDPA_ATTR_DEV_QUEUE_INDEX]);
1213 down_read(&vdpa_dev_lock);
1214 dev = bus_find_device(&vdpa_bus, NULL, devname, vdpa_name_match);
1216 NL_SET_ERR_MSG_MOD(info->extack, "device not found");
1220 vdev = container_of(dev, struct vdpa_device, dev);
1222 NL_SET_ERR_MSG_MOD(info->extack, "unmanaged vdpa device");
1226 err = vdpa_dev_vendor_stats_fill(vdev, msg, info, index);
1230 err = genlmsg_reply(msg, info);
1233 up_read(&vdpa_dev_lock);
1241 up_read(&vdpa_dev_lock);
1245 static const struct nla_policy vdpa_nl_policy[VDPA_ATTR_MAX + 1] = {
1246 [VDPA_ATTR_MGMTDEV_BUS_NAME] = { .type = NLA_NUL_STRING },
1247 [VDPA_ATTR_MGMTDEV_DEV_NAME] = { .type = NLA_STRING },
1248 [VDPA_ATTR_DEV_NAME] = { .type = NLA_STRING },
1249 [VDPA_ATTR_DEV_NET_CFG_MACADDR] = NLA_POLICY_ETH_ADDR,
1250 [VDPA_ATTR_DEV_NET_CFG_MAX_VQP] = { .type = NLA_U16 },
1251 /* virtio spec 1.1 section 5.1.4.1 for valid MTU range */
1252 [VDPA_ATTR_DEV_NET_CFG_MTU] = NLA_POLICY_MIN(NLA_U16, 68),
1253 [VDPA_ATTR_DEV_QUEUE_INDEX] = { .type = NLA_U32 },
1254 [VDPA_ATTR_DEV_FEATURES] = { .type = NLA_U64 },
1257 static const struct genl_ops vdpa_nl_ops[] = {
1259 .cmd = VDPA_CMD_MGMTDEV_GET,
1260 .doit = vdpa_nl_cmd_mgmtdev_get_doit,
1261 .dumpit = vdpa_nl_cmd_mgmtdev_get_dumpit,
1264 .cmd = VDPA_CMD_DEV_NEW,
1265 .doit = vdpa_nl_cmd_dev_add_set_doit,
1266 .flags = GENL_ADMIN_PERM,
1269 .cmd = VDPA_CMD_DEV_DEL,
1270 .doit = vdpa_nl_cmd_dev_del_set_doit,
1271 .flags = GENL_ADMIN_PERM,
1274 .cmd = VDPA_CMD_DEV_GET,
1275 .doit = vdpa_nl_cmd_dev_get_doit,
1276 .dumpit = vdpa_nl_cmd_dev_get_dumpit,
1279 .cmd = VDPA_CMD_DEV_CONFIG_GET,
1280 .doit = vdpa_nl_cmd_dev_config_get_doit,
1281 .dumpit = vdpa_nl_cmd_dev_config_get_dumpit,
1284 .cmd = VDPA_CMD_DEV_VSTATS_GET,
1285 .doit = vdpa_nl_cmd_dev_stats_get_doit,
1286 .flags = GENL_ADMIN_PERM,
1290 static struct genl_family vdpa_nl_family __ro_after_init = {
1291 .name = VDPA_GENL_NAME,
1292 .version = VDPA_GENL_VERSION,
1293 .maxattr = VDPA_ATTR_MAX,
1294 .policy = vdpa_nl_policy,
1296 .module = THIS_MODULE,
1298 .n_ops = ARRAY_SIZE(vdpa_nl_ops),
1299 .resv_start_op = VDPA_CMD_DEV_VSTATS_GET + 1,
1302 static int vdpa_init(void)
1306 err = bus_register(&vdpa_bus);
1309 err = genl_register_family(&vdpa_nl_family);
1315 bus_unregister(&vdpa_bus);
1319 static void __exit vdpa_exit(void)
1321 genl_unregister_family(&vdpa_nl_family);
1322 bus_unregister(&vdpa_bus);
1323 ida_destroy(&vdpa_index_ida);
1325 core_initcall(vdpa_init);
1326 module_exit(vdpa_exit);
1328 MODULE_AUTHOR("Jason Wang <jasowang@redhat.com>");
1329 MODULE_LICENSE("GPL v2");