]> Git Repo - linux.git/blame - drivers/vdpa/vdpa.c
Linux 6.14-rc3
[linux.git] / drivers / vdpa / vdpa.c
CommitLineData
961e9c84
JW
1// SPDX-License-Identifier: GPL-2.0-only
2/*
3 * vDPA bus.
4 *
5 * Copyright (c) 2020, Red Hat. All rights reserved.
6 * Author: Jason Wang <[email protected]>
7 *
8 */
9
10#include <linux/module.h>
11#include <linux/idr.h>
12#include <linux/slab.h>
13#include <linux/vdpa.h>
33b34750
PP
14#include <uapi/linux/vdpa.h>
15#include <net/genetlink.h>
16#include <linux/mod_devicetable.h>
ad69dd0b 17#include <linux/virtio_ids.h>
961e9c84 18
33b34750 19static LIST_HEAD(mdev_head);
fd70a406 20/* A global mutex that protects vdpa management device and device level operations. */
0078ad90 21static DECLARE_RWSEM(vdpa_dev_lock);
961e9c84
JW
22static DEFINE_IDA(vdpa_index_ida);
23
73bc0dbb
EC
24void vdpa_set_status(struct vdpa_device *vdev, u8 status)
25{
a6a51adc 26 down_write(&vdev->cf_lock);
73bc0dbb 27 vdev->config->set_status(vdev, status);
a6a51adc 28 up_write(&vdev->cf_lock);
73bc0dbb
EC
29}
30EXPORT_SYMBOL(vdpa_set_status);
31
33b34750
PP
32static struct genl_family vdpa_nl_family;
33
961e9c84
JW
34static int vdpa_dev_probe(struct device *d)
35{
36 struct vdpa_device *vdev = dev_to_vdpa(d);
37 struct vdpa_driver *drv = drv_to_vdpa(vdev->dev.driver);
c53e5d1b
WZ
38 const struct vdpa_config_ops *ops = vdev->config;
39 u32 max_num, min_num = 1;
961e9c84
JW
40 int ret = 0;
41
99fb2b83
JW
42 d->dma_mask = &d->coherent_dma_mask;
43 ret = dma_set_mask_and_coherent(d, DMA_BIT_MASK(64));
44 if (ret)
45 return ret;
46
c53e5d1b
WZ
47 max_num = ops->get_vq_num_max(vdev);
48 if (ops->get_vq_num_min)
49 min_num = ops->get_vq_num_min(vdev);
50 if (max_num < min_num)
51 return -EINVAL;
52
961e9c84
JW
53 if (drv && drv->probe)
54 ret = drv->probe(vdev);
55
56 return ret;
57}
58
fc7a6209 59static void vdpa_dev_remove(struct device *d)
961e9c84
JW
60{
61 struct vdpa_device *vdev = dev_to_vdpa(d);
62 struct vdpa_driver *drv = drv_to_vdpa(vdev->dev.driver);
63
64 if (drv && drv->remove)
65 drv->remove(vdev);
961e9c84
JW
66}
67
d69d8048 68static int vdpa_dev_match(struct device *dev, const struct device_driver *drv)
539fec78
SG
69{
70 struct vdpa_device *vdev = dev_to_vdpa(dev);
71
72 /* Check override first, and if set, only use the named driver */
73 if (vdev->driver_override)
74 return strcmp(vdev->driver_override, drv->name) == 0;
75
76 /* Currently devices must be supported by all vDPA bus drivers */
77 return 1;
78}
79
80static ssize_t driver_override_store(struct device *dev,
81 struct device_attribute *attr,
82 const char *buf, size_t count)
83{
84 struct vdpa_device *vdev = dev_to_vdpa(dev);
240bf4e6 85 int ret;
539fec78 86
240bf4e6
KK
87 ret = driver_set_override(dev, &vdev->driver_override, buf, count);
88 if (ret)
89 return ret;
539fec78
SG
90
91 return count;
92}
93
94static ssize_t driver_override_show(struct device *dev,
95 struct device_attribute *attr, char *buf)
96{
97 struct vdpa_device *vdev = dev_to_vdpa(dev);
98 ssize_t len;
99
100 device_lock(dev);
7b1b5c7f 101 len = sysfs_emit(buf, "%s\n", vdev->driver_override);
539fec78
SG
102 device_unlock(dev);
103
104 return len;
105}
106static DEVICE_ATTR_RW(driver_override);
107
108static struct attribute *vdpa_dev_attrs[] = {
109 &dev_attr_driver_override.attr,
110 NULL,
111};
112
113static const struct attribute_group vdpa_dev_group = {
114 .attrs = vdpa_dev_attrs,
115};
116__ATTRIBUTE_GROUPS(vdpa_dev);
117
8169ed62 118static const struct bus_type vdpa_bus = {
961e9c84 119 .name = "vdpa",
539fec78
SG
120 .dev_groups = vdpa_dev_groups,
121 .match = vdpa_dev_match,
961e9c84
JW
122 .probe = vdpa_dev_probe,
123 .remove = vdpa_dev_remove,
124};
125
126static void vdpa_release_dev(struct device *d)
127{
128 struct vdpa_device *vdev = dev_to_vdpa(d);
129 const struct vdpa_config_ops *ops = vdev->config;
130
131 if (ops->free)
132 ops->free(vdev);
133
c271fcd9 134 ida_free(&vdpa_index_ida, vdev->index);
539fec78 135 kfree(vdev->driver_override);
961e9c84
JW
136 kfree(vdev);
137}
138
139/**
140 * __vdpa_alloc_device - allocate and initilaize a vDPA device
141 * This allows driver to some prepartion after device is
142 * initialized but before registered.
143 * @parent: the parent device
144 * @config: the bus operations that is supported by this device
d4821902 145 * @ngroups: number of groups supported by this device
db9adcbf 146 * @nas: number of address spaces supported by this device
961e9c84 147 * @size: size of the parent structure that contains private data
fd70a406 148 * @name: name of the vdpa device; optional.
d8945ec4 149 * @use_va: indicate whether virtual address must be used by this device
961e9c84 150 *
24eae8eb 151 * Driver should use vdpa_alloc_device() wrapper macro instead of
961e9c84
JW
152 * using this directly.
153 *
c0a54b4b
PP
154 * Return: Returns an error when parent/config/dma_dev is not set or fail to get
155 * ida.
961e9c84
JW
156 */
157struct vdpa_device *__vdpa_alloc_device(struct device *parent,
158 const struct vdpa_config_ops *config,
db9adcbf 159 unsigned int ngroups, unsigned int nas,
d8945ec4
XY
160 size_t size, const char *name,
161 bool use_va)
961e9c84
JW
162{
163 struct vdpa_device *vdev;
164 int err = -EINVAL;
165
166 if (!config)
167 goto err;
168
169 if (!!config->dma_map != !!config->dma_unmap)
170 goto err;
171
d8945ec4
XY
172 /* It should only work for the device that use on-chip IOMMU */
173 if (use_va && !(config->dma_map || config->set_map))
174 goto err;
175
961e9c84
JW
176 err = -ENOMEM;
177 vdev = kzalloc(size, GFP_KERNEL);
178 if (!vdev)
179 goto err;
180
418eddef 181 err = ida_alloc(&vdpa_index_ida, GFP_KERNEL);
961e9c84
JW
182 if (err < 0)
183 goto err_ida;
184
185 vdev->dev.bus = &vdpa_bus;
186 vdev->dev.parent = parent;
187 vdev->dev.release = vdpa_release_dev;
188 vdev->index = err;
189 vdev->config = config;
452639a6 190 vdev->features_valid = false;
d8945ec4 191 vdev->use_va = use_va;
d4821902 192 vdev->ngroups = ngroups;
db9adcbf 193 vdev->nas = nas;
961e9c84 194
fd70a406
PP
195 if (name)
196 err = dev_set_name(&vdev->dev, "%s", name);
197 else
198 err = dev_set_name(&vdev->dev, "vdpa%u", vdev->index);
961e9c84
JW
199 if (err)
200 goto err_name;
201
a6a51adc 202 init_rwsem(&vdev->cf_lock);
961e9c84
JW
203 device_initialize(&vdev->dev);
204
205 return vdev;
206
207err_name:
c271fcd9 208 ida_free(&vdpa_index_ida, vdev->index);
961e9c84
JW
209err_ida:
210 kfree(vdev);
211err:
212 return ERR_PTR(err);
213}
214EXPORT_SYMBOL_GPL(__vdpa_alloc_device);
215
fd70a406
PP
216static int vdpa_name_match(struct device *dev, const void *data)
217{
218 struct vdpa_device *vdev = container_of(dev, struct vdpa_device, dev);
219
220 return (strcmp(dev_name(&vdev->dev), data) == 0);
221}
222
81d46d69 223static int __vdpa_register_device(struct vdpa_device *vdev, u32 nvqs)
903f7bca
PP
224{
225 struct device *dev;
226
f00bdce0
JW
227 vdev->nvqs = nvqs;
228
0078ad90 229 lockdep_assert_held(&vdpa_dev_lock);
903f7bca
PP
230 dev = bus_find_device(&vdpa_bus, NULL, dev_name(&vdev->dev), vdpa_name_match);
231 if (dev) {
232 put_device(dev);
233 return -EEXIST;
234 }
235 return device_add(&vdev->dev);
236}
237
238/**
239 * _vdpa_register_device - register a vDPA device with vdpa lock held
240 * Caller must have a succeed call of vdpa_alloc_device() before.
241 * Caller must invoke this routine in the management device dev_add()
242 * callback after setting up valid mgmtdev for this vdpa device.
243 * @vdev: the vdpa device to be registered to vDPA bus
f00bdce0 244 * @nvqs: number of virtqueues supported by this device
903f7bca 245 *
c0a54b4b 246 * Return: Returns an error when fail to add device to vDPA bus
903f7bca 247 */
81d46d69 248int _vdpa_register_device(struct vdpa_device *vdev, u32 nvqs)
903f7bca
PP
249{
250 if (!vdev->mdev)
251 return -EINVAL;
252
f00bdce0 253 return __vdpa_register_device(vdev, nvqs);
903f7bca
PP
254}
255EXPORT_SYMBOL_GPL(_vdpa_register_device);
256
961e9c84
JW
257/**
258 * vdpa_register_device - register a vDPA device
ac8b85f9 259 * Callers must have a succeed call of vdpa_alloc_device() before.
961e9c84 260 * @vdev: the vdpa device to be registered to vDPA bus
f00bdce0 261 * @nvqs: number of virtqueues supported by this device
961e9c84 262 *
c0a54b4b 263 * Return: Returns an error when fail to add to vDPA bus
961e9c84 264 */
81d46d69 265int vdpa_register_device(struct vdpa_device *vdev, u32 nvqs)
961e9c84 266{
fd70a406
PP
267 int err;
268
0078ad90 269 down_write(&vdpa_dev_lock);
f00bdce0 270 err = __vdpa_register_device(vdev, nvqs);
0078ad90 271 up_write(&vdpa_dev_lock);
fd70a406 272 return err;
961e9c84
JW
273}
274EXPORT_SYMBOL_GPL(vdpa_register_device);
275
903f7bca
PP
276/**
277 * _vdpa_unregister_device - unregister a vDPA device
278 * Caller must invoke this routine as part of management device dev_del()
279 * callback.
280 * @vdev: the vdpa device to be unregisted from vDPA bus
281 */
282void _vdpa_unregister_device(struct vdpa_device *vdev)
283{
0078ad90 284 lockdep_assert_held(&vdpa_dev_lock);
903f7bca
PP
285 WARN_ON(!vdev->mdev);
286 device_unregister(&vdev->dev);
287}
288EXPORT_SYMBOL_GPL(_vdpa_unregister_device);
289
961e9c84
JW
290/**
291 * vdpa_unregister_device - unregister a vDPA device
292 * @vdev: the vdpa device to be unregisted from vDPA bus
293 */
294void vdpa_unregister_device(struct vdpa_device *vdev)
295{
0078ad90 296 down_write(&vdpa_dev_lock);
961e9c84 297 device_unregister(&vdev->dev);
0078ad90 298 up_write(&vdpa_dev_lock);
961e9c84
JW
299}
300EXPORT_SYMBOL_GPL(vdpa_unregister_device);
301
302/**
303 * __vdpa_register_driver - register a vDPA device driver
304 * @drv: the vdpa device driver to be registered
305 * @owner: module owner of the driver
306 *
c0a54b4b 307 * Return: Returns an err when fail to do the registration
961e9c84
JW
308 */
309int __vdpa_register_driver(struct vdpa_driver *drv, struct module *owner)
310{
311 drv->driver.bus = &vdpa_bus;
312 drv->driver.owner = owner;
313
314 return driver_register(&drv->driver);
315}
316EXPORT_SYMBOL_GPL(__vdpa_register_driver);
317
318/**
319 * vdpa_unregister_driver - unregister a vDPA device driver
320 * @drv: the vdpa device driver to be unregistered
321 */
322void vdpa_unregister_driver(struct vdpa_driver *drv)
323{
324 driver_unregister(&drv->driver);
325}
326EXPORT_SYMBOL_GPL(vdpa_unregister_driver);
327
33b34750
PP
328/**
329 * vdpa_mgmtdev_register - register a vdpa management device
330 *
331 * @mdev: Pointer to vdpa management device
332 * vdpa_mgmtdev_register() register a vdpa management device which supports
333 * vdpa device management.
c0a54b4b
PP
334 * Return: Returns 0 on success or failure when required callback ops are not
335 * initialized.
33b34750
PP
336 */
337int vdpa_mgmtdev_register(struct vdpa_mgmt_dev *mdev)
338{
339 if (!mdev->device || !mdev->ops || !mdev->ops->dev_add || !mdev->ops->dev_del)
340 return -EINVAL;
341
342 INIT_LIST_HEAD(&mdev->list);
0078ad90 343 down_write(&vdpa_dev_lock);
33b34750 344 list_add_tail(&mdev->list, &mdev_head);
0078ad90 345 up_write(&vdpa_dev_lock);
33b34750
PP
346 return 0;
347}
348EXPORT_SYMBOL_GPL(vdpa_mgmtdev_register);
349
903f7bca
PP
350static int vdpa_match_remove(struct device *dev, void *data)
351{
352 struct vdpa_device *vdev = container_of(dev, struct vdpa_device, dev);
353 struct vdpa_mgmt_dev *mdev = vdev->mdev;
354
355 if (mdev == data)
356 mdev->ops->dev_del(mdev, vdev);
357 return 0;
358}
359
33b34750
PP
360void vdpa_mgmtdev_unregister(struct vdpa_mgmt_dev *mdev)
361{
0078ad90 362 down_write(&vdpa_dev_lock);
903f7bca 363
33b34750 364 list_del(&mdev->list);
903f7bca
PP
365
366 /* Filter out all the entries belong to this management device and delete it. */
367 bus_for_each_dev(&vdpa_bus, NULL, mdev, vdpa_match_remove);
368
0078ad90 369 up_write(&vdpa_dev_lock);
33b34750
PP
370}
371EXPORT_SYMBOL_GPL(vdpa_mgmtdev_unregister);
372
30ef7a8a
EC
373static void vdpa_get_config_unlocked(struct vdpa_device *vdev,
374 unsigned int offset,
375 void *buf, unsigned int len)
376{
377 const struct vdpa_config_ops *ops = vdev->config;
378
379 /*
380 * Config accesses aren't supposed to trigger before features are set.
381 * If it does happen we assume a legacy guest.
382 */
383 if (!vdev->features_valid)
e0077cc1 384 vdpa_set_features_unlocked(vdev, 0);
30ef7a8a
EC
385 ops->get_config(vdev, offset, buf, len);
386}
387
6dbb1f16
PP
388/**
389 * vdpa_get_config - Get one or more device configuration fields.
390 * @vdev: vdpa device to operate on
391 * @offset: starting byte offset of the field
392 * @buf: buffer pointer to read to
393 * @len: length of the configuration fields in bytes
394 */
395void vdpa_get_config(struct vdpa_device *vdev, unsigned int offset,
396 void *buf, unsigned int len)
397{
a6a51adc 398 down_read(&vdev->cf_lock);
30ef7a8a 399 vdpa_get_config_unlocked(vdev, offset, buf, len);
a6a51adc 400 up_read(&vdev->cf_lock);
6dbb1f16
PP
401}
402EXPORT_SYMBOL_GPL(vdpa_get_config);
403
404/**
405 * vdpa_set_config - Set one or more device configuration fields.
406 * @vdev: vdpa device to operate on
407 * @offset: starting byte offset of the field
408 * @buf: buffer pointer to read from
409 * @length: length of the configuration fields in bytes
410 */
411void vdpa_set_config(struct vdpa_device *vdev, unsigned int offset,
412 const void *buf, unsigned int length)
413{
a6a51adc 414 down_write(&vdev->cf_lock);
6dbb1f16 415 vdev->config->set_config(vdev, offset, buf, length);
a6a51adc 416 up_write(&vdev->cf_lock);
6dbb1f16
PP
417}
418EXPORT_SYMBOL_GPL(vdpa_set_config);
419
33b34750
PP
420static bool mgmtdev_handle_match(const struct vdpa_mgmt_dev *mdev,
421 const char *busname, const char *devname)
422{
423 /* Bus name is optional for simulated management device, so ignore the
424 * device with bus if bus attribute is provided.
425 */
426 if ((busname && !mdev->device->bus) || (!busname && mdev->device->bus))
427 return false;
428
429 if (!busname && strcmp(dev_name(mdev->device), devname) == 0)
430 return true;
431
432 if (busname && (strcmp(mdev->device->bus->name, busname) == 0) &&
433 (strcmp(dev_name(mdev->device), devname) == 0))
434 return true;
435
436 return false;
437}
438
439static struct vdpa_mgmt_dev *vdpa_mgmtdev_get_from_attr(struct nlattr **attrs)
440{
441 struct vdpa_mgmt_dev *mdev;
442 const char *busname = NULL;
443 const char *devname;
444
445 if (!attrs[VDPA_ATTR_MGMTDEV_DEV_NAME])
446 return ERR_PTR(-EINVAL);
447 devname = nla_data(attrs[VDPA_ATTR_MGMTDEV_DEV_NAME]);
448 if (attrs[VDPA_ATTR_MGMTDEV_BUS_NAME])
449 busname = nla_data(attrs[VDPA_ATTR_MGMTDEV_BUS_NAME]);
450
451 list_for_each_entry(mdev, &mdev_head, list) {
452 if (mgmtdev_handle_match(mdev, busname, devname))
453 return mdev;
454 }
455 return ERR_PTR(-ENODEV);
456}
457
458static int vdpa_nl_mgmtdev_handle_fill(struct sk_buff *msg, const struct vdpa_mgmt_dev *mdev)
459{
460 if (mdev->device->bus &&
461 nla_put_string(msg, VDPA_ATTR_MGMTDEV_BUS_NAME, mdev->device->bus->name))
462 return -EMSGSIZE;
463 if (nla_put_string(msg, VDPA_ATTR_MGMTDEV_DEV_NAME, dev_name(mdev->device)))
464 return -EMSGSIZE;
465 return 0;
466}
467
dbb6f1c4
SWL
468static u64 vdpa_mgmtdev_get_classes(const struct vdpa_mgmt_dev *mdev,
469 unsigned int *nclasses)
470{
471 u64 supported_classes = 0;
472 unsigned int n = 0;
473
474 for (int i = 0; mdev->id_table[i].device; i++) {
475 if (mdev->id_table[i].device > 63)
476 continue;
477 supported_classes |= BIT_ULL(mdev->id_table[i].device);
478 n++;
479 }
480 if (nclasses)
481 *nclasses = n;
482
483 return supported_classes;
484}
485
33b34750
PP
486static int vdpa_mgmtdev_fill(const struct vdpa_mgmt_dev *mdev, struct sk_buff *msg,
487 u32 portid, u32 seq, int flags)
488{
33b34750 489 void *hdr;
33b34750
PP
490 int err;
491
492 hdr = genlmsg_put(msg, portid, seq, &vdpa_nl_family, flags, VDPA_CMD_MGMTDEV_NEW);
493 if (!hdr)
494 return -EMSGSIZE;
495 err = vdpa_nl_mgmtdev_handle_fill(msg, mdev);
496 if (err)
497 goto msg_err;
498
33b34750 499 if (nla_put_u64_64bit(msg, VDPA_ATTR_MGMTDEV_SUPPORTED_CLASSES,
dbb6f1c4
SWL
500 vdpa_mgmtdev_get_classes(mdev, NULL),
501 VDPA_ATTR_UNSPEC)) {
33b34750
PP
502 err = -EMSGSIZE;
503 goto msg_err;
504 }
cd2629f6
EC
505 if (nla_put_u32(msg, VDPA_ATTR_DEV_MGMTDEV_MAX_VQS,
506 mdev->max_supported_vqs)) {
507 err = -EMSGSIZE;
508 goto msg_err;
509 }
510 if (nla_put_u64_64bit(msg, VDPA_ATTR_DEV_SUPPORTED_FEATURES,
511 mdev->supported_features, VDPA_ATTR_PAD)) {
512 err = -EMSGSIZE;
513 goto msg_err;
514 }
33b34750
PP
515
516 genlmsg_end(msg, hdr);
517 return 0;
518
519msg_err:
520 genlmsg_cancel(msg, hdr);
521 return err;
522}
523
524static int vdpa_nl_cmd_mgmtdev_get_doit(struct sk_buff *skb, struct genl_info *info)
525{
526 struct vdpa_mgmt_dev *mdev;
527 struct sk_buff *msg;
528 int err;
529
530 msg = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL);
531 if (!msg)
532 return -ENOMEM;
533
0078ad90 534 down_read(&vdpa_dev_lock);
33b34750
PP
535 mdev = vdpa_mgmtdev_get_from_attr(info->attrs);
536 if (IS_ERR(mdev)) {
0078ad90 537 up_read(&vdpa_dev_lock);
33b34750
PP
538 NL_SET_ERR_MSG_MOD(info->extack, "Fail to find the specified mgmt device");
539 err = PTR_ERR(mdev);
540 goto out;
541 }
542
543 err = vdpa_mgmtdev_fill(mdev, msg, info->snd_portid, info->snd_seq, 0);
0078ad90 544 up_read(&vdpa_dev_lock);
33b34750
PP
545 if (err)
546 goto out;
547 err = genlmsg_reply(msg, info);
548 return err;
549
550out:
551 nlmsg_free(msg);
552 return err;
553}
554
555static int
556vdpa_nl_cmd_mgmtdev_get_dumpit(struct sk_buff *msg, struct netlink_callback *cb)
557{
558 struct vdpa_mgmt_dev *mdev;
559 int start = cb->args[0];
560 int idx = 0;
561 int err;
562
0078ad90 563 down_read(&vdpa_dev_lock);
33b34750
PP
564 list_for_each_entry(mdev, &mdev_head, list) {
565 if (idx < start) {
566 idx++;
567 continue;
568 }
569 err = vdpa_mgmtdev_fill(mdev, msg, NETLINK_CB(cb->skb).portid,
570 cb->nlh->nlmsg_seq, NLM_F_MULTI);
571 if (err)
572 goto out;
573 idx++;
574 }
575out:
0078ad90 576 up_read(&vdpa_dev_lock);
33b34750
PP
577 cb->args[0] = idx;
578 return msg->len;
579}
580
47a1401a
EC
581#define VDPA_DEV_NET_ATTRS_MASK (BIT_ULL(VDPA_ATTR_DEV_NET_CFG_MACADDR) | \
582 BIT_ULL(VDPA_ATTR_DEV_NET_CFG_MTU) | \
583 BIT_ULL(VDPA_ATTR_DEV_NET_CFG_MAX_VQP))
d8ca2fa5 584
dbb6f1c4
SWL
585/*
586 * Bitmask for all per-device features: feature bits VIRTIO_TRANSPORT_F_START
587 * through VIRTIO_TRANSPORT_F_END are unset, i.e. 0xfffffc000fffffff for
588 * all 64bit features. If the features are extended beyond 64 bits, or new
589 * "holes" are reserved for other type of features than per-device, this
590 * macro would have to be updated.
591 */
592#define VIRTIO_DEVICE_F_MASK (~0ULL << (VIRTIO_TRANSPORT_F_END + 1) | \
593 ((1ULL << VIRTIO_TRANSPORT_F_START) - 1))
594
903f7bca
PP
595static int vdpa_nl_cmd_dev_add_set_doit(struct sk_buff *skb, struct genl_info *info)
596{
d8ca2fa5
PP
597 struct vdpa_dev_set_config config = {};
598 struct nlattr **nl_attrs = info->attrs;
903f7bca 599 struct vdpa_mgmt_dev *mdev;
dbb6f1c4 600 unsigned int ncls = 0;
d8ca2fa5 601 const u8 *macaddr;
903f7bca 602 const char *name;
dbb6f1c4 603 u64 classes;
903f7bca
PP
604 int err = 0;
605
606 if (!info->attrs[VDPA_ATTR_DEV_NAME])
607 return -EINVAL;
608
609 name = nla_data(info->attrs[VDPA_ATTR_DEV_NAME]);
610
d8ca2fa5
PP
611 if (nl_attrs[VDPA_ATTR_DEV_NET_CFG_MACADDR]) {
612 macaddr = nla_data(nl_attrs[VDPA_ATTR_DEV_NET_CFG_MACADDR]);
613 memcpy(config.net.mac, macaddr, sizeof(config.net.mac));
47a1401a 614 config.mask |= BIT_ULL(VDPA_ATTR_DEV_NET_CFG_MACADDR);
d8ca2fa5
PP
615 }
616 if (nl_attrs[VDPA_ATTR_DEV_NET_CFG_MTU]) {
617 config.net.mtu =
618 nla_get_u16(nl_attrs[VDPA_ATTR_DEV_NET_CFG_MTU]);
47a1401a 619 config.mask |= BIT_ULL(VDPA_ATTR_DEV_NET_CFG_MTU);
d8ca2fa5 620 }
aba21aff
EC
621 if (nl_attrs[VDPA_ATTR_DEV_NET_CFG_MAX_VQP]) {
622 config.net.max_vq_pairs =
623 nla_get_u16(nl_attrs[VDPA_ATTR_DEV_NET_CFG_MAX_VQP]);
624 if (!config.net.max_vq_pairs) {
625 NL_SET_ERR_MSG_MOD(info->extack,
626 "At least one pair of VQs is required");
627 return -EINVAL;
628 }
629 config.mask |= BIT_ULL(VDPA_ATTR_DEV_NET_CFG_MAX_VQP);
630 }
90fea5a8 631 if (nl_attrs[VDPA_ATTR_DEV_FEATURES]) {
e7d09cd1
SWL
632 u64 missing = 0x0ULL;
633
90fea5a8
JW
634 config.device_features =
635 nla_get_u64(nl_attrs[VDPA_ATTR_DEV_FEATURES]);
e7d09cd1
SWL
636 if (nl_attrs[VDPA_ATTR_DEV_NET_CFG_MACADDR] &&
637 !(config.device_features & BIT_ULL(VIRTIO_NET_F_MAC)))
638 missing |= BIT_ULL(VIRTIO_NET_F_MAC);
639 if (nl_attrs[VDPA_ATTR_DEV_NET_CFG_MTU] &&
640 !(config.device_features & BIT_ULL(VIRTIO_NET_F_MTU)))
641 missing |= BIT_ULL(VIRTIO_NET_F_MTU);
642 if (nl_attrs[VDPA_ATTR_DEV_NET_CFG_MAX_VQP] &&
643 config.net.max_vq_pairs > 1 &&
644 !(config.device_features & BIT_ULL(VIRTIO_NET_F_MQ)))
645 missing |= BIT_ULL(VIRTIO_NET_F_MQ);
646 if (missing) {
647 NL_SET_ERR_MSG_FMT_MOD(info->extack,
648 "Missing features 0x%llx for provided attributes",
649 missing);
650 return -EINVAL;
651 }
90fea5a8
JW
652 config.mask |= BIT_ULL(VDPA_ATTR_DEV_FEATURES);
653 }
d8ca2fa5
PP
654
655 /* Skip checking capability if user didn't prefer to configure any
656 * device networking attributes. It is likely that user might have used
657 * a device specific method to configure such attributes or using device
658 * default attributes.
659 */
660 if ((config.mask & VDPA_DEV_NET_ATTRS_MASK) &&
661 !netlink_capable(skb, CAP_NET_ADMIN))
662 return -EPERM;
663
0078ad90 664 down_write(&vdpa_dev_lock);
903f7bca
PP
665 mdev = vdpa_mgmtdev_get_from_attr(info->attrs);
666 if (IS_ERR(mdev)) {
667 NL_SET_ERR_MSG_MOD(info->extack, "Fail to find the specified management device");
668 err = PTR_ERR(mdev);
669 goto err;
670 }
275487b4 671
d8ca2fa5 672 if ((config.mask & mdev->config_attr_mask) != config.mask) {
275487b4
SWL
673 NL_SET_ERR_MSG_FMT_MOD(info->extack,
674 "Some provided attributes are not supported: 0x%llx",
675 config.mask & ~mdev->config_attr_mask);
d8ca2fa5
PP
676 err = -EOPNOTSUPP;
677 goto err;
678 }
903f7bca 679
dbb6f1c4
SWL
680 classes = vdpa_mgmtdev_get_classes(mdev, &ncls);
681 if (config.mask & VDPA_DEV_NET_ATTRS_MASK &&
682 !(classes & BIT_ULL(VIRTIO_ID_NET))) {
683 NL_SET_ERR_MSG_MOD(info->extack,
684 "Network class attributes provided on unsupported management device");
685 err = -EINVAL;
686 goto err;
687 }
688 if (!(config.mask & VDPA_DEV_NET_ATTRS_MASK) &&
689 config.mask & BIT_ULL(VDPA_ATTR_DEV_FEATURES) &&
690 classes & BIT_ULL(VIRTIO_ID_NET) && ncls > 1 &&
691 config.device_features & VIRTIO_DEVICE_F_MASK) {
692 NL_SET_ERR_MSG_MOD(info->extack,
693 "Management device supports multi-class while device features specified are ambiguous");
694 err = -EINVAL;
695 goto err;
696 }
697
d8ca2fa5 698 err = mdev->ops->dev_add(mdev, name, &config);
903f7bca 699err:
0078ad90 700 up_write(&vdpa_dev_lock);
903f7bca
PP
701 return err;
702}
703
704static int vdpa_nl_cmd_dev_del_set_doit(struct sk_buff *skb, struct genl_info *info)
705{
706 struct vdpa_mgmt_dev *mdev;
707 struct vdpa_device *vdev;
708 struct device *dev;
709 const char *name;
710 int err = 0;
711
712 if (!info->attrs[VDPA_ATTR_DEV_NAME])
713 return -EINVAL;
714 name = nla_data(info->attrs[VDPA_ATTR_DEV_NAME]);
715
0078ad90 716 down_write(&vdpa_dev_lock);
903f7bca
PP
717 dev = bus_find_device(&vdpa_bus, NULL, name, vdpa_name_match);
718 if (!dev) {
719 NL_SET_ERR_MSG_MOD(info->extack, "device not found");
720 err = -ENODEV;
721 goto dev_err;
722 }
723 vdev = container_of(dev, struct vdpa_device, dev);
724 if (!vdev->mdev) {
725 NL_SET_ERR_MSG_MOD(info->extack, "Only user created device can be deleted by user");
726 err = -EINVAL;
727 goto mdev_err;
728 }
729 mdev = vdev->mdev;
730 mdev->ops->dev_del(mdev, vdev);
731mdev_err:
732 put_device(dev);
733dev_err:
0078ad90 734 up_write(&vdpa_dev_lock);
903f7bca
PP
735 return err;
736}
737
bc0d90ee
PP
738static int
739vdpa_dev_fill(struct vdpa_device *vdev, struct sk_buff *msg, u32 portid, u32 seq,
740 int flags, struct netlink_ext_ack *extack)
741{
742 u16 max_vq_size;
e47be840 743 u16 min_vq_size = 1;
bc0d90ee
PP
744 u32 device_id;
745 u32 vendor_id;
746 void *hdr;
747 int err;
748
749 hdr = genlmsg_put(msg, portid, seq, &vdpa_nl_family, flags, VDPA_CMD_DEV_NEW);
750 if (!hdr)
751 return -EMSGSIZE;
752
753 err = vdpa_nl_mgmtdev_handle_fill(msg, vdev->mdev);
754 if (err)
755 goto msg_err;
756
757 device_id = vdev->config->get_device_id(vdev);
758 vendor_id = vdev->config->get_vendor_id(vdev);
759 max_vq_size = vdev->config->get_vq_num_max(vdev);
e47be840
WZ
760 if (vdev->config->get_vq_num_min)
761 min_vq_size = vdev->config->get_vq_num_min(vdev);
bc0d90ee
PP
762
763 err = -EMSGSIZE;
764 if (nla_put_string(msg, VDPA_ATTR_DEV_NAME, dev_name(&vdev->dev)))
765 goto msg_err;
766 if (nla_put_u32(msg, VDPA_ATTR_DEV_ID, device_id))
767 goto msg_err;
768 if (nla_put_u32(msg, VDPA_ATTR_DEV_VENDOR_ID, vendor_id))
769 goto msg_err;
770 if (nla_put_u32(msg, VDPA_ATTR_DEV_MAX_VQS, vdev->nvqs))
771 goto msg_err;
772 if (nla_put_u16(msg, VDPA_ATTR_DEV_MAX_VQ_SIZE, max_vq_size))
773 goto msg_err;
e47be840
WZ
774 if (nla_put_u16(msg, VDPA_ATTR_DEV_MIN_VQ_SIZE, min_vq_size))
775 goto msg_err;
bc0d90ee
PP
776
777 genlmsg_end(msg, hdr);
778 return 0;
779
780msg_err:
781 genlmsg_cancel(msg, hdr);
782 return err;
783}
784
785static int vdpa_nl_cmd_dev_get_doit(struct sk_buff *skb, struct genl_info *info)
786{
787 struct vdpa_device *vdev;
788 struct sk_buff *msg;
789 const char *devname;
790 struct device *dev;
791 int err;
792
793 if (!info->attrs[VDPA_ATTR_DEV_NAME])
794 return -EINVAL;
795 devname = nla_data(info->attrs[VDPA_ATTR_DEV_NAME]);
796 msg = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL);
797 if (!msg)
798 return -ENOMEM;
799
0078ad90 800 down_read(&vdpa_dev_lock);
bc0d90ee
PP
801 dev = bus_find_device(&vdpa_bus, NULL, devname, vdpa_name_match);
802 if (!dev) {
803 NL_SET_ERR_MSG_MOD(info->extack, "device not found");
804 err = -ENODEV;
805 goto err;
806 }
807 vdev = container_of(dev, struct vdpa_device, dev);
808 if (!vdev->mdev) {
809 err = -EINVAL;
810 goto mdev_err;
811 }
812 err = vdpa_dev_fill(vdev, msg, info->snd_portid, info->snd_seq, 0, info->extack);
7a6691f1
EC
813 if (err)
814 goto mdev_err;
815
816 err = genlmsg_reply(msg, info);
817 put_device(dev);
0078ad90 818 up_read(&vdpa_dev_lock);
7a6691f1
EC
819 return err;
820
bc0d90ee
PP
821mdev_err:
822 put_device(dev);
823err:
0078ad90 824 up_read(&vdpa_dev_lock);
7a6691f1 825 nlmsg_free(msg);
bc0d90ee
PP
826 return err;
827}
828
829struct vdpa_dev_dump_info {
830 struct sk_buff *msg;
831 struct netlink_callback *cb;
832 int start_idx;
833 int idx;
834};
835
836static int vdpa_dev_dump(struct device *dev, void *data)
837{
838 struct vdpa_device *vdev = container_of(dev, struct vdpa_device, dev);
839 struct vdpa_dev_dump_info *info = data;
840 int err;
841
842 if (!vdev->mdev)
843 return 0;
844 if (info->idx < info->start_idx) {
845 info->idx++;
846 return 0;
847 }
848 err = vdpa_dev_fill(vdev, info->msg, NETLINK_CB(info->cb->skb).portid,
849 info->cb->nlh->nlmsg_seq, NLM_F_MULTI, info->cb->extack);
850 if (err)
851 return err;
852
853 info->idx++;
854 return 0;
855}
856
857static int vdpa_nl_cmd_dev_get_dumpit(struct sk_buff *msg, struct netlink_callback *cb)
858{
859 struct vdpa_dev_dump_info info;
860
861 info.msg = msg;
862 info.cb = cb;
863 info.start_idx = cb->args[0];
864 info.idx = 0;
865
0078ad90 866 down_read(&vdpa_dev_lock);
bc0d90ee 867 bus_for_each_dev(&vdpa_bus, NULL, &info, vdpa_dev_dump);
0078ad90 868 up_read(&vdpa_dev_lock);
bc0d90ee
PP
869 cb->args[0] = info.idx;
870 return msg->len;
871}
872
8a505711 873static int vdpa_dev_net_mq_config_fill(struct sk_buff *msg, u64 features,
ad69dd0b
PP
874 const struct virtio_net_config *config)
875{
876 u16 val_u16;
877
8a505711
ZL
878 if ((features & BIT_ULL(VIRTIO_NET_F_MQ)) == 0 &&
879 (features & BIT_ULL(VIRTIO_NET_F_RSS)) == 0)
ad69dd0b
PP
880 return 0;
881
35b37c33
ZL
882 val_u16 = __virtio16_to_cpu(true, config->max_virtqueue_pairs);
883
ad69dd0b
PP
884 return nla_put_u16(msg, VDPA_ATTR_DEV_NET_CFG_MAX_VQP, val_u16);
885}
886
41a2ad92
ZL
887static int vdpa_dev_net_mtu_config_fill(struct sk_buff *msg, u64 features,
888 const struct virtio_net_config *config)
889{
890 u16 val_u16;
891
892 if ((features & BIT_ULL(VIRTIO_NET_F_MTU)) == 0)
893 return 0;
894
895 val_u16 = __virtio16_to_cpu(true, config->mtu);
896
897 return nla_put_u16(msg, VDPA_ATTR_DEV_NET_CFG_MTU, val_u16);
898}
899
900static int vdpa_dev_net_mac_config_fill(struct sk_buff *msg, u64 features,
901 const struct virtio_net_config *config)
902{
903 if ((features & BIT_ULL(VIRTIO_NET_F_MAC)) == 0)
904 return 0;
905
906 return nla_put(msg, VDPA_ATTR_DEV_NET_CFG_MACADDR,
907 sizeof(config->mac), config->mac);
908}
909
6e6d3983
SWL
910static int vdpa_dev_net_status_config_fill(struct sk_buff *msg, u64 features,
911 const struct virtio_net_config *config)
912{
913 u16 val_u16;
914
915 if ((features & BIT_ULL(VIRTIO_NET_F_STATUS)) == 0)
916 return 0;
917
918 val_u16 = __virtio16_to_cpu(true, config->status);
919 return nla_put_u16(msg, VDPA_ATTR_DEV_NET_STATUS, val_u16);
920}
921
ad69dd0b
PP
922static int vdpa_dev_net_config_fill(struct vdpa_device *vdev, struct sk_buff *msg)
923{
924 struct virtio_net_config config = {};
c6dac2ec 925 u64 features_device;
ad69dd0b 926
22856510 927 vdev->config->get_config(vdev, 0, &config, sizeof(config));
ad69dd0b 928
22856510 929 features_device = vdev->config->get_device_features(vdev);
ad69dd0b 930
b9e05399 931 if (nla_put_u64_64bit(msg, VDPA_ATTR_DEV_FEATURES, features_device,
612f330e
EC
932 VDPA_ATTR_PAD))
933 return -EMSGSIZE;
ad69dd0b 934
41a2ad92
ZL
935 if (vdpa_dev_net_mtu_config_fill(msg, features_device, &config))
936 return -EMSGSIZE;
937
938 if (vdpa_dev_net_mac_config_fill(msg, features_device, &config))
939 return -EMSGSIZE;
940
6e6d3983
SWL
941 if (vdpa_dev_net_status_config_fill(msg, features_device, &config))
942 return -EMSGSIZE;
943
9d97aa12 944 return vdpa_dev_net_mq_config_fill(msg, features_device, &config);
ad69dd0b
PP
945}
946
c2475a9a
ZL
947static int
948vdpa_dev_blk_capacity_config_fill(struct sk_buff *msg,
949 const struct virtio_blk_config *config)
950{
951 u64 val_u64;
952
953 val_u64 = __virtio64_to_cpu(true, config->capacity);
954
955 return nla_put_u64_64bit(msg, VDPA_ATTR_DEV_BLK_CFG_CAPACITY,
956 val_u64, VDPA_ATTR_PAD);
957}
958
330b8aea
ZL
959static int
960vdpa_dev_blk_seg_size_config_fill(struct sk_buff *msg, u64 features,
961 const struct virtio_blk_config *config)
962{
963 u32 val_u32;
964
965 if ((features & BIT_ULL(VIRTIO_BLK_F_SIZE_MAX)) == 0)
966 return 0;
967
968 val_u32 = __virtio32_to_cpu(true, config->size_max);
969
98a82154 970 return nla_put_u32(msg, VDPA_ATTR_DEV_BLK_CFG_SIZE_MAX, val_u32);
330b8aea
ZL
971}
972
3a1d33fb
ZL
973/* fill the block size*/
974static int
975vdpa_dev_blk_block_size_config_fill(struct sk_buff *msg, u64 features,
976 const struct virtio_blk_config *config)
977{
978 u32 val_u32;
979
980 if ((features & BIT_ULL(VIRTIO_BLK_F_BLK_SIZE)) == 0)
981 return 0;
982
983 val_u32 = __virtio32_to_cpu(true, config->blk_size);
984
985 return nla_put_u32(msg, VDPA_ATTR_DEV_BLK_CFG_BLK_SIZE, val_u32);
986}
987
81f64e1d
ZL
988static int
989vdpa_dev_blk_seg_max_config_fill(struct sk_buff *msg, u64 features,
990 const struct virtio_blk_config *config)
991{
992 u32 val_u32;
993
994 if ((features & BIT_ULL(VIRTIO_BLK_F_SEG_MAX)) == 0)
995 return 0;
996
997 val_u32 = __virtio32_to_cpu(true, config->seg_max);
998
999 return nla_put_u32(msg, VDPA_ATTR_DEV_BLK_CFG_SEG_MAX, val_u32);
1000}
1001
54fb04b0
ZL
1002static int vdpa_dev_blk_mq_config_fill(struct sk_buff *msg, u64 features,
1003 const struct virtio_blk_config *config)
1004{
1005 u16 val_u16;
1006
1007 if ((features & BIT_ULL(VIRTIO_BLK_F_MQ)) == 0)
1008 return 0;
1009
1010 val_u16 = __virtio16_to_cpu(true, config->num_queues);
1011
1012 return nla_put_u16(msg, VDPA_ATTR_DEV_BLK_CFG_NUM_QUEUES, val_u16);
1013}
1014
c9d989b4
ZL
1015static int vdpa_dev_blk_topology_config_fill(struct sk_buff *msg, u64 features,
1016 const struct virtio_blk_config *config)
1017{
1018 u16 min_io_size;
1019 u32 opt_io_size;
1020
1021 if ((features & BIT_ULL(VIRTIO_BLK_F_TOPOLOGY)) == 0)
1022 return 0;
1023
1024 min_io_size = __virtio16_to_cpu(true, config->min_io_size);
1025 opt_io_size = __virtio32_to_cpu(true, config->opt_io_size);
1026
1027 if (nla_put_u8(msg, VDPA_ATTR_DEV_BLK_CFG_PHY_BLK_EXP,
1028 config->physical_block_exp))
1029 return -EMSGSIZE;
1030
1031 if (nla_put_u8(msg, VDPA_ATTR_DEV_BLK_CFG_ALIGN_OFFSET,
1032 config->alignment_offset))
1033 return -EMSGSIZE;
1034
1035 if (nla_put_u16(msg, VDPA_ATTR_DEV_BLK_CFG_MIN_IO_SIZE, min_io_size))
1036 return -EMSGSIZE;
1037
1038 if (nla_put_u32(msg, VDPA_ATTR_DEV_BLK_CFG_OPT_IO_SIZE, opt_io_size))
1039 return -EMSGSIZE;
1040
1041 return 0;
1042}
1043
65848f46
ZL
1044static int vdpa_dev_blk_discard_config_fill(struct sk_buff *msg, u64 features,
1045 const struct virtio_blk_config *config)
1046{
1047 u32 val_u32;
1048
1049 if ((features & BIT_ULL(VIRTIO_BLK_F_DISCARD)) == 0)
1050 return 0;
1051
1052 val_u32 = __virtio32_to_cpu(true, config->max_discard_sectors);
1053 if (nla_put_u32(msg, VDPA_ATTR_DEV_BLK_CFG_MAX_DISCARD_SEC, val_u32))
1054 return -EMSGSIZE;
1055
1056 val_u32 = __virtio32_to_cpu(true, config->max_discard_seg);
1057 if (nla_put_u32(msg, VDPA_ATTR_DEV_BLK_CFG_MAX_DISCARD_SEG, val_u32))
1058 return -EMSGSIZE;
1059
1060 val_u32 = __virtio32_to_cpu(true, config->discard_sector_alignment);
1061 if (nla_put_u32(msg, VDPA_ATTR_DEV_BLK_CFG_DISCARD_SEC_ALIGN, val_u32))
1062 return -EMSGSIZE;
1063
1064 return 0;
1065}
1066
6bdc7846
ZL
1067static int
1068vdpa_dev_blk_write_zeroes_config_fill(struct sk_buff *msg, u64 features,
1069 const struct virtio_blk_config *config)
1070{
1071 u32 val_u32;
1072
1073 if ((features & BIT_ULL(VIRTIO_BLK_F_WRITE_ZEROES)) == 0)
1074 return 0;
1075
1076 val_u32 = __virtio32_to_cpu(true, config->max_write_zeroes_sectors);
1077 if (nla_put_u32(msg, VDPA_ATTR_DEV_BLK_CFG_MAX_WRITE_ZEROES_SEC, val_u32))
1078 return -EMSGSIZE;
1079
1080 val_u32 = __virtio32_to_cpu(true, config->max_write_zeroes_seg);
1081 if (nla_put_u32(msg, VDPA_ATTR_DEV_BLK_CFG_MAX_WRITE_ZEROES_SEG, val_u32))
1082 return -EMSGSIZE;
1083
1084 return 0;
1085}
1086
ae1374b7
ZL
1087static int vdpa_dev_blk_ro_config_fill(struct sk_buff *msg, u64 features)
1088{
1089 u8 ro;
1090
1091 ro = ((features & BIT_ULL(VIRTIO_BLK_F_RO)) == 0) ? 0 : 1;
98a82154 1092 if (nla_put_u8(msg, VDPA_ATTR_DEV_BLK_READ_ONLY, ro))
ae1374b7
ZL
1093 return -EMSGSIZE;
1094
1095 return 0;
1096}
1097
1ac61ddf
ZL
1098static int vdpa_dev_blk_flush_config_fill(struct sk_buff *msg, u64 features)
1099{
1100 u8 flush;
1101
1102 flush = ((features & BIT_ULL(VIRTIO_BLK_F_FLUSH)) == 0) ? 0 : 1;
98a82154 1103 if (nla_put_u8(msg, VDPA_ATTR_DEV_BLK_FLUSH, flush))
1ac61ddf
ZL
1104 return -EMSGSIZE;
1105
1106 return 0;
1107}
1108
c2475a9a
ZL
1109static int vdpa_dev_blk_config_fill(struct vdpa_device *vdev,
1110 struct sk_buff *msg)
1111{
1112 struct virtio_blk_config config = {};
1113 u64 features_device;
1114
1115 vdev->config->get_config(vdev, 0, &config, sizeof(config));
1116
1117 features_device = vdev->config->get_device_features(vdev);
1118
1119 if (nla_put_u64_64bit(msg, VDPA_ATTR_DEV_FEATURES, features_device,
1120 VDPA_ATTR_PAD))
1121 return -EMSGSIZE;
1122
1123 if (vdpa_dev_blk_capacity_config_fill(msg, &config))
1124 return -EMSGSIZE;
1125
330b8aea
ZL
1126 if (vdpa_dev_blk_seg_size_config_fill(msg, features_device, &config))
1127 return -EMSGSIZE;
1128
3a1d33fb
ZL
1129 if (vdpa_dev_blk_block_size_config_fill(msg, features_device, &config))
1130 return -EMSGSIZE;
1131
81f64e1d
ZL
1132 if (vdpa_dev_blk_seg_max_config_fill(msg, features_device, &config))
1133 return -EMSGSIZE;
1134
54fb04b0
ZL
1135 if (vdpa_dev_blk_mq_config_fill(msg, features_device, &config))
1136 return -EMSGSIZE;
1137
c9d989b4
ZL
1138 if (vdpa_dev_blk_topology_config_fill(msg, features_device, &config))
1139 return -EMSGSIZE;
1140
65848f46
ZL
1141 if (vdpa_dev_blk_discard_config_fill(msg, features_device, &config))
1142 return -EMSGSIZE;
1143
6bdc7846
ZL
1144 if (vdpa_dev_blk_write_zeroes_config_fill(msg, features_device, &config))
1145 return -EMSGSIZE;
1146
ae1374b7
ZL
1147 if (vdpa_dev_blk_ro_config_fill(msg, features_device))
1148 return -EMSGSIZE;
1149
1ac61ddf
ZL
1150 if (vdpa_dev_blk_flush_config_fill(msg, features_device))
1151 return -EMSGSIZE;
1152
c2475a9a
ZL
1153 return 0;
1154}
1155
ad69dd0b
PP
1156static int
1157vdpa_dev_config_fill(struct vdpa_device *vdev, struct sk_buff *msg, u32 portid, u32 seq,
1158 int flags, struct netlink_ext_ack *extack)
1159{
c6dac2ec
ZL
1160 u64 features_driver;
1161 u8 status = 0;
ad69dd0b
PP
1162 u32 device_id;
1163 void *hdr;
1164 int err;
1165
a6a51adc 1166 down_read(&vdev->cf_lock);
ad69dd0b
PP
1167 hdr = genlmsg_put(msg, portid, seq, &vdpa_nl_family, flags,
1168 VDPA_CMD_DEV_CONFIG_GET);
30ef7a8a
EC
1169 if (!hdr) {
1170 err = -EMSGSIZE;
1171 goto out;
1172 }
ad69dd0b
PP
1173
1174 if (nla_put_string(msg, VDPA_ATTR_DEV_NAME, dev_name(&vdev->dev))) {
1175 err = -EMSGSIZE;
1176 goto msg_err;
1177 }
1178
1179 device_id = vdev->config->get_device_id(vdev);
1180 if (nla_put_u32(msg, VDPA_ATTR_DEV_ID, device_id)) {
1181 err = -EMSGSIZE;
1182 goto msg_err;
1183 }
1184
c6dac2ec
ZL
1185 /* only read driver features after the feature negotiation is done */
1186 status = vdev->config->get_status(vdev);
1187 if (status & VIRTIO_CONFIG_S_FEATURES_OK) {
1188 features_driver = vdev->config->get_driver_features(vdev);
1189 if (nla_put_u64_64bit(msg, VDPA_ATTR_DEV_NEGOTIATED_FEATURES, features_driver,
1190 VDPA_ATTR_PAD)) {
1191 err = -EMSGSIZE;
1192 goto msg_err;
1193 }
1194 }
1195
ad69dd0b
PP
1196 switch (device_id) {
1197 case VIRTIO_ID_NET:
1198 err = vdpa_dev_net_config_fill(vdev, msg);
1199 break;
c2475a9a
ZL
1200 case VIRTIO_ID_BLOCK:
1201 err = vdpa_dev_blk_config_fill(vdev, msg);
1202 break;
ad69dd0b
PP
1203 default:
1204 err = -EOPNOTSUPP;
1205 break;
1206 }
1207 if (err)
1208 goto msg_err;
1209
a6a51adc 1210 up_read(&vdev->cf_lock);
ad69dd0b
PP
1211 genlmsg_end(msg, hdr);
1212 return 0;
1213
1214msg_err:
1215 genlmsg_cancel(msg, hdr);
30ef7a8a 1216out:
a6a51adc 1217 up_read(&vdev->cf_lock);
ad69dd0b
PP
1218 return err;
1219}
1220
13b00b13
EC
1221static int vdpa_fill_stats_rec(struct vdpa_device *vdev, struct sk_buff *msg,
1222 struct genl_info *info, u32 index)
1223{
1224 struct virtio_net_config config = {};
1225 u64 features;
13b00b13
EC
1226 u8 status;
1227 int err;
1228
1229 status = vdev->config->get_status(vdev);
1230 if (!(status & VIRTIO_CONFIG_S_FEATURES_OK)) {
1231 NL_SET_ERR_MSG_MOD(info->extack, "feature negotiation not complete");
1232 return -EAGAIN;
1233 }
1234 vdpa_get_config_unlocked(vdev, 0, &config, sizeof(config));
1235
13b00b13
EC
1236 features = vdev->config->get_driver_features(vdev);
1237 if (nla_put_u64_64bit(msg, VDPA_ATTR_DEV_NEGOTIATED_FEATURES,
1238 features, VDPA_ATTR_PAD))
1239 return -EMSGSIZE;
1240
1c96d545
JW
1241 err = vdpa_dev_net_mq_config_fill(msg, features, &config);
1242 if (err)
1243 return err;
1244
13b00b13
EC
1245 if (nla_put_u32(msg, VDPA_ATTR_DEV_QUEUE_INDEX, index))
1246 return -EMSGSIZE;
1247
1248 err = vdev->config->get_vendor_vq_stats(vdev, index, msg, info->extack);
1249 if (err)
1250 return err;
1251
1252 return 0;
1253}
1254
1255static int vendor_stats_fill(struct vdpa_device *vdev, struct sk_buff *msg,
1256 struct genl_info *info, u32 index)
1257{
1258 int err;
1259
a6a51adc 1260 down_read(&vdev->cf_lock);
13b00b13
EC
1261 if (!vdev->config->get_vendor_vq_stats) {
1262 err = -EOPNOTSUPP;
1263 goto out;
1264 }
1265
1266 err = vdpa_fill_stats_rec(vdev, msg, info, index);
1267out:
a6a51adc 1268 up_read(&vdev->cf_lock);
13b00b13
EC
1269 return err;
1270}
1271
1272static int vdpa_dev_vendor_stats_fill(struct vdpa_device *vdev,
1273 struct sk_buff *msg,
1274 struct genl_info *info, u32 index)
1275{
1276 u32 device_id;
1277 void *hdr;
1278 int err;
1279 u32 portid = info->snd_portid;
1280 u32 seq = info->snd_seq;
1281 u32 flags = 0;
1282
1283 hdr = genlmsg_put(msg, portid, seq, &vdpa_nl_family, flags,
1284 VDPA_CMD_DEV_VSTATS_GET);
1285 if (!hdr)
1286 return -EMSGSIZE;
1287
1288 if (nla_put_string(msg, VDPA_ATTR_DEV_NAME, dev_name(&vdev->dev))) {
1289 err = -EMSGSIZE;
1290 goto undo_msg;
1291 }
1292
1293 device_id = vdev->config->get_device_id(vdev);
1294 if (nla_put_u32(msg, VDPA_ATTR_DEV_ID, device_id)) {
1295 err = -EMSGSIZE;
1296 goto undo_msg;
1297 }
1298
1299 switch (device_id) {
1300 case VIRTIO_ID_NET:
1301 if (index > VIRTIO_NET_CTRL_MQ_VQ_PAIRS_MAX) {
699209fc 1302 NL_SET_ERR_MSG_MOD(info->extack, "queue index exceeds max value");
13b00b13
EC
1303 err = -ERANGE;
1304 break;
1305 }
1306
1307 err = vendor_stats_fill(vdev, msg, info, index);
1308 break;
1309 default:
1310 err = -EOPNOTSUPP;
1311 break;
1312 }
1313 genlmsg_end(msg, hdr);
1314
1315 return err;
1316
1317undo_msg:
1318 genlmsg_cancel(msg, hdr);
1319 return err;
1320}
1321
ad69dd0b
PP
1322static int vdpa_nl_cmd_dev_config_get_doit(struct sk_buff *skb, struct genl_info *info)
1323{
1324 struct vdpa_device *vdev;
1325 struct sk_buff *msg;
1326 const char *devname;
1327 struct device *dev;
1328 int err;
1329
1330 if (!info->attrs[VDPA_ATTR_DEV_NAME])
1331 return -EINVAL;
1332 devname = nla_data(info->attrs[VDPA_ATTR_DEV_NAME]);
1333 msg = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL);
1334 if (!msg)
1335 return -ENOMEM;
1336
0078ad90 1337 down_read(&vdpa_dev_lock);
ad69dd0b
PP
1338 dev = bus_find_device(&vdpa_bus, NULL, devname, vdpa_name_match);
1339 if (!dev) {
1340 NL_SET_ERR_MSG_MOD(info->extack, "device not found");
1341 err = -ENODEV;
1342 goto dev_err;
1343 }
1344 vdev = container_of(dev, struct vdpa_device, dev);
1345 if (!vdev->mdev) {
1346 NL_SET_ERR_MSG_MOD(info->extack, "unmanaged vdpa device");
1347 err = -EINVAL;
1348 goto mdev_err;
1349 }
1350 err = vdpa_dev_config_fill(vdev, msg, info->snd_portid, info->snd_seq,
1351 0, info->extack);
1352 if (!err)
1353 err = genlmsg_reply(msg, info);
1354
1355mdev_err:
1356 put_device(dev);
1357dev_err:
0078ad90 1358 up_read(&vdpa_dev_lock);
ad69dd0b
PP
1359 if (err)
1360 nlmsg_free(msg);
1361 return err;
1362}
1363
2f87e9cf
CL
1364static int vdpa_dev_net_device_attr_set(struct vdpa_device *vdev,
1365 struct genl_info *info)
1366{
1367 struct vdpa_dev_set_config set_config = {};
1368 struct vdpa_mgmt_dev *mdev = vdev->mdev;
1369 struct nlattr **nl_attrs = info->attrs;
1370 const u8 *macaddr;
1371 int err = -EOPNOTSUPP;
1372
1373 down_write(&vdev->cf_lock);
1374 if (nl_attrs[VDPA_ATTR_DEV_NET_CFG_MACADDR]) {
1375 set_config.mask |= BIT_ULL(VDPA_ATTR_DEV_NET_CFG_MACADDR);
1376 macaddr = nla_data(nl_attrs[VDPA_ATTR_DEV_NET_CFG_MACADDR]);
1377
1378 if (is_valid_ether_addr(macaddr)) {
1379 ether_addr_copy(set_config.net.mac, macaddr);
1380 if (mdev->ops->dev_set_attr) {
1381 err = mdev->ops->dev_set_attr(mdev, vdev,
1382 &set_config);
1383 } else {
1384 NL_SET_ERR_MSG_FMT_MOD(info->extack,
1385 "Operation not supported by the device.");
1386 }
1387 } else {
1388 NL_SET_ERR_MSG_FMT_MOD(info->extack,
1389 "Invalid MAC address");
1390 }
1391 }
1392 up_write(&vdev->cf_lock);
1393 return err;
1394}
1395
1396static int vdpa_nl_cmd_dev_attr_set_doit(struct sk_buff *skb,
1397 struct genl_info *info)
1398{
1399 struct vdpa_device *vdev;
1400 struct device *dev;
1401 const char *name;
1402 u64 classes;
1403 int err = 0;
1404
1405 if (!info->attrs[VDPA_ATTR_DEV_NAME])
1406 return -EINVAL;
1407
1408 name = nla_data(info->attrs[VDPA_ATTR_DEV_NAME]);
1409
1410 down_write(&vdpa_dev_lock);
1411 dev = bus_find_device(&vdpa_bus, NULL, name, vdpa_name_match);
1412 if (!dev) {
1413 NL_SET_ERR_MSG_MOD(info->extack, "device not found");
1414 err = -ENODEV;
1415 goto dev_err;
1416 }
1417 vdev = container_of(dev, struct vdpa_device, dev);
1418 if (!vdev->mdev) {
1419 NL_SET_ERR_MSG_MOD(info->extack, "unmanaged vdpa device");
1420 err = -EINVAL;
1421 goto mdev_err;
1422 }
1423 classes = vdpa_mgmtdev_get_classes(vdev->mdev, NULL);
1424 if (classes & BIT_ULL(VIRTIO_ID_NET)) {
1425 err = vdpa_dev_net_device_attr_set(vdev, info);
1426 } else {
1427 NL_SET_ERR_MSG_FMT_MOD(info->extack, "%s device not supported",
1428 name);
1429 }
1430
1431mdev_err:
1432 put_device(dev);
1433dev_err:
1434 up_write(&vdpa_dev_lock);
1435 return err;
1436}
1437
ad69dd0b
PP
1438static int vdpa_dev_config_dump(struct device *dev, void *data)
1439{
1440 struct vdpa_device *vdev = container_of(dev, struct vdpa_device, dev);
1441 struct vdpa_dev_dump_info *info = data;
1442 int err;
1443
1444 if (!vdev->mdev)
1445 return 0;
1446 if (info->idx < info->start_idx) {
1447 info->idx++;
1448 return 0;
1449 }
1450 err = vdpa_dev_config_fill(vdev, info->msg, NETLINK_CB(info->cb->skb).portid,
1451 info->cb->nlh->nlmsg_seq, NLM_F_MULTI,
1452 info->cb->extack);
1453 if (err)
1454 return err;
1455
1456 info->idx++;
1457 return 0;
1458}
1459
1460static int
1461vdpa_nl_cmd_dev_config_get_dumpit(struct sk_buff *msg, struct netlink_callback *cb)
1462{
1463 struct vdpa_dev_dump_info info;
1464
1465 info.msg = msg;
1466 info.cb = cb;
1467 info.start_idx = cb->args[0];
1468 info.idx = 0;
1469
0078ad90 1470 down_read(&vdpa_dev_lock);
ad69dd0b 1471 bus_for_each_dev(&vdpa_bus, NULL, &info, vdpa_dev_config_dump);
0078ad90 1472 up_read(&vdpa_dev_lock);
ad69dd0b
PP
1473 cb->args[0] = info.idx;
1474 return msg->len;
1475}
1476
13b00b13
EC
1477static int vdpa_nl_cmd_dev_stats_get_doit(struct sk_buff *skb,
1478 struct genl_info *info)
1479{
1480 struct vdpa_device *vdev;
1481 struct sk_buff *msg;
1482 const char *devname;
1483 struct device *dev;
1484 u32 index;
1485 int err;
1486
1487 if (!info->attrs[VDPA_ATTR_DEV_NAME])
1488 return -EINVAL;
1489
1490 if (!info->attrs[VDPA_ATTR_DEV_QUEUE_INDEX])
1491 return -EINVAL;
1492
1493 devname = nla_data(info->attrs[VDPA_ATTR_DEV_NAME]);
1494 msg = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL);
1495 if (!msg)
1496 return -ENOMEM;
1497
1498 index = nla_get_u32(info->attrs[VDPA_ATTR_DEV_QUEUE_INDEX]);
0078ad90 1499 down_read(&vdpa_dev_lock);
13b00b13
EC
1500 dev = bus_find_device(&vdpa_bus, NULL, devname, vdpa_name_match);
1501 if (!dev) {
1502 NL_SET_ERR_MSG_MOD(info->extack, "device not found");
1503 err = -ENODEV;
1504 goto dev_err;
1505 }
1506 vdev = container_of(dev, struct vdpa_device, dev);
1507 if (!vdev->mdev) {
1508 NL_SET_ERR_MSG_MOD(info->extack, "unmanaged vdpa device");
1509 err = -EINVAL;
1510 goto mdev_err;
1511 }
1512 err = vdpa_dev_vendor_stats_fill(vdev, msg, info, index);
1513 if (err)
1514 goto mdev_err;
1515
1516 err = genlmsg_reply(msg, info);
1517
1518 put_device(dev);
0078ad90 1519 up_read(&vdpa_dev_lock);
13b00b13
EC
1520
1521 return err;
1522
1523mdev_err:
1524 put_device(dev);
1525dev_err:
1526 nlmsg_free(msg);
0078ad90 1527 up_read(&vdpa_dev_lock);
13b00b13
EC
1528 return err;
1529}
1530
33b34750
PP
1531static const struct nla_policy vdpa_nl_policy[VDPA_ATTR_MAX + 1] = {
1532 [VDPA_ATTR_MGMTDEV_BUS_NAME] = { .type = NLA_NUL_STRING },
1533 [VDPA_ATTR_MGMTDEV_DEV_NAME] = { .type = NLA_STRING },
903f7bca 1534 [VDPA_ATTR_DEV_NAME] = { .type = NLA_STRING },
d8ca2fa5 1535 [VDPA_ATTR_DEV_NET_CFG_MACADDR] = NLA_POLICY_ETH_ADDR,
5d6ba607 1536 [VDPA_ATTR_DEV_NET_CFG_MAX_VQP] = { .type = NLA_U16 },
d8ca2fa5
PP
1537 /* virtio spec 1.1 section 5.1.4.1 for valid MTU range */
1538 [VDPA_ATTR_DEV_NET_CFG_MTU] = NLA_POLICY_MIN(NLA_U16, 68),
b3003e1b 1539 [VDPA_ATTR_DEV_QUEUE_INDEX] = { .type = NLA_U32 },
79c86515 1540 [VDPA_ATTR_DEV_FEATURES] = { .type = NLA_U64 },
33b34750
PP
1541};
1542
1543static const struct genl_ops vdpa_nl_ops[] = {
1544 {
1545 .cmd = VDPA_CMD_MGMTDEV_GET,
33b34750
PP
1546 .doit = vdpa_nl_cmd_mgmtdev_get_doit,
1547 .dumpit = vdpa_nl_cmd_mgmtdev_get_dumpit,
1548 },
903f7bca
PP
1549 {
1550 .cmd = VDPA_CMD_DEV_NEW,
903f7bca
PP
1551 .doit = vdpa_nl_cmd_dev_add_set_doit,
1552 .flags = GENL_ADMIN_PERM,
1553 },
1554 {
1555 .cmd = VDPA_CMD_DEV_DEL,
903f7bca
PP
1556 .doit = vdpa_nl_cmd_dev_del_set_doit,
1557 .flags = GENL_ADMIN_PERM,
1558 },
bc0d90ee
PP
1559 {
1560 .cmd = VDPA_CMD_DEV_GET,
bc0d90ee
PP
1561 .doit = vdpa_nl_cmd_dev_get_doit,
1562 .dumpit = vdpa_nl_cmd_dev_get_dumpit,
1563 },
ad69dd0b
PP
1564 {
1565 .cmd = VDPA_CMD_DEV_CONFIG_GET,
ad69dd0b
PP
1566 .doit = vdpa_nl_cmd_dev_config_get_doit,
1567 .dumpit = vdpa_nl_cmd_dev_config_get_dumpit,
1568 },
13b00b13
EC
1569 {
1570 .cmd = VDPA_CMD_DEV_VSTATS_GET,
13b00b13
EC
1571 .doit = vdpa_nl_cmd_dev_stats_get_doit,
1572 .flags = GENL_ADMIN_PERM,
1573 },
2f87e9cf
CL
1574 {
1575 .cmd = VDPA_CMD_DEV_ATTR_SET,
1576 .doit = vdpa_nl_cmd_dev_attr_set_doit,
1577 .flags = GENL_ADMIN_PERM,
1578 },
33b34750
PP
1579};
1580
1581static struct genl_family vdpa_nl_family __ro_after_init = {
1582 .name = VDPA_GENL_NAME,
1583 .version = VDPA_GENL_VERSION,
1584 .maxattr = VDPA_ATTR_MAX,
1585 .policy = vdpa_nl_policy,
1586 .netnsok = false,
1587 .module = THIS_MODULE,
1588 .ops = vdpa_nl_ops,
1589 .n_ops = ARRAY_SIZE(vdpa_nl_ops),
9c5d03d3 1590 .resv_start_op = VDPA_CMD_DEV_VSTATS_GET + 1,
33b34750
PP
1591};
1592
961e9c84
JW
1593static int vdpa_init(void)
1594{
33b34750
PP
1595 int err;
1596
1597 err = bus_register(&vdpa_bus);
1598 if (err)
1599 return err;
1600 err = genl_register_family(&vdpa_nl_family);
1601 if (err)
1602 goto err;
1603 return 0;
1604
1605err:
1606 bus_unregister(&vdpa_bus);
1607 return err;
961e9c84
JW
1608}
1609
1610static void __exit vdpa_exit(void)
1611{
33b34750 1612 genl_unregister_family(&vdpa_nl_family);
961e9c84
JW
1613 bus_unregister(&vdpa_bus);
1614 ida_destroy(&vdpa_index_ida);
1615}
1616core_initcall(vdpa_init);
1617module_exit(vdpa_exit);
1618
1619MODULE_AUTHOR("Jason Wang <[email protected]>");
e7909ad6 1620MODULE_DESCRIPTION("vDPA bus");
961e9c84 1621MODULE_LICENSE("GPL v2");
This page took 0.703747 seconds and 4 git commands to generate.