]>
Commit | Line | Data |
---|---|---|
02eb84d0 MT |
1 | /* |
2 | * MSI-X device support | |
3 | * | |
4 | * This module includes support for MSI-X in pci devices. | |
5 | * | |
6 | * Author: Michael S. Tsirkin <[email protected]> | |
7 | * | |
8 | * Copyright (c) 2009, Red Hat Inc, Michael S. Tsirkin ([email protected]) | |
9 | * | |
10 | * This work is licensed under the terms of the GNU GPL, version 2. See | |
11 | * the COPYING file in the top-level directory. | |
6b620ca3 PB |
12 | * |
13 | * Contributions after 2012-01-13 are licensed under the terms of the | |
14 | * GNU GPL, version 2 or (at your option) any later version. | |
02eb84d0 MT |
15 | */ |
16 | ||
97d5408f | 17 | #include "qemu/osdep.h" |
c759b24f MT |
18 | #include "hw/hw.h" |
19 | #include "hw/pci/msi.h" | |
20 | #include "hw/pci/msix.h" | |
21 | #include "hw/pci/pci.h" | |
428c3ece | 22 | #include "hw/xen/xen.h" |
1de7afc9 | 23 | #include "qemu/range.h" |
02eb84d0 | 24 | |
02eb84d0 MT |
25 | #define MSIX_CAP_LENGTH 12 |
26 | ||
2760952b MT |
27 | /* MSI enable bit and maskall bit are in byte 1 in FLAGS register */ |
28 | #define MSIX_CONTROL_OFFSET (PCI_MSIX_FLAGS + 1) | |
02eb84d0 | 29 | #define MSIX_ENABLE_MASK (PCI_MSIX_FLAGS_ENABLE >> 8) |
5b5cb086 | 30 | #define MSIX_MASKALL_MASK (PCI_MSIX_FLAGS_MASKALL >> 8) |
02eb84d0 | 31 | |
4c93bfa9 | 32 | MSIMessage msix_get_message(PCIDevice *dev, unsigned vector) |
bc4caf49 | 33 | { |
d35e428c | 34 | uint8_t *table_entry = dev->msix_table + vector * PCI_MSIX_ENTRY_SIZE; |
bc4caf49 JK |
35 | MSIMessage msg; |
36 | ||
37 | msg.address = pci_get_quad(table_entry + PCI_MSIX_ENTRY_LOWER_ADDR); | |
38 | msg.data = pci_get_long(table_entry + PCI_MSIX_ENTRY_DATA); | |
39 | return msg; | |
40 | } | |
02eb84d0 | 41 | |
932d4a42 AK |
42 | /* |
43 | * Special API for POWER to configure the vectors through | |
44 | * a side channel. Should never be used by devices. | |
45 | */ | |
46 | void msix_set_message(PCIDevice *dev, int vector, struct MSIMessage msg) | |
47 | { | |
48 | uint8_t *table_entry = dev->msix_table + vector * PCI_MSIX_ENTRY_SIZE; | |
49 | ||
50 | pci_set_quad(table_entry + PCI_MSIX_ENTRY_LOWER_ADDR, msg.address); | |
51 | pci_set_long(table_entry + PCI_MSIX_ENTRY_DATA, msg.data); | |
52 | table_entry[PCI_MSIX_ENTRY_VECTOR_CTRL] &= ~PCI_MSIX_ENTRY_CTRL_MASKBIT; | |
53 | } | |
54 | ||
02eb84d0 MT |
55 | static uint8_t msix_pending_mask(int vector) |
56 | { | |
57 | return 1 << (vector % 8); | |
58 | } | |
59 | ||
60 | static uint8_t *msix_pending_byte(PCIDevice *dev, int vector) | |
61 | { | |
d35e428c | 62 | return dev->msix_pba + vector / 8; |
02eb84d0 MT |
63 | } |
64 | ||
65 | static int msix_is_pending(PCIDevice *dev, int vector) | |
66 | { | |
67 | return *msix_pending_byte(dev, vector) & msix_pending_mask(vector); | |
68 | } | |
69 | ||
70f8ee39 | 70 | void msix_set_pending(PCIDevice *dev, unsigned int vector) |
02eb84d0 MT |
71 | { |
72 | *msix_pending_byte(dev, vector) |= msix_pending_mask(vector); | |
73 | } | |
74 | ||
3bdfaabb | 75 | void msix_clr_pending(PCIDevice *dev, int vector) |
02eb84d0 MT |
76 | { |
77 | *msix_pending_byte(dev, vector) &= ~msix_pending_mask(vector); | |
78 | } | |
79 | ||
70f8ee39 | 80 | static bool msix_vector_masked(PCIDevice *dev, unsigned int vector, bool fmask) |
02eb84d0 | 81 | { |
428c3ece | 82 | unsigned offset = vector * PCI_MSIX_ENTRY_SIZE; |
e1e4bf22 | 83 | uint8_t *data = &dev->msix_table[offset + PCI_MSIX_ENTRY_DATA]; |
428c3ece SS |
84 | /* MSIs on Xen can be remapped into pirqs. In those cases, masking |
85 | * and unmasking go through the PV evtchn path. */ | |
e1e4bf22 | 86 | if (xen_enabled() && xen_is_pirq_msi(pci_get_long(data))) { |
428c3ece SS |
87 | return false; |
88 | } | |
89 | return fmask || dev->msix_table[offset + PCI_MSIX_ENTRY_VECTOR_CTRL] & | |
90 | PCI_MSIX_ENTRY_CTRL_MASKBIT; | |
5b5cb086 MT |
91 | } |
92 | ||
70f8ee39 | 93 | bool msix_is_masked(PCIDevice *dev, unsigned int vector) |
5b5cb086 | 94 | { |
ae392c41 MT |
95 | return msix_vector_masked(dev, vector, dev->msix_function_masked); |
96 | } | |
97 | ||
2cdfe53c JK |
98 | static void msix_fire_vector_notifier(PCIDevice *dev, |
99 | unsigned int vector, bool is_masked) | |
100 | { | |
101 | MSIMessage msg; | |
102 | int ret; | |
103 | ||
104 | if (!dev->msix_vector_use_notifier) { | |
105 | return; | |
106 | } | |
107 | if (is_masked) { | |
108 | dev->msix_vector_release_notifier(dev, vector); | |
109 | } else { | |
110 | msg = msix_get_message(dev, vector); | |
111 | ret = dev->msix_vector_use_notifier(dev, vector, msg); | |
112 | assert(ret >= 0); | |
113 | } | |
114 | } | |
115 | ||
ae392c41 MT |
116 | static void msix_handle_mask_update(PCIDevice *dev, int vector, bool was_masked) |
117 | { | |
118 | bool is_masked = msix_is_masked(dev, vector); | |
2cdfe53c | 119 | |
ae392c41 MT |
120 | if (is_masked == was_masked) { |
121 | return; | |
122 | } | |
123 | ||
2cdfe53c JK |
124 | msix_fire_vector_notifier(dev, vector, is_masked); |
125 | ||
ae392c41 | 126 | if (!is_masked && msix_is_pending(dev, vector)) { |
5b5cb086 MT |
127 | msix_clr_pending(dev, vector); |
128 | msix_notify(dev, vector); | |
129 | } | |
130 | } | |
131 | ||
50322249 MT |
132 | static void msix_update_function_masked(PCIDevice *dev) |
133 | { | |
134 | dev->msix_function_masked = !msix_enabled(dev) || | |
135 | (dev->config[dev->msix_cap + MSIX_CONTROL_OFFSET] & MSIX_MASKALL_MASK); | |
136 | } | |
137 | ||
5b5cb086 MT |
138 | /* Handle MSI-X capability config write. */ |
139 | void msix_write_config(PCIDevice *dev, uint32_t addr, | |
140 | uint32_t val, int len) | |
141 | { | |
142 | unsigned enable_pos = dev->msix_cap + MSIX_CONTROL_OFFSET; | |
143 | int vector; | |
50322249 | 144 | bool was_masked; |
5b5cb086 | 145 | |
7c9958b0 | 146 | if (!msix_present(dev) || !range_covers_byte(addr, len, enable_pos)) { |
5b5cb086 MT |
147 | return; |
148 | } | |
149 | ||
50322249 MT |
150 | was_masked = dev->msix_function_masked; |
151 | msix_update_function_masked(dev); | |
152 | ||
5b5cb086 MT |
153 | if (!msix_enabled(dev)) { |
154 | return; | |
155 | } | |
156 | ||
e407bf13 | 157 | pci_device_deassert_intx(dev); |
5b5cb086 | 158 | |
50322249 | 159 | if (dev->msix_function_masked == was_masked) { |
5b5cb086 MT |
160 | return; |
161 | } | |
162 | ||
163 | for (vector = 0; vector < dev->msix_entries_nr; ++vector) { | |
ae392c41 MT |
164 | msix_handle_mask_update(dev, vector, |
165 | msix_vector_masked(dev, vector, was_masked)); | |
5b5cb086 | 166 | } |
02eb84d0 MT |
167 | } |
168 | ||
a8170e5e | 169 | static uint64_t msix_table_mmio_read(void *opaque, hwaddr addr, |
d35e428c | 170 | unsigned size) |
eebcb0a7 AW |
171 | { |
172 | PCIDevice *dev = opaque; | |
eebcb0a7 | 173 | |
d35e428c | 174 | return pci_get_long(dev->msix_table + addr); |
eebcb0a7 AW |
175 | } |
176 | ||
a8170e5e | 177 | static void msix_table_mmio_write(void *opaque, hwaddr addr, |
d35e428c | 178 | uint64_t val, unsigned size) |
02eb84d0 MT |
179 | { |
180 | PCIDevice *dev = opaque; | |
d35e428c | 181 | int vector = addr / PCI_MSIX_ENTRY_SIZE; |
ae392c41 | 182 | bool was_masked; |
9a93b617 | 183 | |
ae392c41 | 184 | was_masked = msix_is_masked(dev, vector); |
d35e428c | 185 | pci_set_long(dev->msix_table + addr, val); |
ae392c41 | 186 | msix_handle_mask_update(dev, vector, was_masked); |
02eb84d0 MT |
187 | } |
188 | ||
d35e428c AW |
189 | static const MemoryRegionOps msix_table_mmio_ops = { |
190 | .read = msix_table_mmio_read, | |
191 | .write = msix_table_mmio_write, | |
68d1e1f5 | 192 | .endianness = DEVICE_LITTLE_ENDIAN, |
d35e428c AW |
193 | .valid = { |
194 | .min_access_size = 4, | |
195 | .max_access_size = 4, | |
196 | }, | |
197 | }; | |
198 | ||
a8170e5e | 199 | static uint64_t msix_pba_mmio_read(void *opaque, hwaddr addr, |
d35e428c AW |
200 | unsigned size) |
201 | { | |
202 | PCIDevice *dev = opaque; | |
bbef882c MT |
203 | if (dev->msix_vector_poll_notifier) { |
204 | unsigned vector_start = addr * 8; | |
205 | unsigned vector_end = MIN(addr + size * 8, dev->msix_entries_nr); | |
206 | dev->msix_vector_poll_notifier(dev, vector_start, vector_end); | |
207 | } | |
d35e428c AW |
208 | |
209 | return pci_get_long(dev->msix_pba + addr); | |
210 | } | |
211 | ||
43b11a91 MAL |
212 | static void msix_pba_mmio_write(void *opaque, hwaddr addr, |
213 | uint64_t val, unsigned size) | |
214 | { | |
215 | } | |
216 | ||
d35e428c AW |
217 | static const MemoryRegionOps msix_pba_mmio_ops = { |
218 | .read = msix_pba_mmio_read, | |
43b11a91 | 219 | .write = msix_pba_mmio_write, |
68d1e1f5 | 220 | .endianness = DEVICE_LITTLE_ENDIAN, |
95524ae8 AK |
221 | .valid = { |
222 | .min_access_size = 4, | |
223 | .max_access_size = 4, | |
224 | }, | |
02eb84d0 MT |
225 | }; |
226 | ||
ae1be0bb MT |
227 | static void msix_mask_all(struct PCIDevice *dev, unsigned nentries) |
228 | { | |
229 | int vector; | |
5b5f1330 | 230 | |
ae1be0bb | 231 | for (vector = 0; vector < nentries; ++vector) { |
01731cfb JK |
232 | unsigned offset = |
233 | vector * PCI_MSIX_ENTRY_SIZE + PCI_MSIX_ENTRY_VECTOR_CTRL; | |
5b5f1330 JK |
234 | bool was_masked = msix_is_masked(dev, vector); |
235 | ||
d35e428c | 236 | dev->msix_table[offset] |= PCI_MSIX_ENTRY_CTRL_MASKBIT; |
5b5f1330 | 237 | msix_handle_mask_update(dev, vector, was_masked); |
ae1be0bb MT |
238 | } |
239 | } | |
240 | ||
5a2c2029 | 241 | /* Initialize the MSI-X structures */ |
02eb84d0 | 242 | int msix_init(struct PCIDevice *dev, unsigned short nentries, |
5a2c2029 AW |
243 | MemoryRegion *table_bar, uint8_t table_bar_nr, |
244 | unsigned table_offset, MemoryRegion *pba_bar, | |
245 | uint8_t pba_bar_nr, unsigned pba_offset, uint8_t cap_pos) | |
02eb84d0 | 246 | { |
5a2c2029 | 247 | int cap; |
d35e428c | 248 | unsigned table_size, pba_size; |
5a2c2029 | 249 | uint8_t *config; |
60ba3cc2 | 250 | |
02eb84d0 | 251 | /* Nothing to do if MSI is not supported by interrupt controller */ |
226419d6 | 252 | if (!msi_nonbroken) { |
02eb84d0 | 253 | return -ENOTSUP; |
60ba3cc2 | 254 | } |
5a2c2029 AW |
255 | |
256 | if (nentries < 1 || nentries > PCI_MSIX_FLAGS_QSIZE + 1) { | |
02eb84d0 | 257 | return -EINVAL; |
5a2c2029 | 258 | } |
02eb84d0 | 259 | |
d35e428c AW |
260 | table_size = nentries * PCI_MSIX_ENTRY_SIZE; |
261 | pba_size = QEMU_ALIGN_UP(nentries, 64) / 8; | |
262 | ||
5a2c2029 AW |
263 | /* Sanity test: table & pba don't overlap, fit within BARs, min aligned */ |
264 | if ((table_bar_nr == pba_bar_nr && | |
265 | ranges_overlap(table_offset, table_size, pba_offset, pba_size)) || | |
266 | table_offset + table_size > memory_region_size(table_bar) || | |
267 | pba_offset + pba_size > memory_region_size(pba_bar) || | |
268 | (table_offset | pba_offset) & PCI_MSIX_FLAGS_BIRMASK) { | |
269 | return -EINVAL; | |
270 | } | |
271 | ||
272 | cap = pci_add_capability(dev, PCI_CAP_ID_MSIX, cap_pos, MSIX_CAP_LENGTH); | |
273 | if (cap < 0) { | |
274 | return cap; | |
275 | } | |
276 | ||
277 | dev->msix_cap = cap; | |
278 | dev->cap_present |= QEMU_PCI_CAP_MSIX; | |
279 | config = dev->config + cap; | |
280 | ||
281 | pci_set_word(config + PCI_MSIX_FLAGS, nentries - 1); | |
282 | dev->msix_entries_nr = nentries; | |
283 | dev->msix_function_masked = true; | |
284 | ||
285 | pci_set_long(config + PCI_MSIX_TABLE, table_offset | table_bar_nr); | |
286 | pci_set_long(config + PCI_MSIX_PBA, pba_offset | pba_bar_nr); | |
287 | ||
288 | /* Make flags bit writable. */ | |
289 | dev->wmask[cap + MSIX_CONTROL_OFFSET] |= MSIX_ENABLE_MASK | | |
290 | MSIX_MASKALL_MASK; | |
02eb84d0 | 291 | |
d35e428c AW |
292 | dev->msix_table = g_malloc0(table_size); |
293 | dev->msix_pba = g_malloc0(pba_size); | |
5a2c2029 AW |
294 | dev->msix_entry_used = g_malloc0(nentries * sizeof *dev->msix_entry_used); |
295 | ||
ae1be0bb | 296 | msix_mask_all(dev, nentries); |
02eb84d0 | 297 | |
40c5dce9 | 298 | memory_region_init_io(&dev->msix_table_mmio, OBJECT(dev), &msix_table_mmio_ops, dev, |
d35e428c | 299 | "msix-table", table_size); |
5a2c2029 | 300 | memory_region_add_subregion(table_bar, table_offset, &dev->msix_table_mmio); |
40c5dce9 | 301 | memory_region_init_io(&dev->msix_pba_mmio, OBJECT(dev), &msix_pba_mmio_ops, dev, |
d35e428c | 302 | "msix-pba", pba_size); |
5a2c2029 | 303 | memory_region_add_subregion(pba_bar, pba_offset, &dev->msix_pba_mmio); |
02eb84d0 | 304 | |
02eb84d0 | 305 | return 0; |
02eb84d0 MT |
306 | } |
307 | ||
53f94925 AW |
308 | int msix_init_exclusive_bar(PCIDevice *dev, unsigned short nentries, |
309 | uint8_t bar_nr) | |
310 | { | |
311 | int ret; | |
312 | char *name; | |
a0ccd212 JW |
313 | uint32_t bar_size = 4096; |
314 | uint32_t bar_pba_offset = bar_size / 2; | |
315 | uint32_t bar_pba_size = (nentries / 8 + 1) * 8; | |
53f94925 AW |
316 | |
317 | /* | |
318 | * Migration compatibility dictates that this remains a 4k | |
319 | * BAR with the vector table in the lower half and PBA in | |
a0ccd212 JW |
320 | * the upper half for nentries which is lower or equal to 128. |
321 | * No need to care about using more than 65 entries for legacy | |
322 | * machine types who has at most 64 queues. | |
53f94925 | 323 | */ |
a0ccd212 JW |
324 | if (nentries * PCI_MSIX_ENTRY_SIZE > bar_pba_offset) { |
325 | bar_pba_offset = nentries * PCI_MSIX_ENTRY_SIZE; | |
326 | } | |
53f94925 | 327 | |
a0ccd212 JW |
328 | if (bar_pba_offset + bar_pba_size > 4096) { |
329 | bar_size = bar_pba_offset + bar_pba_size; | |
330 | } | |
331 | ||
9bff5d81 | 332 | bar_size = pow2ceil(bar_size); |
53f94925 | 333 | |
5f893b4e | 334 | name = g_strdup_printf("%s-msix", dev->name); |
a0ccd212 | 335 | memory_region_init(&dev->msix_exclusive_bar, OBJECT(dev), name, bar_size); |
5f893b4e | 336 | g_free(name); |
53f94925 AW |
337 | |
338 | ret = msix_init(dev, nentries, &dev->msix_exclusive_bar, bar_nr, | |
a0ccd212 JW |
339 | 0, &dev->msix_exclusive_bar, |
340 | bar_nr, bar_pba_offset, | |
341 | 0); | |
53f94925 | 342 | if (ret) { |
53f94925 AW |
343 | return ret; |
344 | } | |
345 | ||
346 | pci_register_bar(dev, bar_nr, PCI_BASE_ADDRESS_SPACE_MEMORY, | |
347 | &dev->msix_exclusive_bar); | |
348 | ||
349 | return 0; | |
350 | } | |
351 | ||
98304c84 MT |
352 | static void msix_free_irq_entries(PCIDevice *dev) |
353 | { | |
354 | int vector; | |
355 | ||
356 | for (vector = 0; vector < dev->msix_entries_nr; ++vector) { | |
357 | dev->msix_entry_used[vector] = 0; | |
358 | msix_clr_pending(dev, vector); | |
359 | } | |
360 | } | |
361 | ||
3cac001e MT |
362 | static void msix_clear_all_vectors(PCIDevice *dev) |
363 | { | |
364 | int vector; | |
365 | ||
366 | for (vector = 0; vector < dev->msix_entries_nr; ++vector) { | |
367 | msix_clr_pending(dev, vector); | |
368 | } | |
369 | } | |
370 | ||
02eb84d0 | 371 | /* Clean up resources for the device. */ |
572992ee | 372 | void msix_uninit(PCIDevice *dev, MemoryRegion *table_bar, MemoryRegion *pba_bar) |
02eb84d0 | 373 | { |
44701ab7 | 374 | if (!msix_present(dev)) { |
572992ee | 375 | return; |
44701ab7 | 376 | } |
02eb84d0 MT |
377 | pci_del_capability(dev, PCI_CAP_ID_MSIX, MSIX_CAP_LENGTH); |
378 | dev->msix_cap = 0; | |
379 | msix_free_irq_entries(dev); | |
380 | dev->msix_entries_nr = 0; | |
5a2c2029 | 381 | memory_region_del_subregion(pba_bar, &dev->msix_pba_mmio); |
d35e428c AW |
382 | g_free(dev->msix_pba); |
383 | dev->msix_pba = NULL; | |
5a2c2029 | 384 | memory_region_del_subregion(table_bar, &dev->msix_table_mmio); |
d35e428c AW |
385 | g_free(dev->msix_table); |
386 | dev->msix_table = NULL; | |
7267c094 | 387 | g_free(dev->msix_entry_used); |
02eb84d0 MT |
388 | dev->msix_entry_used = NULL; |
389 | dev->cap_present &= ~QEMU_PCI_CAP_MSIX; | |
02eb84d0 MT |
390 | } |
391 | ||
53f94925 AW |
392 | void msix_uninit_exclusive_bar(PCIDevice *dev) |
393 | { | |
394 | if (msix_present(dev)) { | |
5a2c2029 | 395 | msix_uninit(dev, &dev->msix_exclusive_bar, &dev->msix_exclusive_bar); |
53f94925 AW |
396 | } |
397 | } | |
398 | ||
02eb84d0 MT |
399 | void msix_save(PCIDevice *dev, QEMUFile *f) |
400 | { | |
9a3e12c8 MT |
401 | unsigned n = dev->msix_entries_nr; |
402 | ||
44701ab7 | 403 | if (!msix_present(dev)) { |
9a3e12c8 | 404 | return; |
72755a70 | 405 | } |
9a3e12c8 | 406 | |
d35e428c AW |
407 | qemu_put_buffer(f, dev->msix_table, n * PCI_MSIX_ENTRY_SIZE); |
408 | qemu_put_buffer(f, dev->msix_pba, (n + 7) / 8); | |
02eb84d0 MT |
409 | } |
410 | ||
411 | /* Should be called after restoring the config space. */ | |
412 | void msix_load(PCIDevice *dev, QEMUFile *f) | |
413 | { | |
414 | unsigned n = dev->msix_entries_nr; | |
2cdfe53c | 415 | unsigned int vector; |
02eb84d0 | 416 | |
44701ab7 | 417 | if (!msix_present(dev)) { |
02eb84d0 | 418 | return; |
98846d73 | 419 | } |
02eb84d0 | 420 | |
3cac001e | 421 | msix_clear_all_vectors(dev); |
d35e428c AW |
422 | qemu_get_buffer(f, dev->msix_table, n * PCI_MSIX_ENTRY_SIZE); |
423 | qemu_get_buffer(f, dev->msix_pba, (n + 7) / 8); | |
50322249 | 424 | msix_update_function_masked(dev); |
2cdfe53c JK |
425 | |
426 | for (vector = 0; vector < n; vector++) { | |
427 | msix_handle_mask_update(dev, vector, true); | |
428 | } | |
02eb84d0 MT |
429 | } |
430 | ||
431 | /* Does device support MSI-X? */ | |
432 | int msix_present(PCIDevice *dev) | |
433 | { | |
434 | return dev->cap_present & QEMU_PCI_CAP_MSIX; | |
435 | } | |
436 | ||
437 | /* Is MSI-X enabled? */ | |
438 | int msix_enabled(PCIDevice *dev) | |
439 | { | |
440 | return (dev->cap_present & QEMU_PCI_CAP_MSIX) && | |
2760952b | 441 | (dev->config[dev->msix_cap + MSIX_CONTROL_OFFSET] & |
02eb84d0 MT |
442 | MSIX_ENABLE_MASK); |
443 | } | |
444 | ||
02eb84d0 MT |
445 | /* Send an MSI-X message */ |
446 | void msix_notify(PCIDevice *dev, unsigned vector) | |
447 | { | |
bc4caf49 | 448 | MSIMessage msg; |
02eb84d0 MT |
449 | |
450 | if (vector >= dev->msix_entries_nr || !dev->msix_entry_used[vector]) | |
451 | return; | |
452 | if (msix_is_masked(dev, vector)) { | |
453 | msix_set_pending(dev, vector); | |
454 | return; | |
455 | } | |
456 | ||
bc4caf49 JK |
457 | msg = msix_get_message(dev, vector); |
458 | ||
38d40ff1 | 459 | msi_send_message(dev, msg); |
02eb84d0 MT |
460 | } |
461 | ||
462 | void msix_reset(PCIDevice *dev) | |
463 | { | |
44701ab7 | 464 | if (!msix_present(dev)) { |
02eb84d0 | 465 | return; |
44701ab7 | 466 | } |
3cac001e | 467 | msix_clear_all_vectors(dev); |
2760952b MT |
468 | dev->config[dev->msix_cap + MSIX_CONTROL_OFFSET] &= |
469 | ~dev->wmask[dev->msix_cap + MSIX_CONTROL_OFFSET]; | |
d35e428c AW |
470 | memset(dev->msix_table, 0, dev->msix_entries_nr * PCI_MSIX_ENTRY_SIZE); |
471 | memset(dev->msix_pba, 0, QEMU_ALIGN_UP(dev->msix_entries_nr, 64) / 8); | |
ae1be0bb | 472 | msix_mask_all(dev, dev->msix_entries_nr); |
02eb84d0 MT |
473 | } |
474 | ||
475 | /* PCI spec suggests that devices make it possible for software to configure | |
476 | * less vectors than supported by the device, but does not specify a standard | |
477 | * mechanism for devices to do so. | |
478 | * | |
479 | * We support this by asking devices to declare vectors software is going to | |
480 | * actually use, and checking this on the notification path. Devices that | |
481 | * don't want to follow the spec suggestion can declare all vectors as used. */ | |
482 | ||
483 | /* Mark vector as used. */ | |
484 | int msix_vector_use(PCIDevice *dev, unsigned vector) | |
485 | { | |
486 | if (vector >= dev->msix_entries_nr) | |
487 | return -EINVAL; | |
488 | dev->msix_entry_used[vector]++; | |
489 | return 0; | |
490 | } | |
491 | ||
492 | /* Mark vector as unused. */ | |
493 | void msix_vector_unuse(PCIDevice *dev, unsigned vector) | |
494 | { | |
98304c84 MT |
495 | if (vector >= dev->msix_entries_nr || !dev->msix_entry_used[vector]) { |
496 | return; | |
497 | } | |
498 | if (--dev->msix_entry_used[vector]) { | |
499 | return; | |
500 | } | |
501 | msix_clr_pending(dev, vector); | |
02eb84d0 | 502 | } |
b5f28bca MT |
503 | |
504 | void msix_unuse_all_vectors(PCIDevice *dev) | |
505 | { | |
44701ab7 | 506 | if (!msix_present(dev)) { |
b5f28bca | 507 | return; |
44701ab7 | 508 | } |
b5f28bca MT |
509 | msix_free_irq_entries(dev); |
510 | } | |
2cdfe53c | 511 | |
cb697aaa JK |
512 | unsigned int msix_nr_vectors_allocated(const PCIDevice *dev) |
513 | { | |
514 | return dev->msix_entries_nr; | |
515 | } | |
516 | ||
2cdfe53c JK |
517 | static int msix_set_notifier_for_vector(PCIDevice *dev, unsigned int vector) |
518 | { | |
519 | MSIMessage msg; | |
520 | ||
521 | if (msix_is_masked(dev, vector)) { | |
522 | return 0; | |
523 | } | |
524 | msg = msix_get_message(dev, vector); | |
525 | return dev->msix_vector_use_notifier(dev, vector, msg); | |
526 | } | |
527 | ||
528 | static void msix_unset_notifier_for_vector(PCIDevice *dev, unsigned int vector) | |
529 | { | |
530 | if (msix_is_masked(dev, vector)) { | |
531 | return; | |
532 | } | |
533 | dev->msix_vector_release_notifier(dev, vector); | |
534 | } | |
535 | ||
536 | int msix_set_vector_notifiers(PCIDevice *dev, | |
537 | MSIVectorUseNotifier use_notifier, | |
bbef882c MT |
538 | MSIVectorReleaseNotifier release_notifier, |
539 | MSIVectorPollNotifier poll_notifier) | |
2cdfe53c JK |
540 | { |
541 | int vector, ret; | |
542 | ||
543 | assert(use_notifier && release_notifier); | |
544 | ||
545 | dev->msix_vector_use_notifier = use_notifier; | |
546 | dev->msix_vector_release_notifier = release_notifier; | |
bbef882c | 547 | dev->msix_vector_poll_notifier = poll_notifier; |
2cdfe53c JK |
548 | |
549 | if ((dev->config[dev->msix_cap + MSIX_CONTROL_OFFSET] & | |
550 | (MSIX_ENABLE_MASK | MSIX_MASKALL_MASK)) == MSIX_ENABLE_MASK) { | |
551 | for (vector = 0; vector < dev->msix_entries_nr; vector++) { | |
552 | ret = msix_set_notifier_for_vector(dev, vector); | |
553 | if (ret < 0) { | |
554 | goto undo; | |
555 | } | |
556 | } | |
557 | } | |
bbef882c MT |
558 | if (dev->msix_vector_poll_notifier) { |
559 | dev->msix_vector_poll_notifier(dev, 0, dev->msix_entries_nr); | |
560 | } | |
2cdfe53c JK |
561 | return 0; |
562 | ||
563 | undo: | |
564 | while (--vector >= 0) { | |
565 | msix_unset_notifier_for_vector(dev, vector); | |
566 | } | |
567 | dev->msix_vector_use_notifier = NULL; | |
568 | dev->msix_vector_release_notifier = NULL; | |
569 | return ret; | |
570 | } | |
571 | ||
572 | void msix_unset_vector_notifiers(PCIDevice *dev) | |
573 | { | |
574 | int vector; | |
575 | ||
576 | assert(dev->msix_vector_use_notifier && | |
577 | dev->msix_vector_release_notifier); | |
578 | ||
579 | if ((dev->config[dev->msix_cap + MSIX_CONTROL_OFFSET] & | |
580 | (MSIX_ENABLE_MASK | MSIX_MASKALL_MASK)) == MSIX_ENABLE_MASK) { | |
581 | for (vector = 0; vector < dev->msix_entries_nr; vector++) { | |
582 | msix_unset_notifier_for_vector(dev, vector); | |
583 | } | |
584 | } | |
585 | dev->msix_vector_use_notifier = NULL; | |
586 | dev->msix_vector_release_notifier = NULL; | |
bbef882c | 587 | dev->msix_vector_poll_notifier = NULL; |
2cdfe53c | 588 | } |
340b50c7 GH |
589 | |
590 | static void put_msix_state(QEMUFile *f, void *pv, size_t size) | |
591 | { | |
592 | msix_save(pv, f); | |
593 | } | |
594 | ||
595 | static int get_msix_state(QEMUFile *f, void *pv, size_t size) | |
596 | { | |
597 | msix_load(pv, f); | |
598 | return 0; | |
599 | } | |
600 | ||
601 | static VMStateInfo vmstate_info_msix = { | |
602 | .name = "msix state", | |
603 | .get = get_msix_state, | |
604 | .put = put_msix_state, | |
605 | }; | |
606 | ||
607 | const VMStateDescription vmstate_msix = { | |
608 | .name = "msix", | |
609 | .fields = (VMStateField[]) { | |
610 | { | |
611 | .name = "msix", | |
612 | .version_id = 0, | |
613 | .field_exists = NULL, | |
614 | .size = 0, /* ouch */ | |
615 | .info = &vmstate_info_msix, | |
616 | .flags = VMS_SINGLE, | |
617 | .offset = 0, | |
618 | }, | |
619 | VMSTATE_END_OF_LIST() | |
620 | } | |
621 | }; |