FreeBSD/Linux Kernel Cross Reference
sys/x86/x86/msi.c
1 /*-
2 * Copyright (c) 2006 Yahoo!, Inc.
3 * All rights reserved.
4 * Written by: John Baldwin <jhb@FreeBSD.org>
5 *
6 * Redistribution and use in source and binary forms, with or without
7 * modification, are permitted provided that the following conditions
8 * are met:
9 * 1. Redistributions of source code must retain the above copyright
10 * notice, this list of conditions and the following disclaimer.
11 * 2. Redistributions in binary form must reproduce the above copyright
12 * notice, this list of conditions and the following disclaimer in the
13 * documentation and/or other materials provided with the distribution.
14 * 3. Neither the name of the author nor the names of any co-contributors
15 * may be used to endorse or promote products derived from this software
16 * without specific prior written permission.
17 *
18 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
19 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
20 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
21 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
22 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
23 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
24 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
25 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
26 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
27 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
28 * SUCH DAMAGE.
29 */
30
31 /*
32 * Support for PCI Message Signalled Interrupts (MSI). MSI interrupts on
33 * x86 are basically APIC messages that the northbridge delivers directly
34 * to the local APICs as if they had come from an I/O APIC.
35 */
36
37 #include <sys/cdefs.h>
38 __FBSDID("$FreeBSD$");
39
40 #include "opt_acpi.h"
41
42 #include <sys/param.h>
43 #include <sys/bus.h>
44 #include <sys/kernel.h>
45 #include <sys/limits.h>
46 #include <sys/lock.h>
47 #include <sys/malloc.h>
48 #include <sys/mutex.h>
49 #include <sys/sx.h>
50 #include <sys/sysctl.h>
51 #include <sys/systm.h>
52 #include <x86/apicreg.h>
53 #include <machine/cputypes.h>
54 #include <machine/md_var.h>
55 #include <machine/frame.h>
56 #include <machine/intr_machdep.h>
57 #include <x86/apicvar.h>
58 #include <x86/iommu/iommu_intrmap.h>
59 #include <machine/specialreg.h>
60 #include <dev/pci/pcivar.h>
61
62 /* Fields in address for Intel MSI messages. */
63 #define MSI_INTEL_ADDR_DEST 0x000ff000
64 #define MSI_INTEL_ADDR_RH 0x00000008
65 # define MSI_INTEL_ADDR_RH_ON 0x00000008
66 # define MSI_INTEL_ADDR_RH_OFF 0x00000000
67 #define MSI_INTEL_ADDR_DM 0x00000004
68 # define MSI_INTEL_ADDR_DM_PHYSICAL 0x00000000
69 # define MSI_INTEL_ADDR_DM_LOGICAL 0x00000004
70
71 /* Fields in data for Intel MSI messages. */
72 #define MSI_INTEL_DATA_TRGRMOD IOART_TRGRMOD /* Trigger mode. */
73 # define MSI_INTEL_DATA_TRGREDG IOART_TRGREDG
74 # define MSI_INTEL_DATA_TRGRLVL IOART_TRGRLVL
75 #define MSI_INTEL_DATA_LEVEL 0x00004000 /* Polarity. */
76 # define MSI_INTEL_DATA_DEASSERT 0x00000000
77 # define MSI_INTEL_DATA_ASSERT 0x00004000
78 #define MSI_INTEL_DATA_DELMOD IOART_DELMOD /* Delivery mode. */
79 # define MSI_INTEL_DATA_DELFIXED IOART_DELFIXED
80 # define MSI_INTEL_DATA_DELLOPRI IOART_DELLOPRI
81 # define MSI_INTEL_DATA_DELSMI IOART_DELSMI
82 # define MSI_INTEL_DATA_DELNMI IOART_DELNMI
83 # define MSI_INTEL_DATA_DELINIT IOART_DELINIT
84 # define MSI_INTEL_DATA_DELEXINT IOART_DELEXINT
85 #define MSI_INTEL_DATA_INTVEC IOART_INTVEC /* Interrupt vector. */
86
87 /*
88 * Build Intel MSI message and data values from a source. AMD64 systems
89 * seem to be compatible, so we use the same function for both.
90 */
91 #define INTEL_ADDR(msi) \
92 (MSI_INTEL_ADDR_BASE | (msi)->msi_cpu << 12 | \
93 MSI_INTEL_ADDR_RH_OFF | MSI_INTEL_ADDR_DM_PHYSICAL)
94 #define INTEL_DATA(msi) \
95 (MSI_INTEL_DATA_TRGREDG | MSI_INTEL_DATA_DELFIXED | (msi)->msi_vector)
96
97 static MALLOC_DEFINE(M_MSI, "msi", "PCI MSI");
98
99 /*
100 * MSI sources are bunched into groups. This is because MSI forces
101 * all of the messages to share the address and data registers and
102 * thus certain properties (such as the local APIC ID target on x86).
103 * Each group has a 'first' source that contains information global to
104 * the group. These fields are marked with (g) below.
105 *
106 * Note that local APIC ID is kind of special. Each message will be
107 * assigned an ID by the system; however, a group will use the ID from
108 * the first message.
109 *
110 * For MSI-X, each message is isolated.
111 */
112 struct msi_intsrc {
113 struct intsrc msi_intsrc;
114 device_t msi_dev; /* Owning device. (g) */
115 struct msi_intsrc *msi_first; /* First source in group. */
116 u_int msi_irq; /* IRQ cookie. */
117 u_int msi_msix; /* MSI-X message. */
118 u_int msi_vector:8; /* IDT vector. */
119 u_int msi_cpu; /* Local APIC ID. (g) */
120 u_int msi_count:8; /* Messages in this group. (g) */
121 u_int msi_maxcount:8; /* Alignment for this group. (g) */
122 u_int *msi_irqs; /* Group's IRQ list. (g) */
123 u_int msi_remap_cookie;
124 };
125
126 static void msi_create_source(void);
127 static void msi_enable_source(struct intsrc *isrc);
128 static void msi_disable_source(struct intsrc *isrc, int eoi);
129 static void msi_eoi_source(struct intsrc *isrc);
130 static void msi_enable_intr(struct intsrc *isrc);
131 static void msi_disable_intr(struct intsrc *isrc);
132 static int msi_vector(struct intsrc *isrc);
133 static int msi_source_pending(struct intsrc *isrc);
134 static int msi_config_intr(struct intsrc *isrc, enum intr_trigger trig,
135 enum intr_polarity pol);
136 static int msi_assign_cpu(struct intsrc *isrc, u_int apic_id);
137
138 struct pic msi_pic = {
139 .pic_enable_source = msi_enable_source,
140 .pic_disable_source = msi_disable_source,
141 .pic_eoi_source = msi_eoi_source,
142 .pic_enable_intr = msi_enable_intr,
143 .pic_disable_intr = msi_disable_intr,
144 .pic_vector = msi_vector,
145 .pic_source_pending = msi_source_pending,
146 .pic_suspend = NULL,
147 .pic_resume = NULL,
148 .pic_config_intr = msi_config_intr,
149 .pic_assign_cpu = msi_assign_cpu,
150 .pic_reprogram_pin = NULL,
151 };
152
153 u_int first_msi_irq;
154
155 u_int num_msi_irqs = 512;
156 SYSCTL_UINT(_machdep, OID_AUTO, num_msi_irqs, CTLFLAG_RDTUN, &num_msi_irqs, 0,
157 "Number of IRQs reserved for MSI and MSI-X interrupts");
158
159 #ifdef SMP
160 /**
161 * Xen hypervisors prior to 4.6.0 do not properly handle updates to
162 * enabled MSI-X table entries. Allow migration of MSI-X interrupts
163 * to be disabled via a tunable. Values have the following meaning:
164 *
165 * -1: automatic detection by FreeBSD
166 * 0: enable migration
167 * 1: disable migration
168 */
169 int msix_disable_migration = -1;
170 SYSCTL_INT(_machdep, OID_AUTO, disable_msix_migration, CTLFLAG_RDTUN,
171 &msix_disable_migration, 0,
172 "Disable migration of MSI-X interrupts between CPUs");
173 #endif
174
175 static int msi_enabled;
176 static u_int msi_last_irq;
177 static struct mtx msi_lock;
178
179 static void
180 msi_enable_source(struct intsrc *isrc)
181 {
182 }
183
184 static void
185 msi_disable_source(struct intsrc *isrc, int eoi)
186 {
187
188 if (eoi == PIC_EOI)
189 lapic_eoi();
190 }
191
192 static void
193 msi_eoi_source(struct intsrc *isrc)
194 {
195
196 lapic_eoi();
197 }
198
199 static void
200 msi_enable_intr(struct intsrc *isrc)
201 {
202 struct msi_intsrc *msi = (struct msi_intsrc *)isrc;
203
204 apic_enable_vector(msi->msi_cpu, msi->msi_vector);
205 }
206
207 static void
208 msi_disable_intr(struct intsrc *isrc)
209 {
210 struct msi_intsrc *msi = (struct msi_intsrc *)isrc;
211
212 apic_disable_vector(msi->msi_cpu, msi->msi_vector);
213 }
214
215 static int
216 msi_vector(struct intsrc *isrc)
217 {
218 struct msi_intsrc *msi = (struct msi_intsrc *)isrc;
219
220 return (msi->msi_irq);
221 }
222
223 static int
224 msi_source_pending(struct intsrc *isrc)
225 {
226
227 return (0);
228 }
229
230 static int
231 msi_config_intr(struct intsrc *isrc, enum intr_trigger trig,
232 enum intr_polarity pol)
233 {
234
235 return (ENODEV);
236 }
237
238 static int
239 msi_assign_cpu(struct intsrc *isrc, u_int apic_id)
240 {
241 struct msi_intsrc *sib, *msi = (struct msi_intsrc *)isrc;
242 int old_vector;
243 u_int old_id;
244 int i, vector;
245
246 /*
247 * Only allow CPUs to be assigned to the first message for an
248 * MSI group.
249 */
250 if (msi->msi_first != msi)
251 return (EINVAL);
252
253 #ifdef SMP
254 if (msix_disable_migration && msi->msi_msix)
255 return (EINVAL);
256 #endif
257
258 /* Store information to free existing irq. */
259 old_vector = msi->msi_vector;
260 old_id = msi->msi_cpu;
261 if (old_id == apic_id)
262 return (0);
263
264 /* Allocate IDT vectors on this cpu. */
265 if (msi->msi_count > 1) {
266 KASSERT(msi->msi_msix == 0, ("MSI-X message group"));
267 vector = apic_alloc_vectors(apic_id, msi->msi_irqs,
268 msi->msi_count, msi->msi_maxcount);
269 } else
270 vector = apic_alloc_vector(apic_id, msi->msi_irq);
271 if (vector == 0)
272 return (ENOSPC);
273
274 msi->msi_cpu = apic_id;
275 msi->msi_vector = vector;
276 if (msi->msi_intsrc.is_handlers > 0)
277 apic_enable_vector(msi->msi_cpu, msi->msi_vector);
278 if (bootverbose)
279 printf("msi: Assigning %s IRQ %d to local APIC %u vector %u\n",
280 msi->msi_msix ? "MSI-X" : "MSI", msi->msi_irq,
281 msi->msi_cpu, msi->msi_vector);
282 for (i = 1; i < msi->msi_count; i++) {
283 sib = (struct msi_intsrc *)intr_lookup_source(msi->msi_irqs[i]);
284 sib->msi_cpu = apic_id;
285 sib->msi_vector = vector + i;
286 if (sib->msi_intsrc.is_handlers > 0)
287 apic_enable_vector(sib->msi_cpu, sib->msi_vector);
288 if (bootverbose)
289 printf(
290 "msi: Assigning MSI IRQ %d to local APIC %u vector %u\n",
291 sib->msi_irq, sib->msi_cpu, sib->msi_vector);
292 }
293 BUS_REMAP_INTR(device_get_parent(msi->msi_dev), msi->msi_dev,
294 msi->msi_irq);
295
296 /*
297 * Free the old vector after the new one is established. This is done
298 * to prevent races where we could miss an interrupt.
299 */
300 if (msi->msi_intsrc.is_handlers > 0)
301 apic_disable_vector(old_id, old_vector);
302 apic_free_vector(old_id, old_vector, msi->msi_irq);
303 for (i = 1; i < msi->msi_count; i++) {
304 sib = (struct msi_intsrc *)intr_lookup_source(msi->msi_irqs[i]);
305 if (sib->msi_intsrc.is_handlers > 0)
306 apic_disable_vector(old_id, old_vector + i);
307 apic_free_vector(old_id, old_vector + i, msi->msi_irqs[i]);
308 }
309 return (0);
310 }
311
312 void
313 msi_init(void)
314 {
315
316 /* Check if we have a supported CPU. */
317 switch (cpu_vendor_id) {
318 case CPU_VENDOR_INTEL:
319 case CPU_VENDOR_AMD:
320 break;
321 case CPU_VENDOR_CENTAUR:
322 if (CPUID_TO_FAMILY(cpu_id) == 0x6 &&
323 CPUID_TO_MODEL(cpu_id) >= 0xf)
324 break;
325 /* FALLTHROUGH */
326 default:
327 return;
328 }
329
330 #ifdef SMP
331 if (msix_disable_migration == -1) {
332 /* The default is to allow migration of MSI-X interrupts. */
333 msix_disable_migration = 0;
334 }
335 #endif
336
337 if (num_msi_irqs == 0)
338 return;
339
340 first_msi_irq = max(MINIMUM_MSI_INT, num_io_irqs);
341 if (num_msi_irqs > UINT_MAX - first_msi_irq)
342 panic("num_msi_irqs too high");
343 num_io_irqs = first_msi_irq + num_msi_irqs;
344
345 msi_enabled = 1;
346 intr_register_pic(&msi_pic);
347 mtx_init(&msi_lock, "msi", NULL, MTX_DEF);
348 }
349
350 static void
351 msi_create_source(void)
352 {
353 struct msi_intsrc *msi;
354 u_int irq;
355
356 mtx_lock(&msi_lock);
357 if (msi_last_irq >= num_msi_irqs) {
358 mtx_unlock(&msi_lock);
359 return;
360 }
361 irq = msi_last_irq + first_msi_irq;
362 msi_last_irq++;
363 mtx_unlock(&msi_lock);
364
365 msi = malloc(sizeof(struct msi_intsrc), M_MSI, M_WAITOK | M_ZERO);
366 msi->msi_intsrc.is_pic = &msi_pic;
367 msi->msi_irq = irq;
368 intr_register_source(&msi->msi_intsrc);
369 nexus_add_irq(irq);
370 }
371
372 /*
373 * Try to allocate 'count' interrupt sources with contiguous IDT values.
374 */
375 int
376 msi_alloc(device_t dev, int count, int maxcount, int *irqs)
377 {
378 struct msi_intsrc *msi, *fsrc;
379 u_int cpu, *mirqs;
380 int cnt, i, vector;
381 #ifdef ACPI_DMAR
382 u_int cookies[count];
383 int error;
384 #endif
385
386 if (!msi_enabled)
387 return (ENXIO);
388
389 if (count > 1)
390 mirqs = malloc(count * sizeof(*mirqs), M_MSI, M_WAITOK);
391 else
392 mirqs = NULL;
393 again:
394 mtx_lock(&msi_lock);
395
396 /* Try to find 'count' free IRQs. */
397 cnt = 0;
398 for (i = first_msi_irq; i < first_msi_irq + num_msi_irqs; i++) {
399 msi = (struct msi_intsrc *)intr_lookup_source(i);
400
401 /* End of allocated sources, so break. */
402 if (msi == NULL)
403 break;
404
405 /* If this is a free one, save its IRQ in the array. */
406 if (msi->msi_dev == NULL) {
407 irqs[cnt] = i;
408 cnt++;
409 if (cnt == count)
410 break;
411 }
412 }
413
414 /* Do we need to create some new sources? */
415 if (cnt < count) {
416 /* If we would exceed the max, give up. */
417 if (i + (count - cnt) > first_msi_irq + num_msi_irqs) {
418 mtx_unlock(&msi_lock);
419 free(mirqs, M_MSI);
420 return (ENXIO);
421 }
422 mtx_unlock(&msi_lock);
423
424 /* We need count - cnt more sources. */
425 while (cnt < count) {
426 msi_create_source();
427 cnt++;
428 }
429 goto again;
430 }
431
432 /* Ok, we now have the IRQs allocated. */
433 KASSERT(cnt == count, ("count mismatch"));
434
435 /* Allocate 'count' IDT vectors. */
436 cpu = intr_next_cpu();
437 vector = apic_alloc_vectors(cpu, irqs, count, maxcount);
438 if (vector == 0) {
439 mtx_unlock(&msi_lock);
440 free(mirqs, M_MSI);
441 return (ENOSPC);
442 }
443
444 #ifdef ACPI_DMAR
445 mtx_unlock(&msi_lock);
446 error = iommu_alloc_msi_intr(dev, cookies, count);
447 mtx_lock(&msi_lock);
448 if (error == EOPNOTSUPP)
449 error = 0;
450 if (error != 0) {
451 for (i = 0; i < count; i++)
452 apic_free_vector(cpu, vector + i, irqs[i]);
453 free(mirqs, M_MSI);
454 return (error);
455 }
456 for (i = 0; i < count; i++) {
457 msi = (struct msi_intsrc *)intr_lookup_source(irqs[i]);
458 msi->msi_remap_cookie = cookies[i];
459 }
460 #endif
461
462 /* Assign IDT vectors and make these messages owned by 'dev'. */
463 fsrc = (struct msi_intsrc *)intr_lookup_source(irqs[0]);
464 for (i = 0; i < count; i++) {
465 msi = (struct msi_intsrc *)intr_lookup_source(irqs[i]);
466 msi->msi_cpu = cpu;
467 msi->msi_dev = dev;
468 msi->msi_vector = vector + i;
469 if (bootverbose)
470 printf(
471 "msi: routing MSI IRQ %d to local APIC %u vector %u\n",
472 msi->msi_irq, msi->msi_cpu, msi->msi_vector);
473 msi->msi_first = fsrc;
474 KASSERT(msi->msi_intsrc.is_handlers == 0,
475 ("dead MSI has handlers"));
476 }
477 fsrc->msi_count = count;
478 fsrc->msi_maxcount = maxcount;
479 if (count > 1)
480 bcopy(irqs, mirqs, count * sizeof(*mirqs));
481 fsrc->msi_irqs = mirqs;
482 mtx_unlock(&msi_lock);
483 return (0);
484 }
485
486 int
487 msi_release(int *irqs, int count)
488 {
489 struct msi_intsrc *msi, *first;
490 int i;
491
492 mtx_lock(&msi_lock);
493 first = (struct msi_intsrc *)intr_lookup_source(irqs[0]);
494 if (first == NULL) {
495 mtx_unlock(&msi_lock);
496 return (ENOENT);
497 }
498
499 /* Make sure this isn't an MSI-X message. */
500 if (first->msi_msix) {
501 mtx_unlock(&msi_lock);
502 return (EINVAL);
503 }
504
505 /* Make sure this message is allocated to a group. */
506 if (first->msi_first == NULL) {
507 mtx_unlock(&msi_lock);
508 return (ENXIO);
509 }
510
511 /*
512 * Make sure this is the start of a group and that we are releasing
513 * the entire group.
514 */
515 if (first->msi_first != first || first->msi_count != count) {
516 mtx_unlock(&msi_lock);
517 return (EINVAL);
518 }
519 KASSERT(first->msi_dev != NULL, ("unowned group"));
520
521 /* Clear all the extra messages in the group. */
522 for (i = 1; i < count; i++) {
523 msi = (struct msi_intsrc *)intr_lookup_source(irqs[i]);
524 KASSERT(msi->msi_first == first, ("message not in group"));
525 KASSERT(msi->msi_dev == first->msi_dev, ("owner mismatch"));
526 #ifdef ACPI_DMAR
527 iommu_unmap_msi_intr(first->msi_dev, msi->msi_remap_cookie);
528 #endif
529 msi->msi_first = NULL;
530 msi->msi_dev = NULL;
531 apic_free_vector(msi->msi_cpu, msi->msi_vector, msi->msi_irq);
532 msi->msi_vector = 0;
533 }
534
535 /* Clear out the first message. */
536 #ifdef ACPI_DMAR
537 mtx_unlock(&msi_lock);
538 iommu_unmap_msi_intr(first->msi_dev, first->msi_remap_cookie);
539 mtx_lock(&msi_lock);
540 #endif
541 first->msi_first = NULL;
542 first->msi_dev = NULL;
543 apic_free_vector(first->msi_cpu, first->msi_vector, first->msi_irq);
544 first->msi_vector = 0;
545 first->msi_count = 0;
546 first->msi_maxcount = 0;
547 free(first->msi_irqs, M_MSI);
548 first->msi_irqs = NULL;
549
550 mtx_unlock(&msi_lock);
551 return (0);
552 }
553
554 int
555 msi_map(int irq, uint64_t *addr, uint32_t *data)
556 {
557 struct msi_intsrc *msi;
558 int error;
559 #ifdef ACPI_DMAR
560 struct msi_intsrc *msi1;
561 int i, k;
562 #endif
563
564 mtx_lock(&msi_lock);
565 msi = (struct msi_intsrc *)intr_lookup_source(irq);
566 if (msi == NULL) {
567 mtx_unlock(&msi_lock);
568 return (ENOENT);
569 }
570
571 /* Make sure this message is allocated to a device. */
572 if (msi->msi_dev == NULL) {
573 mtx_unlock(&msi_lock);
574 return (ENXIO);
575 }
576
577 /*
578 * If this message isn't an MSI-X message, make sure it's part
579 * of a group, and switch to the first message in the
580 * group.
581 */
582 if (!msi->msi_msix) {
583 if (msi->msi_first == NULL) {
584 mtx_unlock(&msi_lock);
585 return (ENXIO);
586 }
587 msi = msi->msi_first;
588 }
589
590 #ifdef ACPI_DMAR
591 if (!msi->msi_msix) {
592 for (k = msi->msi_count - 1, i = first_msi_irq; k > 0 &&
593 i < first_msi_irq + num_msi_irqs; i++) {
594 if (i == msi->msi_irq)
595 continue;
596 msi1 = (struct msi_intsrc *)intr_lookup_source(i);
597 if (!msi1->msi_msix && msi1->msi_first == msi) {
598 mtx_unlock(&msi_lock);
599 iommu_map_msi_intr(msi1->msi_dev,
600 msi1->msi_cpu, msi1->msi_vector,
601 msi1->msi_remap_cookie, NULL, NULL);
602 k--;
603 mtx_lock(&msi_lock);
604 }
605 }
606 }
607 mtx_unlock(&msi_lock);
608 error = iommu_map_msi_intr(msi->msi_dev, msi->msi_cpu,
609 msi->msi_vector, msi->msi_remap_cookie, addr, data);
610 #else
611 mtx_unlock(&msi_lock);
612 error = EOPNOTSUPP;
613 #endif
614 if (error == EOPNOTSUPP) {
615 *addr = INTEL_ADDR(msi);
616 *data = INTEL_DATA(msi);
617 error = 0;
618 }
619 return (error);
620 }
621
622 int
623 msix_alloc(device_t dev, int *irq)
624 {
625 struct msi_intsrc *msi;
626 u_int cpu;
627 int i, vector;
628 #ifdef ACPI_DMAR
629 u_int cookie;
630 int error;
631 #endif
632
633 if (!msi_enabled)
634 return (ENXIO);
635
636 again:
637 mtx_lock(&msi_lock);
638
639 /* Find a free IRQ. */
640 for (i = first_msi_irq; i < first_msi_irq + num_msi_irqs; i++) {
641 msi = (struct msi_intsrc *)intr_lookup_source(i);
642
643 /* End of allocated sources, so break. */
644 if (msi == NULL)
645 break;
646
647 /* Stop at the first free source. */
648 if (msi->msi_dev == NULL)
649 break;
650 }
651
652 /* Are all IRQs in use? */
653 if (i == first_msi_irq + num_msi_irqs) {
654 mtx_unlock(&msi_lock);
655 return (ENXIO);
656 }
657
658 /* Do we need to create a new source? */
659 if (msi == NULL) {
660 mtx_unlock(&msi_lock);
661
662 /* Create a new source. */
663 msi_create_source();
664 goto again;
665 }
666
667 /* Allocate an IDT vector. */
668 cpu = intr_next_cpu();
669 vector = apic_alloc_vector(cpu, i);
670 if (vector == 0) {
671 mtx_unlock(&msi_lock);
672 return (ENOSPC);
673 }
674
675 msi->msi_dev = dev;
676 #ifdef ACPI_DMAR
677 mtx_unlock(&msi_lock);
678 error = iommu_alloc_msi_intr(dev, &cookie, 1);
679 mtx_lock(&msi_lock);
680 if (error == EOPNOTSUPP)
681 error = 0;
682 if (error != 0) {
683 msi->msi_dev = NULL;
684 apic_free_vector(cpu, vector, i);
685 return (error);
686 }
687 msi->msi_remap_cookie = cookie;
688 #endif
689
690 if (bootverbose)
691 printf("msi: routing MSI-X IRQ %d to local APIC %u vector %u\n",
692 msi->msi_irq, cpu, vector);
693
694 /* Setup source. */
695 msi->msi_cpu = cpu;
696 msi->msi_first = msi;
697 msi->msi_vector = vector;
698 msi->msi_msix = 1;
699 msi->msi_count = 1;
700 msi->msi_maxcount = 1;
701 msi->msi_irqs = NULL;
702
703 KASSERT(msi->msi_intsrc.is_handlers == 0, ("dead MSI-X has handlers"));
704 mtx_unlock(&msi_lock);
705
706 *irq = i;
707 return (0);
708 }
709
710 int
711 msix_release(int irq)
712 {
713 struct msi_intsrc *msi;
714
715 mtx_lock(&msi_lock);
716 msi = (struct msi_intsrc *)intr_lookup_source(irq);
717 if (msi == NULL) {
718 mtx_unlock(&msi_lock);
719 return (ENOENT);
720 }
721
722 /* Make sure this is an MSI-X message. */
723 if (!msi->msi_msix) {
724 mtx_unlock(&msi_lock);
725 return (EINVAL);
726 }
727
728 KASSERT(msi->msi_dev != NULL, ("unowned message"));
729
730 /* Clear out the message. */
731 #ifdef ACPI_DMAR
732 mtx_unlock(&msi_lock);
733 iommu_unmap_msi_intr(msi->msi_dev, msi->msi_remap_cookie);
734 mtx_lock(&msi_lock);
735 #endif
736 msi->msi_first = NULL;
737 msi->msi_dev = NULL;
738 apic_free_vector(msi->msi_cpu, msi->msi_vector, msi->msi_irq);
739 msi->msi_vector = 0;
740 msi->msi_msix = 0;
741 msi->msi_count = 0;
742 msi->msi_maxcount = 0;
743
744 mtx_unlock(&msi_lock);
745 return (0);
746 }
Cache object: e24accc3004699d65bc9c827631632f1
|