]> Git Repo - qemu.git/blame - hw/xics.c
pseries: Rework implementation of TCE bypass
[qemu.git] / hw / xics.c
CommitLineData
b5cec4c5
DG
1/*
2 * QEMU PowerPC pSeries Logical Partition (aka sPAPR) hardware System Emulator
3 *
4 * PAPR Virtualized Interrupt System, aka ICS/ICP aka xics
5 *
6 * Copyright (c) 2010,2011 David Gibson, IBM Corporation.
7 *
8 * Permission is hereby granted, free of charge, to any person obtaining a copy
9 * of this software and associated documentation files (the "Software"), to deal
10 * in the Software without restriction, including without limitation the rights
11 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
12 * copies of the Software, and to permit persons to whom the Software is
13 * furnished to do so, subject to the following conditions:
14 *
15 * The above copyright notice and this permission notice shall be included in
16 * all copies or substantial portions of the Software.
17 *
18 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
19 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
20 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
21 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
22 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
23 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
24 * THE SOFTWARE.
25 *
26 */
27
28#include "hw.h"
29#include "hw/spapr.h"
30#include "hw/xics.h"
31
b5cec4c5
DG
32/*
33 * ICP: Presentation layer
34 */
35
36struct icp_server_state {
37 uint32_t xirr;
38 uint8_t pending_priority;
39 uint8_t mfrr;
40 qemu_irq output;
41};
42
43#define XISR_MASK 0x00ffffff
44#define CPPR_MASK 0xff000000
45
46#define XISR(ss) (((ss)->xirr) & XISR_MASK)
47#define CPPR(ss) (((ss)->xirr) >> 24)
48
49struct ics_state;
50
51struct icp_state {
52 long nr_servers;
53 struct icp_server_state *ss;
54 struct ics_state *ics;
55};
56
57static void ics_reject(struct ics_state *ics, int nr);
58static void ics_resend(struct ics_state *ics);
59static void ics_eoi(struct ics_state *ics, int nr);
60
61static void icp_check_ipi(struct icp_state *icp, int server)
62{
63 struct icp_server_state *ss = icp->ss + server;
64
65 if (XISR(ss) && (ss->pending_priority <= ss->mfrr)) {
66 return;
67 }
68
69 if (XISR(ss)) {
70 ics_reject(icp->ics, XISR(ss));
71 }
72
73 ss->xirr = (ss->xirr & ~XISR_MASK) | XICS_IPI;
74 ss->pending_priority = ss->mfrr;
75 qemu_irq_raise(ss->output);
76}
77
78static void icp_resend(struct icp_state *icp, int server)
79{
80 struct icp_server_state *ss = icp->ss + server;
81
82 if (ss->mfrr < CPPR(ss)) {
83 icp_check_ipi(icp, server);
84 }
85 ics_resend(icp->ics);
86}
87
88static void icp_set_cppr(struct icp_state *icp, int server, uint8_t cppr)
89{
90 struct icp_server_state *ss = icp->ss + server;
91 uint8_t old_cppr;
92 uint32_t old_xisr;
93
94 old_cppr = CPPR(ss);
95 ss->xirr = (ss->xirr & ~CPPR_MASK) | (cppr << 24);
96
97 if (cppr < old_cppr) {
98 if (XISR(ss) && (cppr <= ss->pending_priority)) {
99 old_xisr = XISR(ss);
100 ss->xirr &= ~XISR_MASK; /* Clear XISR */
101 qemu_irq_lower(ss->output);
102 ics_reject(icp->ics, old_xisr);
103 }
104 } else {
105 if (!XISR(ss)) {
106 icp_resend(icp, server);
107 }
108 }
109}
110
111static void icp_set_mfrr(struct icp_state *icp, int nr, uint8_t mfrr)
112{
113 struct icp_server_state *ss = icp->ss + nr;
114
115 ss->mfrr = mfrr;
116 if (mfrr < CPPR(ss)) {
117 icp_check_ipi(icp, nr);
118 }
119}
120
121static uint32_t icp_accept(struct icp_server_state *ss)
122{
123 uint32_t xirr;
124
125 qemu_irq_lower(ss->output);
126 xirr = ss->xirr;
127 ss->xirr = ss->pending_priority << 24;
128 return xirr;
129}
130
131static void icp_eoi(struct icp_state *icp, int server, uint32_t xirr)
132{
133 struct icp_server_state *ss = icp->ss + server;
134
b5cec4c5
DG
135 /* Send EOI -> ICS */
136 ss->xirr = (ss->xirr & ~CPPR_MASK) | (xirr & CPPR_MASK);
d07fee7e 137 ics_eoi(icp->ics, xirr & XISR_MASK);
b5cec4c5
DG
138 if (!XISR(ss)) {
139 icp_resend(icp, server);
140 }
141}
142
143static void icp_irq(struct icp_state *icp, int server, int nr, uint8_t priority)
144{
145 struct icp_server_state *ss = icp->ss + server;
146
147 if ((priority >= CPPR(ss))
148 || (XISR(ss) && (ss->pending_priority <= priority))) {
149 ics_reject(icp->ics, nr);
150 } else {
151 if (XISR(ss)) {
152 ics_reject(icp->ics, XISR(ss));
153 }
154 ss->xirr = (ss->xirr & ~XISR_MASK) | (nr & XISR_MASK);
155 ss->pending_priority = priority;
156 qemu_irq_raise(ss->output);
157 }
158}
159
160/*
161 * ICS: Source layer
162 */
163
164struct ics_irq_state {
165 int server;
166 uint8_t priority;
167 uint8_t saved_priority;
98ca8c02
DG
168#define XICS_STATUS_ASSERTED 0x1
169#define XICS_STATUS_SENT 0x2
170#define XICS_STATUS_REJECTED 0x4
171#define XICS_STATUS_MASKED_PENDING 0x8
172 uint8_t status;
ff9d2afa 173 bool lsi;
b5cec4c5
DG
174};
175
176struct ics_state {
177 int nr_irqs;
178 int offset;
179 qemu_irq *qirqs;
180 struct ics_irq_state *irqs;
181 struct icp_state *icp;
182};
183
184static int ics_valid_irq(struct ics_state *ics, uint32_t nr)
185{
186 return (nr >= ics->offset)
187 && (nr < (ics->offset + ics->nr_irqs));
188}
189
d07fee7e
DG
190static void resend_msi(struct ics_state *ics, int srcno)
191{
192 struct ics_irq_state *irq = ics->irqs + srcno;
193
194 /* FIXME: filter by server#? */
98ca8c02
DG
195 if (irq->status & XICS_STATUS_REJECTED) {
196 irq->status &= ~XICS_STATUS_REJECTED;
d07fee7e
DG
197 if (irq->priority != 0xff) {
198 icp_irq(ics->icp, irq->server, srcno + ics->offset,
199 irq->priority);
200 }
201 }
202}
203
204static void resend_lsi(struct ics_state *ics, int srcno)
205{
206 struct ics_irq_state *irq = ics->irqs + srcno;
207
98ca8c02
DG
208 if ((irq->priority != 0xff)
209 && (irq->status & XICS_STATUS_ASSERTED)
210 && !(irq->status & XICS_STATUS_SENT)) {
211 irq->status |= XICS_STATUS_SENT;
d07fee7e
DG
212 icp_irq(ics->icp, irq->server, srcno + ics->offset, irq->priority);
213 }
214}
215
216static void set_irq_msi(struct ics_state *ics, int srcno, int val)
b5cec4c5 217{
cc67b9c8 218 struct ics_irq_state *irq = ics->irqs + srcno;
b5cec4c5
DG
219
220 if (val) {
221 if (irq->priority == 0xff) {
98ca8c02 222 irq->status |= XICS_STATUS_MASKED_PENDING;
b5cec4c5
DG
223 /* masked pending */ ;
224 } else {
cc67b9c8 225 icp_irq(ics->icp, irq->server, srcno + ics->offset, irq->priority);
b5cec4c5
DG
226 }
227 }
228}
229
d07fee7e 230static void set_irq_lsi(struct ics_state *ics, int srcno, int val)
b5cec4c5 231{
d07fee7e 232 struct ics_irq_state *irq = ics->irqs + srcno;
b5cec4c5 233
98ca8c02
DG
234 if (val) {
235 irq->status |= XICS_STATUS_ASSERTED;
236 } else {
237 irq->status &= ~XICS_STATUS_ASSERTED;
238 }
d07fee7e 239 resend_lsi(ics, srcno);
b5cec4c5
DG
240}
241
d07fee7e 242static void ics_set_irq(void *opaque, int srcno, int val)
b5cec4c5 243{
d07fee7e
DG
244 struct ics_state *ics = (struct ics_state *)opaque;
245 struct ics_irq_state *irq = ics->irqs + srcno;
b5cec4c5 246
ff9d2afa 247 if (irq->lsi) {
d07fee7e
DG
248 set_irq_lsi(ics, srcno, val);
249 } else {
250 set_irq_msi(ics, srcno, val);
251 }
252}
b5cec4c5 253
d07fee7e
DG
254static void write_xive_msi(struct ics_state *ics, int srcno)
255{
256 struct ics_irq_state *irq = ics->irqs + srcno;
257
98ca8c02
DG
258 if (!(irq->status & XICS_STATUS_MASKED_PENDING)
259 || (irq->priority == 0xff)) {
d07fee7e 260 return;
b5cec4c5 261 }
d07fee7e 262
98ca8c02 263 irq->status &= ~XICS_STATUS_MASKED_PENDING;
d07fee7e 264 icp_irq(ics->icp, irq->server, srcno + ics->offset, irq->priority);
b5cec4c5
DG
265}
266
d07fee7e 267static void write_xive_lsi(struct ics_state *ics, int srcno)
b5cec4c5 268{
d07fee7e
DG
269 resend_lsi(ics, srcno);
270}
271
272static void ics_write_xive(struct ics_state *ics, int nr, int server,
273 uint8_t priority)
274{
275 int srcno = nr - ics->offset;
276 struct ics_irq_state *irq = ics->irqs + srcno;
b5cec4c5
DG
277
278 irq->server = server;
279 irq->priority = priority;
280
ff9d2afa 281 if (irq->lsi) {
d07fee7e
DG
282 write_xive_lsi(ics, srcno);
283 } else {
284 write_xive_msi(ics, srcno);
b5cec4c5 285 }
b5cec4c5
DG
286}
287
288static void ics_reject(struct ics_state *ics, int nr)
289{
d07fee7e
DG
290 struct ics_irq_state *irq = ics->irqs + nr - ics->offset;
291
98ca8c02
DG
292 irq->status |= XICS_STATUS_REJECTED; /* Irrelevant but harmless for LSI */
293 irq->status &= ~XICS_STATUS_SENT; /* Irrelevant but harmless for MSI */
b5cec4c5
DG
294}
295
296static void ics_resend(struct ics_state *ics)
297{
d07fee7e
DG
298 int i;
299
300 for (i = 0; i < ics->nr_irqs; i++) {
301 struct ics_irq_state *irq = ics->irqs + i;
302
303 /* FIXME: filter by server#? */
ff9d2afa 304 if (irq->lsi) {
d07fee7e
DG
305 resend_lsi(ics, i);
306 } else {
307 resend_msi(ics, i);
308 }
309 }
b5cec4c5
DG
310}
311
312static void ics_eoi(struct ics_state *ics, int nr)
313{
d07fee7e
DG
314 int srcno = nr - ics->offset;
315 struct ics_irq_state *irq = ics->irqs + srcno;
316
ff9d2afa 317 if (irq->lsi) {
98ca8c02 318 irq->status &= ~XICS_STATUS_SENT;
d07fee7e 319 }
b5cec4c5
DG
320}
321
322/*
323 * Exported functions
324 */
325
a307d594 326qemu_irq xics_get_qirq(struct icp_state *icp, int irq)
b5cec4c5
DG
327{
328 if ((irq < icp->ics->offset)
329 || (irq >= (icp->ics->offset + icp->ics->nr_irqs))) {
330 return NULL;
331 }
332
a307d594
AK
333 return icp->ics->qirqs[irq - icp->ics->offset];
334}
335
ff9d2afa 336void xics_set_irq_type(struct icp_state *icp, int irq, bool lsi)
a307d594
AK
337{
338 assert((irq >= icp->ics->offset)
339 && (irq < (icp->ics->offset + icp->ics->nr_irqs)));
d07fee7e 340
ff9d2afa 341 icp->ics->irqs[irq - icp->ics->offset].lsi = lsi;
b5cec4c5
DG
342}
343
e2684c0b 344static target_ulong h_cppr(CPUPPCState *env, sPAPREnvironment *spapr,
b5cec4c5
DG
345 target_ulong opcode, target_ulong *args)
346{
347 target_ulong cppr = args[0];
348
349 icp_set_cppr(spapr->icp, env->cpu_index, cppr);
350 return H_SUCCESS;
351}
352
e2684c0b 353static target_ulong h_ipi(CPUPPCState *env, sPAPREnvironment *spapr,
b5cec4c5
DG
354 target_ulong opcode, target_ulong *args)
355{
356 target_ulong server = args[0];
357 target_ulong mfrr = args[1];
358
359 if (server >= spapr->icp->nr_servers) {
360 return H_PARAMETER;
361 }
362
363 icp_set_mfrr(spapr->icp, server, mfrr);
364 return H_SUCCESS;
365
366}
367
e2684c0b 368static target_ulong h_xirr(CPUPPCState *env, sPAPREnvironment *spapr,
b5cec4c5
DG
369 target_ulong opcode, target_ulong *args)
370{
371 uint32_t xirr = icp_accept(spapr->icp->ss + env->cpu_index);
372
373 args[0] = xirr;
374 return H_SUCCESS;
375}
376
e2684c0b 377static target_ulong h_eoi(CPUPPCState *env, sPAPREnvironment *spapr,
b5cec4c5
DG
378 target_ulong opcode, target_ulong *args)
379{
380 target_ulong xirr = args[0];
381
382 icp_eoi(spapr->icp, env->cpu_index, xirr);
383 return H_SUCCESS;
384}
385
386static void rtas_set_xive(sPAPREnvironment *spapr, uint32_t token,
387 uint32_t nargs, target_ulong args,
388 uint32_t nret, target_ulong rets)
389{
390 struct ics_state *ics = spapr->icp->ics;
391 uint32_t nr, server, priority;
392
393 if ((nargs != 3) || (nret != 1)) {
394 rtas_st(rets, 0, -3);
395 return;
396 }
397
398 nr = rtas_ld(args, 0);
399 server = rtas_ld(args, 1);
400 priority = rtas_ld(args, 2);
401
402 if (!ics_valid_irq(ics, nr) || (server >= ics->icp->nr_servers)
403 || (priority > 0xff)) {
404 rtas_st(rets, 0, -3);
405 return;
406 }
407
d07fee7e 408 ics_write_xive(ics, nr, server, priority);
b5cec4c5
DG
409
410 rtas_st(rets, 0, 0); /* Success */
411}
412
413static void rtas_get_xive(sPAPREnvironment *spapr, uint32_t token,
414 uint32_t nargs, target_ulong args,
415 uint32_t nret, target_ulong rets)
416{
417 struct ics_state *ics = spapr->icp->ics;
418 uint32_t nr;
419
420 if ((nargs != 1) || (nret != 3)) {
421 rtas_st(rets, 0, -3);
422 return;
423 }
424
425 nr = rtas_ld(args, 0);
426
427 if (!ics_valid_irq(ics, nr)) {
428 rtas_st(rets, 0, -3);
429 return;
430 }
431
432 rtas_st(rets, 0, 0); /* Success */
433 rtas_st(rets, 1, ics->irqs[nr - ics->offset].server);
434 rtas_st(rets, 2, ics->irqs[nr - ics->offset].priority);
435}
436
437static void rtas_int_off(sPAPREnvironment *spapr, uint32_t token,
438 uint32_t nargs, target_ulong args,
439 uint32_t nret, target_ulong rets)
440{
441 struct ics_state *ics = spapr->icp->ics;
442 uint32_t nr;
443
444 if ((nargs != 1) || (nret != 1)) {
445 rtas_st(rets, 0, -3);
446 return;
447 }
448
449 nr = rtas_ld(args, 0);
450
451 if (!ics_valid_irq(ics, nr)) {
452 rtas_st(rets, 0, -3);
453 return;
454 }
455
456 /* This is a NOP for now, since the described PAPR semantics don't
457 * seem to gel with what Linux does */
458#if 0
459 struct ics_irq_state *irq = xics->irqs + (nr - xics->offset);
460
461 irq->saved_priority = irq->priority;
cc67b9c8 462 ics_write_xive_msi(xics, nr, irq->server, 0xff);
b5cec4c5
DG
463#endif
464
465 rtas_st(rets, 0, 0); /* Success */
466}
467
468static void rtas_int_on(sPAPREnvironment *spapr, uint32_t token,
469 uint32_t nargs, target_ulong args,
470 uint32_t nret, target_ulong rets)
471{
472 struct ics_state *ics = spapr->icp->ics;
473 uint32_t nr;
474
475 if ((nargs != 1) || (nret != 1)) {
476 rtas_st(rets, 0, -3);
477 return;
478 }
479
480 nr = rtas_ld(args, 0);
481
482 if (!ics_valid_irq(ics, nr)) {
483 rtas_st(rets, 0, -3);
484 return;
485 }
486
487 /* This is a NOP for now, since the described PAPR semantics don't
488 * seem to gel with what Linux does */
489#if 0
490 struct ics_irq_state *irq = xics->irqs + (nr - xics->offset);
491
cc67b9c8 492 ics_write_xive_msi(xics, nr, irq->server, irq->saved_priority);
b5cec4c5
DG
493#endif
494
495 rtas_st(rets, 0, 0); /* Success */
496}
497
256b408a
DG
498static void xics_reset(void *opaque)
499{
500 struct icp_state *icp = (struct icp_state *)opaque;
501 struct ics_state *ics = icp->ics;
502 int i;
503
504 for (i = 0; i < icp->nr_servers; i++) {
505 icp->ss[i].xirr = 0;
506 icp->ss[i].pending_priority = 0;
507 icp->ss[i].mfrr = 0xff;
508 /* Make all outputs are deasserted */
509 qemu_set_irq(icp->ss[i].output, 0);
510 }
511
512 for (i = 0; i < ics->nr_irqs; i++) {
513 /* Reset everything *except* the type */
514 ics->irqs[i].server = 0;
98ca8c02 515 ics->irqs[i].status = 0;
256b408a
DG
516 ics->irqs[i].priority = 0xff;
517 ics->irqs[i].saved_priority = 0xff;
518 }
519}
520
c7a5c0c9 521struct icp_state *xics_system_init(int nr_irqs)
b5cec4c5 522{
e2684c0b 523 CPUPPCState *env;
c7a5c0c9 524 int max_server_num;
b5cec4c5
DG
525 struct icp_state *icp;
526 struct ics_state *ics;
527
c7a5c0c9
DG
528 max_server_num = -1;
529 for (env = first_cpu; env != NULL; env = env->next_cpu) {
530 if (env->cpu_index > max_server_num) {
531 max_server_num = env->cpu_index;
532 }
533 }
534
7267c094 535 icp = g_malloc0(sizeof(*icp));
c7a5c0c9 536 icp->nr_servers = max_server_num + 1;
7267c094 537 icp->ss = g_malloc0(icp->nr_servers*sizeof(struct icp_server_state));
c7a5c0c9 538
c7a5c0c9
DG
539 for (env = first_cpu; env != NULL; env = env->next_cpu) {
540 struct icp_server_state *ss = &icp->ss[env->cpu_index];
b5cec4c5 541
c7a5c0c9 542 switch (PPC_INPUT(env)) {
b5cec4c5 543 case PPC_FLAGS_INPUT_POWER7:
c7a5c0c9 544 ss->output = env->irq_inputs[POWER7_INPUT_INT];
b5cec4c5
DG
545 break;
546
547 case PPC_FLAGS_INPUT_970:
c7a5c0c9 548 ss->output = env->irq_inputs[PPC970_INPUT_INT];
b5cec4c5
DG
549 break;
550
551 default:
552 hw_error("XICS interrupt model does not support this CPU bus "
553 "model\n");
554 exit(1);
555 }
b5cec4c5
DG
556 }
557
7267c094 558 ics = g_malloc0(sizeof(*ics));
b5cec4c5
DG
559 ics->nr_irqs = nr_irqs;
560 ics->offset = 16;
7267c094 561 ics->irqs = g_malloc0(nr_irqs * sizeof(struct ics_irq_state));
b5cec4c5
DG
562
563 icp->ics = ics;
564 ics->icp = icp;
565
d07fee7e 566 ics->qirqs = qemu_allocate_irqs(ics_set_irq, ics, nr_irqs);
b5cec4c5
DG
567
568 spapr_register_hypercall(H_CPPR, h_cppr);
569 spapr_register_hypercall(H_IPI, h_ipi);
570 spapr_register_hypercall(H_XIRR, h_xirr);
571 spapr_register_hypercall(H_EOI, h_eoi);
572
573 spapr_rtas_register("ibm,set-xive", rtas_set_xive);
574 spapr_rtas_register("ibm,get-xive", rtas_get_xive);
575 spapr_rtas_register("ibm,int-off", rtas_int_off);
576 spapr_rtas_register("ibm,int-on", rtas_int_on);
577
256b408a
DG
578 qemu_register_reset(xics_reset, icp);
579
b5cec4c5
DG
580 return icp;
581}
This page took 0.37314 seconds and 4 git commands to generate.