1 // SPDX-License-Identifier: GPL-2.0
3 * Driver for s390 eadm subchannels
5 * Copyright IBM Corp. 2012
9 #include <linux/kernel_stat.h>
10 #include <linux/completion.h>
11 #include <linux/workqueue.h>
12 #include <linux/spinlock.h>
13 #include <linux/device.h>
14 #include <linux/module.h>
15 #include <linux/timer.h>
16 #include <linux/slab.h>
17 #include <linux/list.h>
20 #include <asm/css_chars.h>
21 #include <asm/debug.h>
33 MODULE_DESCRIPTION("driver for s390 eadm subchannels");
34 MODULE_LICENSE("GPL");
36 #define EADM_TIMEOUT (7 * HZ)
37 static DEFINE_SPINLOCK(list_lock);
38 static LIST_HEAD(eadm_list);
40 static debug_info_t *eadm_debug;
42 #define EADM_LOG(imp, txt) do { \
43 debug_text_event(eadm_debug, imp, txt); \
46 static void EADM_LOG_HEX(int level, void *data, int length)
48 debug_event(eadm_debug, level, data, length);
51 static void orb_init(union orb *orb)
53 memset(orb, 0, sizeof(union orb));
54 orb->eadm.compat1 = 1;
55 orb->eadm.compat2 = 1;
60 static int eadm_subchannel_start(struct subchannel *sch, struct aob *aob)
62 union orb *orb = &get_eadm_private(sch)->orb;
66 orb->eadm.aob = virt_to_dma32(aob);
67 orb->eadm.intparm = (u32)virt_to_phys(sch);
68 orb->eadm.key = PAGE_DEFAULT_KEY >> 4;
71 EADM_LOG_HEX(6, &sch->schid, sizeof(sch->schid));
73 cc = ssch(sch->schid, orb);
76 sch->schib.scsw.eadm.actl |= SCSW_ACTL_START_PEND;
78 case 1: /* status pending */
81 case 3: /* not operational */
87 static int eadm_subchannel_clear(struct subchannel *sch)
91 cc = csch(sch->schid);
95 sch->schib.scsw.eadm.actl |= SCSW_ACTL_CLEAR_PEND;
99 static void eadm_subchannel_timeout(struct timer_list *t)
101 struct eadm_private *private = from_timer(private, t, timer);
102 struct subchannel *sch = private->sch;
104 spin_lock_irq(&sch->lock);
105 EADM_LOG(1, "timeout");
106 EADM_LOG_HEX(1, &sch->schid, sizeof(sch->schid));
107 if (eadm_subchannel_clear(sch))
108 EADM_LOG(0, "clear failed");
109 spin_unlock_irq(&sch->lock);
112 static void eadm_subchannel_set_timeout(struct subchannel *sch, int expires)
114 struct eadm_private *private = get_eadm_private(sch);
117 del_timer(&private->timer);
119 mod_timer(&private->timer, jiffies + expires);
122 static void eadm_subchannel_irq(struct subchannel *sch)
124 struct eadm_private *private = get_eadm_private(sch);
125 struct eadm_scsw *scsw = &sch->schib.scsw.eadm;
126 struct irb *irb = this_cpu_ptr(&cio_irb);
127 blk_status_t error = BLK_STS_OK;
130 EADM_LOG_HEX(6, irb, sizeof(*irb));
132 inc_irq_stat(IRQIO_ADM);
134 if ((scsw->stctl & (SCSW_STCTL_ALERT_STATUS | SCSW_STCTL_STATUS_PEND))
135 && scsw->eswf == 1 && irb->esw.eadm.erw.r)
136 error = BLK_STS_IOERR;
138 if (scsw->fctl & SCSW_FCTL_CLEAR_FUNC)
139 error = BLK_STS_TIMEOUT;
141 eadm_subchannel_set_timeout(sch, 0);
143 if (private->state != EADM_BUSY) {
144 EADM_LOG(1, "irq unsol");
145 EADM_LOG_HEX(1, irb, sizeof(*irb));
146 private->state = EADM_NOT_OPER;
147 css_sched_sch_todo(sch, SCH_TODO_EVAL);
150 scm_irq_handler(dma32_to_virt(scsw->aob), error);
151 private->state = EADM_IDLE;
153 if (private->completion)
154 complete(private->completion);
157 static struct subchannel *eadm_get_idle_sch(void)
159 struct eadm_private *private;
160 struct subchannel *sch;
163 spin_lock_irqsave(&list_lock, flags);
164 list_for_each_entry(private, &eadm_list, head) {
166 spin_lock(&sch->lock);
167 if (private->state == EADM_IDLE) {
168 private->state = EADM_BUSY;
169 list_move_tail(&private->head, &eadm_list);
170 spin_unlock(&sch->lock);
171 spin_unlock_irqrestore(&list_lock, flags);
175 spin_unlock(&sch->lock);
177 spin_unlock_irqrestore(&list_lock, flags);
182 int eadm_start_aob(struct aob *aob)
184 struct eadm_private *private;
185 struct subchannel *sch;
189 sch = eadm_get_idle_sch();
193 spin_lock_irqsave(&sch->lock, flags);
194 eadm_subchannel_set_timeout(sch, EADM_TIMEOUT);
195 ret = eadm_subchannel_start(sch, aob);
199 /* Handle start subchannel failure. */
200 eadm_subchannel_set_timeout(sch, 0);
201 private = get_eadm_private(sch);
202 private->state = EADM_NOT_OPER;
203 css_sched_sch_todo(sch, SCH_TODO_EVAL);
206 spin_unlock_irqrestore(&sch->lock, flags);
210 EXPORT_SYMBOL_GPL(eadm_start_aob);
212 static int eadm_subchannel_probe(struct subchannel *sch)
214 struct eadm_private *private;
217 private = kzalloc(sizeof(*private), GFP_KERNEL | GFP_DMA);
221 INIT_LIST_HEAD(&private->head);
222 timer_setup(&private->timer, eadm_subchannel_timeout, 0);
224 spin_lock_irq(&sch->lock);
225 set_eadm_private(sch, private);
226 private->state = EADM_IDLE;
228 sch->isc = EADM_SCH_ISC;
229 ret = cio_enable_subchannel(sch, (u32)virt_to_phys(sch));
231 set_eadm_private(sch, NULL);
232 spin_unlock_irq(&sch->lock);
236 spin_unlock_irq(&sch->lock);
238 spin_lock_irq(&list_lock);
239 list_add(&private->head, &eadm_list);
240 spin_unlock_irq(&list_lock);
245 static void eadm_quiesce(struct subchannel *sch)
247 struct eadm_private *private = get_eadm_private(sch);
248 DECLARE_COMPLETION_ONSTACK(completion);
251 spin_lock_irq(&sch->lock);
252 if (private->state != EADM_BUSY)
255 if (eadm_subchannel_clear(sch))
258 private->completion = &completion;
259 spin_unlock_irq(&sch->lock);
261 wait_for_completion_io(&completion);
263 spin_lock_irq(&sch->lock);
264 private->completion = NULL;
267 eadm_subchannel_set_timeout(sch, 0);
269 ret = cio_disable_subchannel(sch);
270 } while (ret == -EBUSY);
272 spin_unlock_irq(&sch->lock);
275 static void eadm_subchannel_remove(struct subchannel *sch)
277 struct eadm_private *private = get_eadm_private(sch);
279 spin_lock_irq(&list_lock);
280 list_del(&private->head);
281 spin_unlock_irq(&list_lock);
285 spin_lock_irq(&sch->lock);
286 set_eadm_private(sch, NULL);
287 spin_unlock_irq(&sch->lock);
292 static void eadm_subchannel_shutdown(struct subchannel *sch)
298 * eadm_subchannel_sch_event - process subchannel event
300 * @process: non-zero if function is called in process context
302 * An unspecified event occurred for this subchannel. Adjust data according
303 * to the current operational state of the subchannel. Return zero when the
304 * event has been handled sufficiently or -EAGAIN when this function should
305 * be called again in process context.
307 static int eadm_subchannel_sch_event(struct subchannel *sch, int process)
309 struct eadm_private *private;
312 spin_lock_irqsave(&sch->lock, flags);
313 if (!device_is_registered(&sch->dev))
316 if (work_pending(&sch->todo_work))
319 if (cio_update_schib(sch)) {
320 css_sched_sch_todo(sch, SCH_TODO_UNREG);
323 private = get_eadm_private(sch);
324 if (private->state == EADM_NOT_OPER)
325 private->state = EADM_IDLE;
328 spin_unlock_irqrestore(&sch->lock, flags);
333 static struct css_device_id eadm_subchannel_ids[] = {
334 { .match_flags = 0x1, .type = SUBCHANNEL_TYPE_ADM, },
335 { /* end of list */ },
337 MODULE_DEVICE_TABLE(css, eadm_subchannel_ids);
339 static struct css_driver eadm_subchannel_driver = {
341 .name = "eadm_subchannel",
342 .owner = THIS_MODULE,
344 .subchannel_type = eadm_subchannel_ids,
345 .irq = eadm_subchannel_irq,
346 .probe = eadm_subchannel_probe,
347 .remove = eadm_subchannel_remove,
348 .shutdown = eadm_subchannel_shutdown,
349 .sch_event = eadm_subchannel_sch_event,
352 static int __init eadm_sch_init(void)
356 if (!css_general_characteristics.eadm)
359 eadm_debug = debug_register("eadm_log", 16, 1, 16);
363 debug_register_view(eadm_debug, &debug_hex_ascii_view);
364 debug_set_level(eadm_debug, 2);
366 isc_register(EADM_SCH_ISC);
367 ret = css_driver_register(&eadm_subchannel_driver);
374 isc_unregister(EADM_SCH_ISC);
375 debug_unregister(eadm_debug);
379 static void __exit eadm_sch_exit(void)
381 css_driver_unregister(&eadm_subchannel_driver);
382 isc_unregister(EADM_SCH_ISC);
383 debug_unregister(eadm_debug);
385 module_init(eadm_sch_init);
386 module_exit(eadm_sch_exit);