2 * linux/kernel/softirq.c
4 * Copyright (C) 1992 Linus Torvalds
6 * Rewritten. Old one was good in 2.2, but in 2.3 it was immoral. --ANK (990903)
9 #include <linux/module.h>
10 #include <linux/kernel_stat.h>
11 #include <linux/interrupt.h>
12 #include <linux/init.h>
14 #include <linux/notifier.h>
15 #include <linux/percpu.h>
16 #include <linux/cpu.h>
17 #include <linux/kthread.h>
18 #include <linux/rcupdate.h>
19 #include <linux/smp.h>
23 - No shared variables, all the data are CPU local.
24 - If a softirq needs serialization, let it serialize itself
26 - Even if softirq is serialized, only local cpu is marked for
27 execution. Hence, we get something sort of weak cpu binding.
28 Though it is still not clear, will it result in better locality
32 - NET RX softirq. It is multithreaded and does not require
33 any global serialization.
34 - NET TX softirq. It kicks software netdevice queues, hence
35 it is logically serialized per device, but this serialization
36 is invisible to common code.
37 - Tasklets: serialized wrt itself.
40 #ifndef __ARCH_IRQ_STAT
41 irq_cpustat_t irq_stat[NR_CPUS] ____cacheline_aligned;
42 EXPORT_SYMBOL(irq_stat);
45 static struct softirq_action softirq_vec[32] __cacheline_aligned_in_smp;
47 static DEFINE_PER_CPU(struct task_struct *, ksoftirqd);
50 * we cannot loop indefinitely here to avoid userspace starvation,
51 * but we also don't want to introduce a worst case 1/HZ latency
52 * to the pending events, so lets the scheduler to balance
53 * the softirq load for us.
55 static inline void wakeup_softirqd(void)
57 /* Interrupts are disabled: no need to stop preemption */
58 struct task_struct *tsk = __get_cpu_var(ksoftirqd);
60 if (tsk && tsk->state != TASK_RUNNING)
65 * This one is for softirq.c-internal use,
66 * where hardirqs are disabled legitimately:
68 #ifdef CONFIG_TRACE_IRQFLAGS
69 static void __local_bh_disable(unsigned long ip)
73 WARN_ON_ONCE(in_irq());
75 raw_local_irq_save(flags);
76 add_preempt_count(SOFTIRQ_OFFSET);
78 * Were softirqs turned off above:
80 if (softirq_count() == SOFTIRQ_OFFSET)
81 trace_softirqs_off(ip);
82 raw_local_irq_restore(flags);
84 #else /* !CONFIG_TRACE_IRQFLAGS */
85 static inline void __local_bh_disable(unsigned long ip)
87 add_preempt_count(SOFTIRQ_OFFSET);
90 #endif /* CONFIG_TRACE_IRQFLAGS */
92 void local_bh_disable(void)
94 __local_bh_disable((unsigned long)__builtin_return_address(0));
97 EXPORT_SYMBOL(local_bh_disable);
99 void __local_bh_enable(void)
101 WARN_ON_ONCE(in_irq());
104 * softirqs should never be enabled by __local_bh_enable(),
105 * it always nests inside local_bh_enable() sections:
107 WARN_ON_ONCE(softirq_count() == SOFTIRQ_OFFSET);
109 sub_preempt_count(SOFTIRQ_OFFSET);
111 EXPORT_SYMBOL_GPL(__local_bh_enable);
114 * Special-case - softirqs can safely be enabled in
115 * cond_resched_softirq(), or by __do_softirq(),
116 * without processing still-pending softirqs:
118 void _local_bh_enable(void)
120 WARN_ON_ONCE(in_irq());
121 WARN_ON_ONCE(!irqs_disabled());
123 if (softirq_count() == SOFTIRQ_OFFSET)
124 trace_softirqs_on((unsigned long)__builtin_return_address(0));
125 sub_preempt_count(SOFTIRQ_OFFSET);
128 EXPORT_SYMBOL(_local_bh_enable);
130 void local_bh_enable(void)
132 #ifdef CONFIG_TRACE_IRQFLAGS
135 WARN_ON_ONCE(in_irq());
137 WARN_ON_ONCE(irqs_disabled());
139 #ifdef CONFIG_TRACE_IRQFLAGS
140 local_irq_save(flags);
143 * Are softirqs going to be turned on now:
145 if (softirq_count() == SOFTIRQ_OFFSET)
146 trace_softirqs_on((unsigned long)__builtin_return_address(0));
148 * Keep preemption disabled until we are done with
149 * softirq processing:
151 sub_preempt_count(SOFTIRQ_OFFSET - 1);
153 if (unlikely(!in_interrupt() && local_softirq_pending()))
157 #ifdef CONFIG_TRACE_IRQFLAGS
158 local_irq_restore(flags);
160 preempt_check_resched();
162 EXPORT_SYMBOL(local_bh_enable);
164 void local_bh_enable_ip(unsigned long ip)
166 #ifdef CONFIG_TRACE_IRQFLAGS
169 WARN_ON_ONCE(in_irq());
171 local_irq_save(flags);
174 * Are softirqs going to be turned on now:
176 if (softirq_count() == SOFTIRQ_OFFSET)
177 trace_softirqs_on(ip);
179 * Keep preemption disabled until we are done with
180 * softirq processing:
182 sub_preempt_count(SOFTIRQ_OFFSET - 1);
184 if (unlikely(!in_interrupt() && local_softirq_pending()))
188 #ifdef CONFIG_TRACE_IRQFLAGS
189 local_irq_restore(flags);
191 preempt_check_resched();
193 EXPORT_SYMBOL(local_bh_enable_ip);
196 * We restart softirq processing MAX_SOFTIRQ_RESTART times,
197 * and we fall back to softirqd after that.
199 * This number has been established via experimentation.
200 * The two things to balance is latency against fairness -
201 * we want to handle softirqs as soon as possible, but they
202 * should not be able to lock up the box.
204 #define MAX_SOFTIRQ_RESTART 10
206 asmlinkage void __do_softirq(void)
208 struct softirq_action *h;
210 int max_restart = MAX_SOFTIRQ_RESTART;
213 pending = local_softirq_pending();
214 account_system_vtime(current);
216 __local_bh_disable((unsigned long)__builtin_return_address(0));
217 trace_softirq_enter();
219 cpu = smp_processor_id();
221 /* Reset the pending bitmask before enabling irqs */
222 set_softirq_pending(0);
231 rcu_bh_qsctr_inc(cpu);
239 pending = local_softirq_pending();
240 if (pending && --max_restart)
246 trace_softirq_exit();
248 account_system_vtime(current);
252 #ifndef __ARCH_HAS_DO_SOFTIRQ
254 asmlinkage void do_softirq(void)
262 local_irq_save(flags);
264 pending = local_softirq_pending();
269 local_irq_restore(flags);
272 EXPORT_SYMBOL(do_softirq);
277 * Enter an interrupt context.
281 account_system_vtime(current);
282 add_preempt_count(HARDIRQ_OFFSET);
283 trace_hardirq_enter();
286 #ifdef __ARCH_IRQ_EXIT_IRQS_DISABLED
287 # define invoke_softirq() __do_softirq()
289 # define invoke_softirq() do_softirq()
293 * Exit an interrupt context. Process softirqs if needed and possible:
297 account_system_vtime(current);
298 trace_hardirq_exit();
299 sub_preempt_count(IRQ_EXIT_OFFSET);
300 if (!in_interrupt() && local_softirq_pending())
302 preempt_enable_no_resched();
306 * This function must run with irqs disabled!
308 inline fastcall void raise_softirq_irqoff(unsigned int nr)
310 __raise_softirq_irqoff(nr);
313 * If we're in an interrupt or softirq, we're done
314 * (this also catches softirq-disabled code). We will
315 * actually run the softirq once we return from
316 * the irq or softirq.
318 * Otherwise we wake up ksoftirqd to make sure we
319 * schedule the softirq soon.
325 EXPORT_SYMBOL(raise_softirq_irqoff);
327 void fastcall raise_softirq(unsigned int nr)
331 local_irq_save(flags);
332 raise_softirq_irqoff(nr);
333 local_irq_restore(flags);
336 void open_softirq(int nr, void (*action)(struct softirq_action*), void *data)
338 softirq_vec[nr].data = data;
339 softirq_vec[nr].action = action;
345 struct tasklet_struct *list;
348 /* Some compilers disobey section attribute on statics when not
350 static DEFINE_PER_CPU(struct tasklet_head, tasklet_vec) = { NULL };
351 static DEFINE_PER_CPU(struct tasklet_head, tasklet_hi_vec) = { NULL };
353 void fastcall __tasklet_schedule(struct tasklet_struct *t)
357 local_irq_save(flags);
358 t->next = __get_cpu_var(tasklet_vec).list;
359 __get_cpu_var(tasklet_vec).list = t;
360 raise_softirq_irqoff(TASKLET_SOFTIRQ);
361 local_irq_restore(flags);
364 EXPORT_SYMBOL(__tasklet_schedule);
366 void fastcall __tasklet_hi_schedule(struct tasklet_struct *t)
370 local_irq_save(flags);
371 t->next = __get_cpu_var(tasklet_hi_vec).list;
372 __get_cpu_var(tasklet_hi_vec).list = t;
373 raise_softirq_irqoff(HI_SOFTIRQ);
374 local_irq_restore(flags);
377 EXPORT_SYMBOL(__tasklet_hi_schedule);
379 static void tasklet_action(struct softirq_action *a)
381 struct tasklet_struct *list;
384 list = __get_cpu_var(tasklet_vec).list;
385 __get_cpu_var(tasklet_vec).list = NULL;
389 struct tasklet_struct *t = list;
393 if (tasklet_trylock(t)) {
394 if (!atomic_read(&t->count)) {
395 if (!test_and_clear_bit(TASKLET_STATE_SCHED, &t->state))
405 t->next = __get_cpu_var(tasklet_vec).list;
406 __get_cpu_var(tasklet_vec).list = t;
407 __raise_softirq_irqoff(TASKLET_SOFTIRQ);
412 static void tasklet_hi_action(struct softirq_action *a)
414 struct tasklet_struct *list;
417 list = __get_cpu_var(tasklet_hi_vec).list;
418 __get_cpu_var(tasklet_hi_vec).list = NULL;
422 struct tasklet_struct *t = list;
426 if (tasklet_trylock(t)) {
427 if (!atomic_read(&t->count)) {
428 if (!test_and_clear_bit(TASKLET_STATE_SCHED, &t->state))
438 t->next = __get_cpu_var(tasklet_hi_vec).list;
439 __get_cpu_var(tasklet_hi_vec).list = t;
440 __raise_softirq_irqoff(HI_SOFTIRQ);
446 void tasklet_init(struct tasklet_struct *t,
447 void (*func)(unsigned long), unsigned long data)
451 atomic_set(&t->count, 0);
456 EXPORT_SYMBOL(tasklet_init);
458 void tasklet_kill(struct tasklet_struct *t)
461 printk("Attempt to kill tasklet from interrupt\n");
463 while (test_and_set_bit(TASKLET_STATE_SCHED, &t->state)) {
466 while (test_bit(TASKLET_STATE_SCHED, &t->state));
468 tasklet_unlock_wait(t);
469 clear_bit(TASKLET_STATE_SCHED, &t->state);
472 EXPORT_SYMBOL(tasklet_kill);
474 void __init softirq_init(void)
476 open_softirq(TASKLET_SOFTIRQ, tasklet_action, NULL);
477 open_softirq(HI_SOFTIRQ, tasklet_hi_action, NULL);
480 static int ksoftirqd(void * __bind_cpu)
482 set_user_nice(current, 19);
483 current->flags |= PF_NOFREEZE;
485 set_current_state(TASK_INTERRUPTIBLE);
487 while (!kthread_should_stop()) {
489 if (!local_softirq_pending()) {
490 preempt_enable_no_resched();
495 __set_current_state(TASK_RUNNING);
497 while (local_softirq_pending()) {
498 /* Preempt disable stops cpu going offline.
499 If already offline, we'll be on wrong CPU:
501 if (cpu_is_offline((long)__bind_cpu))
504 preempt_enable_no_resched();
509 set_current_state(TASK_INTERRUPTIBLE);
511 __set_current_state(TASK_RUNNING);
516 /* Wait for kthread_stop */
517 set_current_state(TASK_INTERRUPTIBLE);
518 while (!kthread_should_stop()) {
520 set_current_state(TASK_INTERRUPTIBLE);
522 __set_current_state(TASK_RUNNING);
526 #ifdef CONFIG_HOTPLUG_CPU
528 * tasklet_kill_immediate is called to remove a tasklet which can already be
529 * scheduled for execution on @cpu.
531 * Unlike tasklet_kill, this function removes the tasklet
532 * _immediately_, even if the tasklet is in TASKLET_STATE_SCHED state.
534 * When this function is called, @cpu must be in the CPU_DEAD state.
536 void tasklet_kill_immediate(struct tasklet_struct *t, unsigned int cpu)
538 struct tasklet_struct **i;
540 BUG_ON(cpu_online(cpu));
541 BUG_ON(test_bit(TASKLET_STATE_RUN, &t->state));
543 if (!test_bit(TASKLET_STATE_SCHED, &t->state))
546 /* CPU is dead, so no lock needed. */
547 for (i = &per_cpu(tasklet_vec, cpu).list; *i; i = &(*i)->next) {
556 static void takeover_tasklets(unsigned int cpu)
558 struct tasklet_struct **i;
560 /* CPU is dead, so no lock needed. */
563 /* Find end, append list for that CPU. */
564 for (i = &__get_cpu_var(tasklet_vec).list; *i; i = &(*i)->next);
565 *i = per_cpu(tasklet_vec, cpu).list;
566 per_cpu(tasklet_vec, cpu).list = NULL;
567 raise_softirq_irqoff(TASKLET_SOFTIRQ);
569 for (i = &__get_cpu_var(tasklet_hi_vec).list; *i; i = &(*i)->next);
570 *i = per_cpu(tasklet_hi_vec, cpu).list;
571 per_cpu(tasklet_hi_vec, cpu).list = NULL;
572 raise_softirq_irqoff(HI_SOFTIRQ);
576 #endif /* CONFIG_HOTPLUG_CPU */
578 static int __cpuinit cpu_callback(struct notifier_block *nfb,
579 unsigned long action,
582 int hotcpu = (unsigned long)hcpu;
583 struct task_struct *p;
587 p = kthread_create(ksoftirqd, hcpu, "ksoftirqd/%d", hotcpu);
589 printk("ksoftirqd for %i failed\n", hotcpu);
592 kthread_bind(p, hotcpu);
593 per_cpu(ksoftirqd, hotcpu) = p;
596 wake_up_process(per_cpu(ksoftirqd, hotcpu));
598 #ifdef CONFIG_HOTPLUG_CPU
599 case CPU_UP_CANCELED:
600 if (!per_cpu(ksoftirqd, hotcpu))
602 /* Unbind so it can run. Fall thru. */
603 kthread_bind(per_cpu(ksoftirqd, hotcpu),
604 any_online_cpu(cpu_online_map));
606 p = per_cpu(ksoftirqd, hotcpu);
607 per_cpu(ksoftirqd, hotcpu) = NULL;
609 takeover_tasklets(hotcpu);
611 #endif /* CONFIG_HOTPLUG_CPU */
616 static struct notifier_block __cpuinitdata cpu_nfb = {
617 .notifier_call = cpu_callback
620 __init int spawn_ksoftirqd(void)
622 void *cpu = (void *)(long)smp_processor_id();
623 int err = cpu_callback(&cpu_nfb, CPU_UP_PREPARE, cpu);
625 BUG_ON(err == NOTIFY_BAD);
626 cpu_callback(&cpu_nfb, CPU_ONLINE, cpu);
627 register_cpu_notifier(&cpu_nfb);
633 * Call a function on all processors
635 int on_each_cpu(void (*func) (void *info), void *info, int retry, int wait)
640 ret = smp_call_function(func, info, retry, wait);
647 EXPORT_SYMBOL(on_each_cpu);