blob: a385ea0ab39ea677ba5705ec3ff4856fc5346a89 (
plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
|
--- b/kernel/softirq.c
+++ a/kernel/softirq.c
@@ -79,16 +79,12 @@
/*
* If ksoftirqd is scheduled, we do not want to process pending softirqs
+ * right now. Let ksoftirqd handle this at its own rate, to get fairness.
- * right now. Let ksoftirqd handle this at its own rate, to get fairness,
- * unless we're doing some of the synchronous softirqs.
*/
+static bool ksoftirqd_running(void)
-#define SOFTIRQ_NOW_MASK ((1 << HI_SOFTIRQ) | (1 << TASKLET_SOFTIRQ))
-static bool ksoftirqd_running(unsigned long pending)
{
struct task_struct *tsk = __this_cpu_read(ksoftirqd);
- if (pending & SOFTIRQ_NOW_MASK)
- return false;
return tsk && (tsk->state == TASK_RUNNING);
}
@@ -328,7 +324,7 @@
pending = local_softirq_pending();
+ if (pending && !ksoftirqd_running())
- if (pending && !ksoftirqd_running(pending))
do_softirq_own_stack();
local_irq_restore(flags);
@@ -355,7 +351,7 @@
static inline void invoke_softirq(void)
{
+ if (ksoftirqd_running())
- if (ksoftirqd_running(local_softirq_pending()))
return;
if (!force_irqthreads) {
|