]>
Commit | Line | Data |
---|---|---|
1da177e4 LT |
1 | /* |
2 | * mm/thrash.c | |
3 | * | |
4 | * Copyright (C) 2004, Red Hat, Inc. | |
5 | * Copyright (C) 2004, Rik van Riel <[email protected]> | |
6 | * Released under the GPL, see the file COPYING for details. | |
7 | * | |
8 | * Simple token based thrashing protection, using the algorithm | |
e21c7ffd | 9 | * described in: http://www.cse.ohio-state.edu/hpcs/WWW/HTML/publications/abs05-1.html |
7602bdf2 AC |
10 | * |
11 | * Sep 2006, Ashwin Chaugule <[email protected]> | |
12 | * Improved algorithm to pass token: | |
13 | * Each task has a priority which is incremented if it contended | |
14 | * for the token in an interval less than its previous attempt. | |
15 | * If the token is acquired, that task's priority is boosted to prevent | |
16 | * the token from bouncing around too often and to let the task make | |
17 | * some progress in its execution. | |
1da177e4 | 18 | */ |
7602bdf2 | 19 | |
1da177e4 LT |
20 | #include <linux/jiffies.h> |
21 | #include <linux/mm.h> | |
22 | #include <linux/sched.h> | |
23 | #include <linux/swap.h> | |
a433658c | 24 | #include <linux/memcontrol.h> |
1da177e4 | 25 | |
83cd81a3 KM |
26 | #include <trace/events/vmscan.h> |
27 | ||
d7911ef3 KM |
28 | #define TOKEN_AGING_INTERVAL (0xFF) |
29 | ||
1da177e4 | 30 | static DEFINE_SPINLOCK(swap_token_lock); |
7602bdf2 | 31 | struct mm_struct *swap_token_mm; |
22d5368a | 32 | static struct mem_cgroup *swap_token_memcg; |
1da177e4 | 33 | |
a433658c KM |
34 | #ifdef CONFIG_CGROUP_MEM_RES_CTLR |
35 | static struct mem_cgroup *swap_token_memcg_from_mm(struct mm_struct *mm) | |
36 | { | |
37 | struct mem_cgroup *memcg; | |
38 | ||
39 | memcg = try_get_mem_cgroup_from_mm(mm); | |
40 | if (memcg) | |
41 | css_put(mem_cgroup_css(memcg)); | |
42 | ||
43 | return memcg; | |
44 | } | |
45 | #else | |
46 | static struct mem_cgroup *swap_token_memcg_from_mm(struct mm_struct *mm) | |
47 | { | |
48 | return NULL; | |
49 | } | |
50 | #endif | |
51 | ||
a5c9b696 | 52 | void grab_swap_token(struct mm_struct *mm) |
1da177e4 | 53 | { |
7602bdf2 | 54 | int current_interval; |
83cd81a3 | 55 | unsigned int old_prio = mm->token_priority; |
53bb01f5 KM |
56 | static unsigned int global_faults; |
57 | static unsigned int last_aging; | |
1da177e4 | 58 | |
7602bdf2 | 59 | global_faults++; |
1da177e4 | 60 | |
a5c9b696 | 61 | current_interval = global_faults - mm->faultstamp; |
1da177e4 | 62 | |
7602bdf2 AC |
63 | if (!spin_trylock(&swap_token_lock)) |
64 | return; | |
1da177e4 | 65 | |
7602bdf2 | 66 | /* First come first served */ |
a433658c KM |
67 | if (!swap_token_mm) |
68 | goto replace_token; | |
69 | ||
45ebb840 KM |
70 | /* |
71 | * Usually, we don't need priority aging because long interval faults | |
72 | * makes priority decrease quickly. But there is one exception. If the | |
73 | * token owner task is sleeping, it never make long interval faults. | |
74 | * Thus, we need a priority aging mechanism instead. The requirements | |
75 | * of priority aging are | |
76 | * 1) An aging interval is reasonable enough long. Too short aging | |
77 | * interval makes quick swap token lost and decrease performance. | |
78 | * 2) The swap token owner task have to get priority aging even if | |
79 | * it's under sleep. | |
80 | */ | |
d7911ef3 KM |
81 | if ((global_faults - last_aging) > TOKEN_AGING_INTERVAL) { |
82 | swap_token_mm->token_priority /= 2; | |
83 | last_aging = global_faults; | |
84 | } | |
85 | ||
a433658c KM |
86 | if (mm == swap_token_mm) { |
87 | mm->token_priority += 2; | |
83cd81a3 | 88 | goto update_priority; |
7602bdf2 | 89 | } |
1da177e4 | 90 | |
a433658c KM |
91 | if (current_interval < mm->last_interval) |
92 | mm->token_priority++; | |
93 | else { | |
94 | if (likely(mm->token_priority > 0)) | |
95 | mm->token_priority--; | |
1da177e4 | 96 | } |
7602bdf2 | 97 | |
a433658c KM |
98 | /* Check if we deserve the token */ |
99 | if (mm->token_priority > swap_token_mm->token_priority) | |
100 | goto replace_token; | |
101 | ||
83cd81a3 | 102 | update_priority: |
d7911ef3 | 103 | trace_update_swap_token_priority(mm, old_prio, swap_token_mm); |
83cd81a3 | 104 | |
7602bdf2 | 105 | out: |
a5c9b696 HD |
106 | mm->faultstamp = global_faults; |
107 | mm->last_interval = current_interval; | |
7602bdf2 | 108 | spin_unlock(&swap_token_lock); |
a433658c KM |
109 | return; |
110 | ||
111 | replace_token: | |
112 | mm->token_priority += 2; | |
83cd81a3 | 113 | trace_replace_swap_token(swap_token_mm, mm); |
a433658c KM |
114 | swap_token_mm = mm; |
115 | swap_token_memcg = swap_token_memcg_from_mm(mm); | |
d7911ef3 | 116 | last_aging = global_faults; |
a433658c | 117 | goto out; |
1da177e4 LT |
118 | } |
119 | ||
120 | /* Called on process exit. */ | |
121 | void __put_swap_token(struct mm_struct *mm) | |
122 | { | |
123 | spin_lock(&swap_token_lock); | |
a433658c | 124 | if (likely(mm == swap_token_mm)) { |
83cd81a3 | 125 | trace_put_swap_token(swap_token_mm); |
7602bdf2 | 126 | swap_token_mm = NULL; |
a433658c KM |
127 | swap_token_memcg = NULL; |
128 | } | |
1da177e4 LT |
129 | spin_unlock(&swap_token_lock); |
130 | } | |
a433658c KM |
131 | |
132 | static bool match_memcg(struct mem_cgroup *a, struct mem_cgroup *b) | |
133 | { | |
134 | if (!a) | |
135 | return true; | |
136 | if (!b) | |
137 | return true; | |
138 | if (a == b) | |
139 | return true; | |
140 | return false; | |
141 | } | |
142 | ||
143 | void disable_swap_token(struct mem_cgroup *memcg) | |
144 | { | |
145 | /* memcg reclaim don't disable unrelated mm token. */ | |
146 | if (match_memcg(memcg, swap_token_memcg)) { | |
147 | spin_lock(&swap_token_lock); | |
148 | if (match_memcg(memcg, swap_token_memcg)) { | |
83cd81a3 | 149 | trace_disable_swap_token(swap_token_mm); |
a433658c KM |
150 | swap_token_mm = NULL; |
151 | swap_token_memcg = NULL; | |
152 | } | |
153 | spin_unlock(&swap_token_lock); | |
154 | } | |
155 | } |