]>
Commit | Line | Data |
---|---|---|
8256e47c MD |
1 | /* |
2 | * Copyright (C) 2007 Mathieu Desnoyers | |
3 | * | |
4 | * This program is free software; you can redistribute it and/or modify | |
5 | * it under the terms of the GNU General Public License as published by | |
6 | * the Free Software Foundation; either version 2 of the License, or | |
7 | * (at your option) any later version. | |
8 | * | |
9 | * This program is distributed in the hope that it will be useful, | |
10 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | |
11 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | |
12 | * GNU General Public License for more details. | |
13 | * | |
14 | * You should have received a copy of the GNU General Public License | |
15 | * along with this program; if not, write to the Free Software | |
16 | * Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. | |
17 | */ | |
18 | #include <linux/module.h> | |
19 | #include <linux/mutex.h> | |
20 | #include <linux/types.h> | |
21 | #include <linux/jhash.h> | |
22 | #include <linux/list.h> | |
23 | #include <linux/rcupdate.h> | |
24 | #include <linux/marker.h> | |
25 | #include <linux/err.h> | |
1aeb272c | 26 | #include <linux/slab.h> |
8256e47c MD |
27 | |
28 | extern struct marker __start___markers[]; | |
29 | extern struct marker __stop___markers[]; | |
30 | ||
fb40bd78 | 31 | /* Set to 1 to enable marker debug output */ |
ab883af5 | 32 | static const int marker_debug; |
fb40bd78 | 33 | |
8256e47c | 34 | /* |
314de8a9 | 35 | * markers_mutex nests inside module_mutex. Markers mutex protects the builtin |
fb40bd78 | 36 | * and module markers and the hash table. |
8256e47c MD |
37 | */ |
38 | static DEFINE_MUTEX(markers_mutex); | |
39 | ||
8256e47c MD |
40 | /* |
41 | * Marker hash table, containing the active markers. | |
42 | * Protected by module_mutex. | |
43 | */ | |
44 | #define MARKER_HASH_BITS 6 | |
45 | #define MARKER_TABLE_SIZE (1 << MARKER_HASH_BITS) | |
4de62748 | 46 | static struct hlist_head marker_table[MARKER_TABLE_SIZE]; |
8256e47c | 47 | |
fb40bd78 MD |
48 | /* |
49 | * Note about RCU : | |
50 | * It is used to make sure every handler has finished using its private data | |
51 | * between two consecutive operation (add or remove) on a given marker. It is | |
52 | * also used to delay the free of multiple probes array until a quiescent state | |
53 | * is reached. | |
54 | * marker entries modifications are protected by the markers_mutex. | |
55 | */ | |
8256e47c MD |
56 | struct marker_entry { |
57 | struct hlist_node hlist; | |
58 | char *format; | |
dc102a8f MD |
59 | /* Probe wrapper */ |
60 | void (*call)(const struct marker *mdata, void *call_private, ...); | |
fb40bd78 MD |
61 | struct marker_probe_closure single; |
62 | struct marker_probe_closure *multi; | |
8256e47c | 63 | int refcount; /* Number of times armed. 0 if disarmed. */ |
ed86a590 MD |
64 | struct rcu_head rcu; |
65 | void *oldptr; | |
1b7ae37c | 66 | int rcu_pending; |
de4fc64f | 67 | unsigned char ptype:1; |
0eec481e | 68 | unsigned char format_allocated:1; |
8256e47c MD |
69 | char name[0]; /* Contains name'\0'format'\0' */ |
70 | }; | |
71 | ||
8256e47c MD |
72 | /** |
73 | * __mark_empty_function - Empty probe callback | |
fb40bd78 MD |
74 | * @probe_private: probe private data |
75 | * @call_private: call site private data | |
8256e47c MD |
76 | * @fmt: format string |
77 | * @...: variable argument list | |
78 | * | |
79 | * Empty callback provided as a probe to the markers. By providing this to a | |
80 | * disabled marker, we make sure the execution flow is always valid even | |
81 | * though the function pointer change and the marker enabling are two distinct | |
82 | * operations that modifies the execution flow of preemptible code. | |
83 | */ | |
021aeb05 | 84 | notrace void __mark_empty_function(void *probe_private, void *call_private, |
fb40bd78 | 85 | const char *fmt, va_list *args) |
8256e47c MD |
86 | { |
87 | } | |
88 | EXPORT_SYMBOL_GPL(__mark_empty_function); | |
89 | ||
fb40bd78 MD |
90 | /* |
91 | * marker_probe_cb Callback that prepares the variable argument list for probes. | |
92 | * @mdata: pointer of type struct marker | |
93 | * @call_private: caller site private data | |
fb40bd78 MD |
94 | * @...: Variable argument list. |
95 | * | |
96 | * Since we do not use "typical" pointer based RCU in the 1 argument case, we | |
97 | * need to put a full smp_rmb() in this branch. This is why we do not use | |
98 | * rcu_dereference() for the pointer read. | |
99 | */ | |
021aeb05 MD |
100 | notrace void marker_probe_cb(const struct marker *mdata, |
101 | void *call_private, ...) | |
fb40bd78 MD |
102 | { |
103 | va_list args; | |
104 | char ptype; | |
105 | ||
106 | /* | |
e2d3b75d MD |
107 | * rcu_read_lock_sched does two things : disabling preemption to make |
108 | * sure the teardown of the callbacks can be done correctly when they | |
109 | * are in modules and they insure RCU read coherency. | |
fb40bd78 | 110 | */ |
021aeb05 | 111 | rcu_read_lock_sched_notrace(); |
58336114 | 112 | ptype = mdata->ptype; |
fb40bd78 MD |
113 | if (likely(!ptype)) { |
114 | marker_probe_func *func; | |
115 | /* Must read the ptype before ptr. They are not data dependant, | |
116 | * so we put an explicit smp_rmb() here. */ | |
117 | smp_rmb(); | |
58336114 | 118 | func = mdata->single.func; |
fb40bd78 MD |
119 | /* Must read the ptr before private data. They are not data |
120 | * dependant, so we put an explicit smp_rmb() here. */ | |
121 | smp_rmb(); | |
dc102a8f MD |
122 | va_start(args, call_private); |
123 | func(mdata->single.probe_private, call_private, mdata->format, | |
124 | &args); | |
fb40bd78 MD |
125 | va_end(args); |
126 | } else { | |
127 | struct marker_probe_closure *multi; | |
128 | int i; | |
5def9a3a MD |
129 | /* |
130 | * Read mdata->ptype before mdata->multi. | |
131 | */ | |
132 | smp_rmb(); | |
133 | multi = mdata->multi; | |
fb40bd78 MD |
134 | /* |
135 | * multi points to an array, therefore accessing the array | |
136 | * depends on reading multi. However, even in this case, | |
137 | * we must insure that the pointer is read _before_ the array | |
138 | * data. Same as rcu_dereference, but we need a full smp_rmb() | |
139 | * in the fast path, so put the explicit barrier here. | |
140 | */ | |
141 | smp_read_barrier_depends(); | |
fb40bd78 | 142 | for (i = 0; multi[i].func; i++) { |
dc102a8f MD |
143 | va_start(args, call_private); |
144 | multi[i].func(multi[i].probe_private, call_private, | |
145 | mdata->format, &args); | |
fb40bd78 MD |
146 | va_end(args); |
147 | } | |
148 | } | |
021aeb05 | 149 | rcu_read_unlock_sched_notrace(); |
fb40bd78 MD |
150 | } |
151 | EXPORT_SYMBOL_GPL(marker_probe_cb); | |
152 | ||
153 | /* | |
154 | * marker_probe_cb Callback that does not prepare the variable argument list. | |
155 | * @mdata: pointer of type struct marker | |
156 | * @call_private: caller site private data | |
fb40bd78 MD |
157 | * @...: Variable argument list. |
158 | * | |
159 | * Should be connected to markers "MARK_NOARGS". | |
160 | */ | |
021aeb05 MD |
161 | static notrace void marker_probe_cb_noarg(const struct marker *mdata, |
162 | void *call_private, ...) | |
fb40bd78 MD |
163 | { |
164 | va_list args; /* not initialized */ | |
165 | char ptype; | |
166 | ||
021aeb05 | 167 | rcu_read_lock_sched_notrace(); |
58336114 | 168 | ptype = mdata->ptype; |
fb40bd78 MD |
169 | if (likely(!ptype)) { |
170 | marker_probe_func *func; | |
171 | /* Must read the ptype before ptr. They are not data dependant, | |
172 | * so we put an explicit smp_rmb() here. */ | |
173 | smp_rmb(); | |
58336114 | 174 | func = mdata->single.func; |
fb40bd78 MD |
175 | /* Must read the ptr before private data. They are not data |
176 | * dependant, so we put an explicit smp_rmb() here. */ | |
177 | smp_rmb(); | |
dc102a8f MD |
178 | func(mdata->single.probe_private, call_private, mdata->format, |
179 | &args); | |
fb40bd78 MD |
180 | } else { |
181 | struct marker_probe_closure *multi; | |
182 | int i; | |
5def9a3a MD |
183 | /* |
184 | * Read mdata->ptype before mdata->multi. | |
185 | */ | |
186 | smp_rmb(); | |
187 | multi = mdata->multi; | |
fb40bd78 MD |
188 | /* |
189 | * multi points to an array, therefore accessing the array | |
190 | * depends on reading multi. However, even in this case, | |
191 | * we must insure that the pointer is read _before_ the array | |
192 | * data. Same as rcu_dereference, but we need a full smp_rmb() | |
193 | * in the fast path, so put the explicit barrier here. | |
194 | */ | |
195 | smp_read_barrier_depends(); | |
fb40bd78 | 196 | for (i = 0; multi[i].func; i++) |
dc102a8f MD |
197 | multi[i].func(multi[i].probe_private, call_private, |
198 | mdata->format, &args); | |
fb40bd78 | 199 | } |
021aeb05 | 200 | rcu_read_unlock_sched_notrace(); |
fb40bd78 | 201 | } |
fb40bd78 | 202 | |
ed86a590 MD |
203 | static void free_old_closure(struct rcu_head *head) |
204 | { | |
205 | struct marker_entry *entry = container_of(head, | |
206 | struct marker_entry, rcu); | |
207 | kfree(entry->oldptr); | |
208 | /* Make sure we free the data before setting the pending flag to 0 */ | |
209 | smp_wmb(); | |
210 | entry->rcu_pending = 0; | |
211 | } | |
212 | ||
fb40bd78 MD |
213 | static void debug_print_probes(struct marker_entry *entry) |
214 | { | |
215 | int i; | |
216 | ||
217 | if (!marker_debug) | |
218 | return; | |
219 | ||
220 | if (!entry->ptype) { | |
221 | printk(KERN_DEBUG "Single probe : %p %p\n", | |
222 | entry->single.func, | |
223 | entry->single.probe_private); | |
224 | } else { | |
225 | for (i = 0; entry->multi[i].func; i++) | |
226 | printk(KERN_DEBUG "Multi probe %d : %p %p\n", i, | |
227 | entry->multi[i].func, | |
228 | entry->multi[i].probe_private); | |
229 | } | |
230 | } | |
231 | ||
232 | static struct marker_probe_closure * | |
233 | marker_entry_add_probe(struct marker_entry *entry, | |
234 | marker_probe_func *probe, void *probe_private) | |
235 | { | |
236 | int nr_probes = 0; | |
237 | struct marker_probe_closure *old, *new; | |
238 | ||
239 | WARN_ON(!probe); | |
240 | ||
241 | debug_print_probes(entry); | |
242 | old = entry->multi; | |
243 | if (!entry->ptype) { | |
244 | if (entry->single.func == probe && | |
245 | entry->single.probe_private == probe_private) | |
246 | return ERR_PTR(-EBUSY); | |
247 | if (entry->single.func == __mark_empty_function) { | |
248 | /* 0 -> 1 probes */ | |
249 | entry->single.func = probe; | |
250 | entry->single.probe_private = probe_private; | |
251 | entry->refcount = 1; | |
252 | entry->ptype = 0; | |
253 | debug_print_probes(entry); | |
254 | return NULL; | |
255 | } else { | |
256 | /* 1 -> 2 probes */ | |
257 | nr_probes = 1; | |
258 | old = NULL; | |
259 | } | |
260 | } else { | |
261 | /* (N -> N+1), (N != 0, 1) probes */ | |
262 | for (nr_probes = 0; old[nr_probes].func; nr_probes++) | |
263 | if (old[nr_probes].func == probe | |
264 | && old[nr_probes].probe_private | |
265 | == probe_private) | |
266 | return ERR_PTR(-EBUSY); | |
267 | } | |
268 | /* + 2 : one for new probe, one for NULL func */ | |
269 | new = kzalloc((nr_probes + 2) * sizeof(struct marker_probe_closure), | |
270 | GFP_KERNEL); | |
271 | if (new == NULL) | |
272 | return ERR_PTR(-ENOMEM); | |
273 | if (!old) | |
274 | new[0] = entry->single; | |
275 | else | |
276 | memcpy(new, old, | |
277 | nr_probes * sizeof(struct marker_probe_closure)); | |
278 | new[nr_probes].func = probe; | |
279 | new[nr_probes].probe_private = probe_private; | |
280 | entry->refcount = nr_probes + 1; | |
281 | entry->multi = new; | |
282 | entry->ptype = 1; | |
283 | debug_print_probes(entry); | |
284 | return old; | |
285 | } | |
286 | ||
287 | static struct marker_probe_closure * | |
288 | marker_entry_remove_probe(struct marker_entry *entry, | |
289 | marker_probe_func *probe, void *probe_private) | |
290 | { | |
291 | int nr_probes = 0, nr_del = 0, i; | |
292 | struct marker_probe_closure *old, *new; | |
293 | ||
294 | old = entry->multi; | |
295 | ||
296 | debug_print_probes(entry); | |
297 | if (!entry->ptype) { | |
298 | /* 0 -> N is an error */ | |
299 | WARN_ON(entry->single.func == __mark_empty_function); | |
300 | /* 1 -> 0 probes */ | |
301 | WARN_ON(probe && entry->single.func != probe); | |
302 | WARN_ON(entry->single.probe_private != probe_private); | |
303 | entry->single.func = __mark_empty_function; | |
304 | entry->refcount = 0; | |
305 | entry->ptype = 0; | |
306 | debug_print_probes(entry); | |
307 | return NULL; | |
308 | } else { | |
309 | /* (N -> M), (N > 1, M >= 0) probes */ | |
310 | for (nr_probes = 0; old[nr_probes].func; nr_probes++) { | |
311 | if ((!probe || old[nr_probes].func == probe) | |
312 | && old[nr_probes].probe_private | |
313 | == probe_private) | |
314 | nr_del++; | |
315 | } | |
316 | } | |
317 | ||
318 | if (nr_probes - nr_del == 0) { | |
319 | /* N -> 0, (N > 1) */ | |
320 | entry->single.func = __mark_empty_function; | |
321 | entry->refcount = 0; | |
322 | entry->ptype = 0; | |
323 | } else if (nr_probes - nr_del == 1) { | |
324 | /* N -> 1, (N > 1) */ | |
325 | for (i = 0; old[i].func; i++) | |
326 | if ((probe && old[i].func != probe) || | |
327 | old[i].probe_private != probe_private) | |
328 | entry->single = old[i]; | |
329 | entry->refcount = 1; | |
330 | entry->ptype = 0; | |
331 | } else { | |
332 | int j = 0; | |
333 | /* N -> M, (N > 1, M > 1) */ | |
334 | /* + 1 for NULL */ | |
335 | new = kzalloc((nr_probes - nr_del + 1) | |
336 | * sizeof(struct marker_probe_closure), GFP_KERNEL); | |
337 | if (new == NULL) | |
338 | return ERR_PTR(-ENOMEM); | |
339 | for (i = 0; old[i].func; i++) | |
340 | if ((probe && old[i].func != probe) || | |
341 | old[i].probe_private != probe_private) | |
342 | new[j++] = old[i]; | |
343 | entry->refcount = nr_probes - nr_del; | |
344 | entry->ptype = 1; | |
345 | entry->multi = new; | |
346 | } | |
347 | debug_print_probes(entry); | |
348 | return old; | |
349 | } | |
350 | ||
8256e47c MD |
351 | /* |
352 | * Get marker if the marker is present in the marker hash table. | |
353 | * Must be called with markers_mutex held. | |
354 | * Returns NULL if not present. | |
355 | */ | |
356 | static struct marker_entry *get_marker(const char *name) | |
357 | { | |
358 | struct hlist_head *head; | |
359 | struct hlist_node *node; | |
360 | struct marker_entry *e; | |
361 | u32 hash = jhash(name, strlen(name), 0); | |
362 | ||
363 | head = &marker_table[hash & ((1 << MARKER_HASH_BITS)-1)]; | |
364 | hlist_for_each_entry(e, node, head, hlist) { | |
365 | if (!strcmp(name, e->name)) | |
366 | return e; | |
367 | } | |
368 | return NULL; | |
369 | } | |
370 | ||
371 | /* | |
372 | * Add the marker to the marker hash table. Must be called with markers_mutex | |
373 | * held. | |
374 | */ | |
fb40bd78 | 375 | static struct marker_entry *add_marker(const char *name, const char *format) |
8256e47c MD |
376 | { |
377 | struct hlist_head *head; | |
378 | struct hlist_node *node; | |
379 | struct marker_entry *e; | |
380 | size_t name_len = strlen(name) + 1; | |
381 | size_t format_len = 0; | |
382 | u32 hash = jhash(name, name_len-1, 0); | |
383 | ||
384 | if (format) | |
385 | format_len = strlen(format) + 1; | |
386 | head = &marker_table[hash & ((1 << MARKER_HASH_BITS)-1)]; | |
387 | hlist_for_each_entry(e, node, head, hlist) { | |
388 | if (!strcmp(name, e->name)) { | |
389 | printk(KERN_NOTICE | |
fb40bd78 MD |
390 | "Marker %s busy\n", name); |
391 | return ERR_PTR(-EBUSY); /* Already there */ | |
8256e47c MD |
392 | } |
393 | } | |
394 | /* | |
395 | * Using kmalloc here to allocate a variable length element. Could | |
396 | * cause some memory fragmentation if overused. | |
397 | */ | |
398 | e = kmalloc(sizeof(struct marker_entry) + name_len + format_len, | |
399 | GFP_KERNEL); | |
400 | if (!e) | |
fb40bd78 | 401 | return ERR_PTR(-ENOMEM); |
8256e47c MD |
402 | memcpy(&e->name[0], name, name_len); |
403 | if (format) { | |
404 | e->format = &e->name[name_len]; | |
405 | memcpy(e->format, format, format_len); | |
fb40bd78 MD |
406 | if (strcmp(e->format, MARK_NOARGS) == 0) |
407 | e->call = marker_probe_cb_noarg; | |
408 | else | |
409 | e->call = marker_probe_cb; | |
8256e47c MD |
410 | trace_mark(core_marker_format, "name %s format %s", |
411 | e->name, e->format); | |
fb40bd78 | 412 | } else { |
8256e47c | 413 | e->format = NULL; |
fb40bd78 MD |
414 | e->call = marker_probe_cb; |
415 | } | |
416 | e->single.func = __mark_empty_function; | |
417 | e->single.probe_private = NULL; | |
418 | e->multi = NULL; | |
419 | e->ptype = 0; | |
0eec481e | 420 | e->format_allocated = 0; |
8256e47c | 421 | e->refcount = 0; |
ed86a590 | 422 | e->rcu_pending = 0; |
8256e47c | 423 | hlist_add_head(&e->hlist, head); |
fb40bd78 | 424 | return e; |
8256e47c MD |
425 | } |
426 | ||
427 | /* | |
428 | * Remove the marker from the marker hash table. Must be called with mutex_lock | |
429 | * held. | |
430 | */ | |
fb40bd78 | 431 | static int remove_marker(const char *name) |
8256e47c MD |
432 | { |
433 | struct hlist_head *head; | |
434 | struct hlist_node *node; | |
435 | struct marker_entry *e; | |
436 | int found = 0; | |
437 | size_t len = strlen(name) + 1; | |
8256e47c MD |
438 | u32 hash = jhash(name, len-1, 0); |
439 | ||
440 | head = &marker_table[hash & ((1 << MARKER_HASH_BITS)-1)]; | |
441 | hlist_for_each_entry(e, node, head, hlist) { | |
442 | if (!strcmp(name, e->name)) { | |
443 | found = 1; | |
444 | break; | |
445 | } | |
446 | } | |
fb40bd78 MD |
447 | if (!found) |
448 | return -ENOENT; | |
449 | if (e->single.func != __mark_empty_function) | |
450 | return -EBUSY; | |
451 | hlist_del(&e->hlist); | |
0eec481e LJ |
452 | if (e->format_allocated) |
453 | kfree(e->format); | |
ed86a590 MD |
454 | /* Make sure the call_rcu has been executed */ |
455 | if (e->rcu_pending) | |
456 | rcu_barrier_sched(); | |
fb40bd78 MD |
457 | kfree(e); |
458 | return 0; | |
8256e47c MD |
459 | } |
460 | ||
461 | /* | |
462 | * Set the mark_entry format to the format found in the element. | |
463 | */ | |
0eec481e | 464 | static int marker_set_format(struct marker_entry *entry, const char *format) |
8256e47c | 465 | { |
0eec481e LJ |
466 | entry->format = kstrdup(format, GFP_KERNEL); |
467 | if (!entry->format) | |
8256e47c | 468 | return -ENOMEM; |
0eec481e LJ |
469 | entry->format_allocated = 1; |
470 | ||
8256e47c | 471 | trace_mark(core_marker_format, "name %s format %s", |
0eec481e | 472 | entry->name, entry->format); |
8256e47c MD |
473 | return 0; |
474 | } | |
475 | ||
476 | /* | |
477 | * Sets the probe callback corresponding to one marker. | |
478 | */ | |
0eec481e | 479 | static int set_marker(struct marker_entry *entry, struct marker *elem, |
fb40bd78 | 480 | int active) |
8256e47c | 481 | { |
c1df1bd2 | 482 | int ret = 0; |
0eec481e | 483 | WARN_ON(strcmp(entry->name, elem->name) != 0); |
8256e47c | 484 | |
0eec481e LJ |
485 | if (entry->format) { |
486 | if (strcmp(entry->format, elem->format) != 0) { | |
8256e47c MD |
487 | printk(KERN_NOTICE |
488 | "Format mismatch for probe %s " | |
489 | "(%s), marker (%s)\n", | |
0eec481e LJ |
490 | entry->name, |
491 | entry->format, | |
8256e47c MD |
492 | elem->format); |
493 | return -EPERM; | |
494 | } | |
495 | } else { | |
496 | ret = marker_set_format(entry, elem->format); | |
497 | if (ret) | |
498 | return ret; | |
499 | } | |
fb40bd78 MD |
500 | |
501 | /* | |
502 | * probe_cb setup (statically known) is done here. It is | |
503 | * asynchronous with the rest of execution, therefore we only | |
504 | * pass from a "safe" callback (with argument) to an "unsafe" | |
505 | * callback (does not set arguments). | |
506 | */ | |
0eec481e | 507 | elem->call = entry->call; |
fb40bd78 MD |
508 | /* |
509 | * Sanity check : | |
510 | * We only update the single probe private data when the ptr is | |
511 | * set to a _non_ single probe! (0 -> 1 and N -> 1, N != 1) | |
512 | */ | |
513 | WARN_ON(elem->single.func != __mark_empty_function | |
0eec481e LJ |
514 | && elem->single.probe_private != entry->single.probe_private |
515 | && !elem->ptype); | |
516 | elem->single.probe_private = entry->single.probe_private; | |
fb40bd78 MD |
517 | /* |
518 | * Make sure the private data is valid when we update the | |
519 | * single probe ptr. | |
520 | */ | |
521 | smp_wmb(); | |
0eec481e | 522 | elem->single.func = entry->single.func; |
fb40bd78 MD |
523 | /* |
524 | * We also make sure that the new probe callbacks array is consistent | |
525 | * before setting a pointer to it. | |
526 | */ | |
0eec481e | 527 | rcu_assign_pointer(elem->multi, entry->multi); |
fb40bd78 MD |
528 | /* |
529 | * Update the function or multi probe array pointer before setting the | |
530 | * ptype. | |
531 | */ | |
532 | smp_wmb(); | |
0eec481e | 533 | elem->ptype = entry->ptype; |
c1df1bd2 MD |
534 | |
535 | if (elem->tp_name && (active ^ elem->state)) { | |
536 | WARN_ON(!elem->tp_cb); | |
537 | /* | |
538 | * It is ok to directly call the probe registration because type | |
539 | * checking has been done in the __trace_mark_tp() macro. | |
540 | */ | |
541 | ||
542 | if (active) { | |
543 | /* | |
544 | * try_module_get should always succeed because we hold | |
545 | * lock_module() to get the tp_cb address. | |
546 | */ | |
547 | ret = try_module_get(__module_text_address( | |
548 | (unsigned long)elem->tp_cb)); | |
549 | BUG_ON(!ret); | |
550 | ret = tracepoint_probe_register_noupdate( | |
551 | elem->tp_name, | |
552 | elem->tp_cb); | |
553 | } else { | |
554 | ret = tracepoint_probe_unregister_noupdate( | |
555 | elem->tp_name, | |
556 | elem->tp_cb); | |
557 | /* | |
558 | * tracepoint_probe_update_all() must be called | |
559 | * before the module containing tp_cb is unloaded. | |
560 | */ | |
561 | module_put(__module_text_address( | |
562 | (unsigned long)elem->tp_cb)); | |
563 | } | |
564 | } | |
fb40bd78 MD |
565 | elem->state = active; |
566 | ||
c1df1bd2 | 567 | return ret; |
8256e47c MD |
568 | } |
569 | ||
570 | /* | |
571 | * Disable a marker and its probe callback. | |
fd3c36f8 MD |
572 | * Note: only waiting an RCU period after setting elem->call to the empty |
573 | * function insures that the original callback is not used anymore. This insured | |
e2d3b75d | 574 | * by rcu_read_lock_sched around the call site. |
8256e47c MD |
575 | */ |
576 | static void disable_marker(struct marker *elem) | |
577 | { | |
c1df1bd2 MD |
578 | int ret; |
579 | ||
fb40bd78 | 580 | /* leave "call" as is. It is known statically. */ |
c1df1bd2 MD |
581 | if (elem->tp_name && elem->state) { |
582 | WARN_ON(!elem->tp_cb); | |
583 | /* | |
584 | * It is ok to directly call the probe registration because type | |
585 | * checking has been done in the __trace_mark_tp() macro. | |
586 | */ | |
587 | ret = tracepoint_probe_unregister_noupdate(elem->tp_name, | |
588 | elem->tp_cb); | |
589 | WARN_ON(ret); | |
590 | /* | |
591 | * tracepoint_probe_update_all() must be called | |
592 | * before the module containing tp_cb is unloaded. | |
593 | */ | |
594 | module_put(__module_text_address((unsigned long)elem->tp_cb)); | |
595 | } | |
8256e47c | 596 | elem->state = 0; |
fb40bd78 MD |
597 | elem->single.func = __mark_empty_function; |
598 | /* Update the function before setting the ptype */ | |
599 | smp_wmb(); | |
600 | elem->ptype = 0; /* single probe */ | |
8256e47c MD |
601 | /* |
602 | * Leave the private data and id there, because removal is racy and | |
fd3c36f8 MD |
603 | * should be done only after an RCU period. These are never used until |
604 | * the next initialization anyway. | |
8256e47c MD |
605 | */ |
606 | } | |
607 | ||
608 | /** | |
609 | * marker_update_probe_range - Update a probe range | |
610 | * @begin: beginning of the range | |
611 | * @end: end of the range | |
8256e47c MD |
612 | * |
613 | * Updates the probe callback corresponding to a range of markers. | |
8256e47c MD |
614 | */ |
615 | void marker_update_probe_range(struct marker *begin, | |
fb40bd78 | 616 | struct marker *end) |
8256e47c MD |
617 | { |
618 | struct marker *iter; | |
619 | struct marker_entry *mark_entry; | |
620 | ||
314de8a9 | 621 | mutex_lock(&markers_mutex); |
8256e47c MD |
622 | for (iter = begin; iter < end; iter++) { |
623 | mark_entry = get_marker(iter->name); | |
fb40bd78 | 624 | if (mark_entry) { |
0eec481e | 625 | set_marker(mark_entry, iter, !!mark_entry->refcount); |
8256e47c MD |
626 | /* |
627 | * ignore error, continue | |
628 | */ | |
8256e47c MD |
629 | } else { |
630 | disable_marker(iter); | |
631 | } | |
632 | } | |
314de8a9 | 633 | mutex_unlock(&markers_mutex); |
8256e47c MD |
634 | } |
635 | ||
636 | /* | |
637 | * Update probes, removing the faulty probes. | |
fb40bd78 MD |
638 | * |
639 | * Internal callback only changed before the first probe is connected to it. | |
640 | * Single probe private data can only be changed on 0 -> 1 and 2 -> 1 | |
641 | * transitions. All other transitions will leave the old private data valid. | |
642 | * This makes the non-atomicity of the callback/private data updates valid. | |
643 | * | |
644 | * "special case" updates : | |
645 | * 0 -> 1 callback | |
646 | * 1 -> 0 callback | |
647 | * 1 -> 2 callbacks | |
648 | * 2 -> 1 callbacks | |
649 | * Other updates all behave the same, just like the 2 -> 3 or 3 -> 2 updates. | |
650 | * Site effect : marker_set_format may delete the marker entry (creating a | |
651 | * replacement). | |
8256e47c | 652 | */ |
fb40bd78 | 653 | static void marker_update_probes(void) |
8256e47c | 654 | { |
8256e47c | 655 | /* Core kernel markers */ |
fb40bd78 | 656 | marker_update_probe_range(__start___markers, __stop___markers); |
8256e47c | 657 | /* Markers in modules. */ |
fb40bd78 | 658 | module_update_markers(); |
c1df1bd2 | 659 | tracepoint_probe_update_all(); |
8256e47c MD |
660 | } |
661 | ||
662 | /** | |
663 | * marker_probe_register - Connect a probe to a marker | |
664 | * @name: marker name | |
665 | * @format: format string | |
666 | * @probe: probe handler | |
fb40bd78 | 667 | * @probe_private: probe private data |
8256e47c MD |
668 | * |
669 | * private data must be a valid allocated memory address, or NULL. | |
670 | * Returns 0 if ok, error value on error. | |
fb40bd78 | 671 | * The probe address must at least be aligned on the architecture pointer size. |
8256e47c MD |
672 | */ |
673 | int marker_probe_register(const char *name, const char *format, | |
fb40bd78 | 674 | marker_probe_func *probe, void *probe_private) |
8256e47c MD |
675 | { |
676 | struct marker_entry *entry; | |
314de8a9 | 677 | int ret = 0; |
fb40bd78 | 678 | struct marker_probe_closure *old; |
8256e47c MD |
679 | |
680 | mutex_lock(&markers_mutex); | |
681 | entry = get_marker(name); | |
fb40bd78 MD |
682 | if (!entry) { |
683 | entry = add_marker(name, format); | |
48043bcd | 684 | if (IS_ERR(entry)) |
fb40bd78 | 685 | ret = PTR_ERR(entry); |
48043bcd LJ |
686 | } else if (format) { |
687 | if (!entry->format) | |
0eec481e | 688 | ret = marker_set_format(entry, format); |
48043bcd LJ |
689 | else if (strcmp(entry->format, format)) |
690 | ret = -EPERM; | |
8256e47c | 691 | } |
48043bcd LJ |
692 | if (ret) |
693 | goto end; | |
694 | ||
ed86a590 MD |
695 | /* |
696 | * If we detect that a call_rcu is pending for this marker, | |
697 | * make sure it's executed now. | |
698 | */ | |
699 | if (entry->rcu_pending) | |
700 | rcu_barrier_sched(); | |
fb40bd78 MD |
701 | old = marker_entry_add_probe(entry, probe, probe_private); |
702 | if (IS_ERR(old)) { | |
703 | ret = PTR_ERR(old); | |
8256e47c | 704 | goto end; |
fb40bd78 | 705 | } |
314de8a9 | 706 | mutex_unlock(&markers_mutex); |
2bdba316 | 707 | marker_update_probes(); |
fb40bd78 MD |
708 | mutex_lock(&markers_mutex); |
709 | entry = get_marker(name); | |
2bdba316 MD |
710 | if (!entry) |
711 | goto end; | |
ed86a590 MD |
712 | if (entry->rcu_pending) |
713 | rcu_barrier_sched(); | |
714 | entry->oldptr = old; | |
715 | entry->rcu_pending = 1; | |
716 | /* write rcu_pending before calling the RCU callback */ | |
717 | smp_wmb(); | |
718 | call_rcu_sched(&entry->rcu, free_old_closure); | |
8256e47c MD |
719 | end: |
720 | mutex_unlock(&markers_mutex); | |
8256e47c MD |
721 | return ret; |
722 | } | |
723 | EXPORT_SYMBOL_GPL(marker_probe_register); | |
724 | ||
725 | /** | |
726 | * marker_probe_unregister - Disconnect a probe from a marker | |
727 | * @name: marker name | |
fb40bd78 MD |
728 | * @probe: probe function pointer |
729 | * @probe_private: probe private data | |
8256e47c MD |
730 | * |
731 | * Returns the private data given to marker_probe_register, or an ERR_PTR(). | |
fb40bd78 MD |
732 | * We do not need to call a synchronize_sched to make sure the probes have |
733 | * finished running before doing a module unload, because the module unload | |
734 | * itself uses stop_machine(), which insures that every preempt disabled section | |
735 | * have finished. | |
8256e47c | 736 | */ |
fb40bd78 MD |
737 | int marker_probe_unregister(const char *name, |
738 | marker_probe_func *probe, void *probe_private) | |
8256e47c | 739 | { |
8256e47c | 740 | struct marker_entry *entry; |
fb40bd78 | 741 | struct marker_probe_closure *old; |
544adb41 | 742 | int ret = -ENOENT; |
8256e47c MD |
743 | |
744 | mutex_lock(&markers_mutex); | |
745 | entry = get_marker(name); | |
544adb41 | 746 | if (!entry) |
8256e47c | 747 | goto end; |
ed86a590 MD |
748 | if (entry->rcu_pending) |
749 | rcu_barrier_sched(); | |
fb40bd78 | 750 | old = marker_entry_remove_probe(entry, probe, probe_private); |
314de8a9 | 751 | mutex_unlock(&markers_mutex); |
2bdba316 | 752 | marker_update_probes(); |
fb40bd78 MD |
753 | mutex_lock(&markers_mutex); |
754 | entry = get_marker(name); | |
544adb41 JJ |
755 | if (!entry) |
756 | goto end; | |
ed86a590 MD |
757 | if (entry->rcu_pending) |
758 | rcu_barrier_sched(); | |
759 | entry->oldptr = old; | |
760 | entry->rcu_pending = 1; | |
761 | /* write rcu_pending before calling the RCU callback */ | |
762 | smp_wmb(); | |
763 | call_rcu_sched(&entry->rcu, free_old_closure); | |
fb40bd78 | 764 | remove_marker(name); /* Ignore busy error message */ |
544adb41 | 765 | ret = 0; |
8256e47c MD |
766 | end: |
767 | mutex_unlock(&markers_mutex); | |
fb40bd78 | 768 | return ret; |
8256e47c MD |
769 | } |
770 | EXPORT_SYMBOL_GPL(marker_probe_unregister); | |
771 | ||
fb40bd78 MD |
772 | static struct marker_entry * |
773 | get_marker_from_private_data(marker_probe_func *probe, void *probe_private) | |
8256e47c | 774 | { |
8256e47c | 775 | struct marker_entry *entry; |
8256e47c | 776 | unsigned int i; |
fb40bd78 MD |
777 | struct hlist_head *head; |
778 | struct hlist_node *node; | |
8256e47c | 779 | |
8256e47c MD |
780 | for (i = 0; i < MARKER_TABLE_SIZE; i++) { |
781 | head = &marker_table[i]; | |
782 | hlist_for_each_entry(entry, node, head, hlist) { | |
fb40bd78 MD |
783 | if (!entry->ptype) { |
784 | if (entry->single.func == probe | |
785 | && entry->single.probe_private | |
786 | == probe_private) | |
787 | return entry; | |
788 | } else { | |
789 | struct marker_probe_closure *closure; | |
790 | closure = entry->multi; | |
791 | for (i = 0; closure[i].func; i++) { | |
792 | if (closure[i].func == probe && | |
793 | closure[i].probe_private | |
794 | == probe_private) | |
795 | return entry; | |
796 | } | |
8256e47c MD |
797 | } |
798 | } | |
799 | } | |
fb40bd78 | 800 | return NULL; |
8256e47c | 801 | } |
8256e47c MD |
802 | |
803 | /** | |
fb40bd78 MD |
804 | * marker_probe_unregister_private_data - Disconnect a probe from a marker |
805 | * @probe: probe function | |
806 | * @probe_private: probe private data | |
8256e47c | 807 | * |
fb40bd78 MD |
808 | * Unregister a probe by providing the registered private data. |
809 | * Only removes the first marker found in hash table. | |
810 | * Return 0 on success or error value. | |
811 | * We do not need to call a synchronize_sched to make sure the probes have | |
812 | * finished running before doing a module unload, because the module unload | |
813 | * itself uses stop_machine(), which insures that every preempt disabled section | |
814 | * have finished. | |
8256e47c | 815 | */ |
fb40bd78 MD |
816 | int marker_probe_unregister_private_data(marker_probe_func *probe, |
817 | void *probe_private) | |
8256e47c MD |
818 | { |
819 | struct marker_entry *entry; | |
314de8a9 | 820 | int ret = 0; |
fb40bd78 | 821 | struct marker_probe_closure *old; |
8256e47c MD |
822 | |
823 | mutex_lock(&markers_mutex); | |
fb40bd78 | 824 | entry = get_marker_from_private_data(probe, probe_private); |
8256e47c MD |
825 | if (!entry) { |
826 | ret = -ENOENT; | |
827 | goto end; | |
828 | } | |
ed86a590 MD |
829 | if (entry->rcu_pending) |
830 | rcu_barrier_sched(); | |
fb40bd78 | 831 | old = marker_entry_remove_probe(entry, NULL, probe_private); |
8256e47c | 832 | mutex_unlock(&markers_mutex); |
2bdba316 | 833 | marker_update_probes(); |
8256e47c | 834 | mutex_lock(&markers_mutex); |
fb40bd78 | 835 | entry = get_marker_from_private_data(probe, probe_private); |
2bdba316 MD |
836 | if (!entry) |
837 | goto end; | |
ed86a590 MD |
838 | if (entry->rcu_pending) |
839 | rcu_barrier_sched(); | |
840 | entry->oldptr = old; | |
841 | entry->rcu_pending = 1; | |
842 | /* write rcu_pending before calling the RCU callback */ | |
843 | smp_wmb(); | |
844 | call_rcu_sched(&entry->rcu, free_old_closure); | |
fb40bd78 | 845 | remove_marker(entry->name); /* Ignore busy error message */ |
8256e47c MD |
846 | end: |
847 | mutex_unlock(&markers_mutex); | |
8256e47c MD |
848 | return ret; |
849 | } | |
fb40bd78 | 850 | EXPORT_SYMBOL_GPL(marker_probe_unregister_private_data); |
8256e47c MD |
851 | |
852 | /** | |
853 | * marker_get_private_data - Get a marker's probe private data | |
854 | * @name: marker name | |
fb40bd78 MD |
855 | * @probe: probe to match |
856 | * @num: get the nth matching probe's private data | |
8256e47c | 857 | * |
fb40bd78 MD |
858 | * Returns the nth private data pointer (starting from 0) matching, or an |
859 | * ERR_PTR. | |
8256e47c MD |
860 | * Returns the private data pointer, or an ERR_PTR. |
861 | * The private data pointer should _only_ be dereferenced if the caller is the | |
862 | * owner of the data, or its content could vanish. This is mostly used to | |
863 | * confirm that a caller is the owner of a registered probe. | |
864 | */ | |
fb40bd78 MD |
865 | void *marker_get_private_data(const char *name, marker_probe_func *probe, |
866 | int num) | |
8256e47c MD |
867 | { |
868 | struct hlist_head *head; | |
869 | struct hlist_node *node; | |
870 | struct marker_entry *e; | |
871 | size_t name_len = strlen(name) + 1; | |
872 | u32 hash = jhash(name, name_len-1, 0); | |
fb40bd78 | 873 | int i; |
8256e47c MD |
874 | |
875 | head = &marker_table[hash & ((1 << MARKER_HASH_BITS)-1)]; | |
876 | hlist_for_each_entry(e, node, head, hlist) { | |
877 | if (!strcmp(name, e->name)) { | |
fb40bd78 MD |
878 | if (!e->ptype) { |
879 | if (num == 0 && e->single.func == probe) | |
880 | return e->single.probe_private; | |
fb40bd78 MD |
881 | } else { |
882 | struct marker_probe_closure *closure; | |
883 | int match = 0; | |
884 | closure = e->multi; | |
885 | for (i = 0; closure[i].func; i++) { | |
886 | if (closure[i].func != probe) | |
887 | continue; | |
888 | if (match++ == num) | |
889 | return closure[i].probe_private; | |
890 | } | |
891 | } | |
5d9881ea | 892 | break; |
8256e47c MD |
893 | } |
894 | } | |
895 | return ERR_PTR(-ENOENT); | |
896 | } | |
897 | EXPORT_SYMBOL_GPL(marker_get_private_data); | |
a419246a | 898 | |
227a8375 IM |
899 | #ifdef CONFIG_MODULES |
900 | ||
a419246a MD |
901 | int marker_module_notify(struct notifier_block *self, |
902 | unsigned long val, void *data) | |
903 | { | |
904 | struct module *mod = data; | |
905 | ||
906 | switch (val) { | |
907 | case MODULE_STATE_COMING: | |
908 | marker_update_probe_range(mod->markers, | |
909 | mod->markers + mod->num_markers); | |
910 | break; | |
911 | case MODULE_STATE_GOING: | |
912 | marker_update_probe_range(mod->markers, | |
913 | mod->markers + mod->num_markers); | |
914 | break; | |
915 | } | |
916 | return 0; | |
917 | } | |
918 | ||
919 | struct notifier_block marker_module_nb = { | |
920 | .notifier_call = marker_module_notify, | |
921 | .priority = 0, | |
922 | }; | |
923 | ||
924 | static int init_markers(void) | |
925 | { | |
926 | return register_module_notifier(&marker_module_nb); | |
927 | } | |
928 | __initcall(init_markers); | |
227a8375 IM |
929 | |
930 | #endif /* CONFIG_MODULES */ |