/* * Copyright (C) 2008 Mathieu Desnoyers * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation; either version 2 of the License, or * (at your option) any later version. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, write to the Free Software * Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. */ #include #include #include #include #include #include #include #include #include #include #include extern struct tracepoint * const __start___tracepoints_ptrs[]; extern struct tracepoint * const __stop___tracepoints_ptrs[]; /* Set to 1 to enable tracepoint debug output */ static const int tracepoint_debug; /* * Tracepoints mutex protects the builtin and module tracepoints and the hash * table, as well as the local module list. */ static DEFINE_MUTEX(tracepoints_mutex); #ifdef CONFIG_MODULES /* Local list of struct module */ static LIST_HEAD(tracepoint_module_list); #endif /* CONFIG_MODULES */ /* * Tracepoint hash table, containing the active tracepoints. * Protected by tracepoints_mutex. */ #define TRACEPOINT_HASH_BITS 6 #define TRACEPOINT_TABLE_SIZE (1 << TRACEPOINT_HASH_BITS) static struct hlist_head tracepoint_table[TRACEPOINT_TABLE_SIZE]; /* * Note about RCU : * It is used to delay the free of multiple probes array until a quiescent * state is reached. * Tracepoint entries modifications are protected by the tracepoints_mutex. */ struct tracepoint_entry { struct hlist_node hlist; struct tracepoint_func *funcs; int refcount; /* Number of times armed. 0 if disarmed. */ char name[0]; }; struct tp_probes { union { struct rcu_head rcu; struct list_head list; } u; struct tracepoint_func probes[0]; }; static inline void *allocate_probes(int count) { struct tp_probes *p = kmalloc(count * sizeof(struct tracepoint_func) + sizeof(struct tp_probes), GFP_KERNEL); return p == NULL ? NULL : p->probes; } static void rcu_free_old_probes(struct rcu_head *head) { kfree(container_of(head, struct tp_probes, u.rcu)); } static inline void release_probes(struct tracepoint_func *old) { if (old) { struct tp_probes *tp_probes = container_of(old, struct tp_probes, probes[0]); call_rcu_sched(&tp_probes->u.rcu, rcu_free_old_probes); } } static void debug_print_probes(struct tracepoint_entry *entry) { int i; if (!tracepoint_debug || !entry->funcs) return; for (i = 0; entry->funcs[i].func; i++) printk(KERN_DEBUG "Probe %d : %p\n", i, entry->funcs[i].func); } static struct tracepoint_func * tracepoint_entry_add_probe(struct tracepoint_entry *entry, void *probe, void *data) { int nr_probes = 0; struct tracepoint_func *old, *new; WARN_ON(!probe); debug_print_probes(entry); old = entry->funcs; if (old) { /* (N -> N+1), (N != 0, 1) probes */ for (nr_probes = 0; old[nr_probes].func; nr_probes++) if (old[nr_probes].func == probe && old[nr_probes].data == data) return ERR_PTR(-EEXIST); } /* + 2 : one for new probe, one for NULL func */ new = allocate_probes(nr_probes + 2); if (new == NULL) return ERR_PTR(-ENOMEM); if (old) memcpy(new, old, nr_probes * sizeof(struct tracepoint_func)); new[nr_probes].func = probe; new[nr_probes].data = data; new[nr_probes + 1].func = NULL; entry->refcount = nr_probes + 1; entry->funcs = new; debug_print_probes(entry); return old; } static void * tracepoint_entry_remove_probe(struct tracepoint_entry *entry, void *probe, void *data) { int nr_probes = 0, nr_del = 0, i; struct tracepoint_func *old, *new; old = entry->funcs; if (!old) return ERR_PTR(-ENOENT); debug_print_probes(entry); /* (N -> M), (N > 1, M >= 0) probes */ for (nr_probes = 0; old[nr_probes].func; nr_probes++) { if (!probe || (old[nr_probes].func == probe && old[nr_probes].data == data)) nr_del++; } if (nr_probes - nr_del == 0) { /* N -> 0, (N > 1) */ entry->funcs = NULL; entry->refcount = 0; debug_print_probes(entry); return old; } else { int j = 0; /* N -> M, (N > 1, M > 0) */ /* + 1 for NULL */ new = allocate_probes(nr_probes - nr_del + 1); if (new == NULL) return ERR_PTR(-ENOMEM); for (i = 0; old[i].func; i++) if (probe && (old[i].func != probe || old[i].data != data)) new[j++] = old[i]; new[nr_probes - nr_del].func = NULL; entry->refcount = nr_probes - nr_del; entry->funcs = new; } debug_print_probes(entry); return old; } /* * Get tracepoint if the tracepoint is present in the tracepoint hash table. * Must be called with tracepoints_mutex held. * Returns NULL if not present. */ static struct tracepoint_entry *get_tracepoint(const char *name) { struct hlist_head *head; struct hlist_node *node; struct tracepoint_entry *e; u32 hash = jhash(name, strlen(name), 0); head = &tracepoint_table[hash & (TRACEPOINT_TABLE_SIZE - 1)]; hlist_for_each_entry(e, node, head, hlist) { if (!strcmp(name, e->name)) return e; } return NULL; } /* * Add the tracepoint to the tracepoint hash table. Must be called with * tracepoints_mutex held. */ static struct tracepoint_entry *add_tracepoint(const char *name) { struct hlist_head *head; struct hlist_node *node; struct tracepoint_entry *e; size_t name_len = strlen(name) + 1; u32 hash = jhash(name, name_len-1, 0); head = &tracepoint_table[hash & (TRACEPOINT_TABLE_SIZE - 1)]; hlist_for_each_entry(e, node, head, hlist) { if (!strcmp(name, e->name)) { printk(KERN_NOTICE "tracepoint %s busy\n", name); return ERR_PTR(-EEXIST); /* Already there */ } } /* * Using kmalloc here to allocate a variable length element. Could * cause some memory fragmentation if overused. */ e = kmalloc(sizeof(struct tracepoint_entry) + name_len, GFP_KERNEL); if (!e) return ERR_PTR(-ENOMEM); memcpy(&e->name[0], name, name_len); e->funcs = NULL; e->refcount = 0; hlist_add_head(&e->hlist, head); return e; } /* * Remove the tracepoint from the tracepoint hash table. Must be called with * mutex_lock held. */ static inline void remove_tracepoint(struct tracepoint_entry *e) { hlist_del(&e->hlist); kfree(e); } /* * Sets the probe callback corresponding to one tracepoint. */ static void set_tracepoint(struct tracepoint_entry **entry, struct tracepoint *elem, int active) { WARN_ON(strcmp((*entry)->name, elem->name) != 0); if (elem->regfunc && !static_key_enabled(&elem->key) && active) elem->regfunc(); else if (elem->unregfunc && static_key_enabled(&elem->key) && !active) elem->unregfunc(); /* * rcu_assign_pointer has a smp_wmb() which makes sure that the new * probe callbacks array is consistent before setting a pointer to it. * This array is referenced by __DO_TRACE from * include/linux/tracepoints.h. A matching smp_read_barrier_depends() * is used. */ rcu_assign_pointer(elem->funcs, (*entry)->funcs); if (active && !static_key_enabled(&elem->key)) static_key_slow_inc(&elem->key); else if (!active && static_key_enabled(&elem->key)) static_key_slow_dec(&elem->key); } /* * Disable a tracepoint and its probe callback. * Note: only waiting an RCU period after setting elem->call to the empty * function insures that the original callback is not used anymore. This insured * by preempt_disable around the call site. */ static void disable_tracepoint(struct tracepoint *elem) { if (elem->unregfunc && static_key_enabled(&elem->key)) elem->unregfunc(); if (static_key_enabled(&elem->key)) static_key_slow_dec(&elem->key); rcu_assign_pointer(elem->funcs, NULL); } /** * tracepoint_update_probe_range - Update a probe range * @begin: beginning of the range * @end: end of the range * * Updates the probe callback corresponding to a range of tracepoints. * Called with tracepoints_mutex held. */ static void tracepoint_update_probe_range(struct tracepoint * const *begin, struct tracepoint * const *end) { struct tracepoint * const *iter; struct tracepoint_entry *mark_entry; if (!begin) return; for (iter = begin; iter < end; iter++) { mark_entry = get_tracepoint((*iter)->name); if (mark_entry) { set_tracepoint(&mark_entry, *iter, !!mark_entry->refcount); } else { disable_tracepoint(*iter); } } } #ifdef CONFIG_MODULES void module_update_tracepoints(void) { struct tp_module *tp_mod; list_for_each_entry(tp_mod, &tracepoint_module_list, list) tracepoint_update_probe_range(tp_mod->tracepoints_ptrs, tp_mod->tracepoints_ptrs + tp_mod->num_tracepoints); } #else /* CONFIG_MODULES */ void module_update_tracepoints(void) { } #endif /* CONFIG_MODULES */ /* * Update probes, removing the faulty probes. * Called with tracepoints_mutex held. */ static void tracepoint_update_probes(void) { /* Core kernel tracepoints */ tracepoint_update_probe_range(__start___tracepoints_ptrs, __stop___tracepoints_ptrs); /* tracepoints in modules. */ module_update_tracepoints(); } static struct tracepoint_func * tracepoint_add_probe(const char *name, void *probe, void *data) { struct tracepoint_entry *entry; struct tracepoint_func *old; entry = get_tracepoint(name); if (!entry) { entry = add_tracepoint(name); if (IS_ERR(entry)) return (struct tracepoint_func *)entry; } old = tracepoint_entry_add_probe(entry, probe, data); if (IS_ERR(old) && !entry->refcount) remove_tracepoint(entry); return old; } /** * tracepoint_probe_register - Connect a probe to a tracepoint * @name: tracepoint name * @probe: probe handler * * Returns 0 if ok, error value on error. * The probe address must at least be aligned on the architecture pointer size. */ int tracepoint_probe_register(const char *name, void *probe, void *data) { struct tracepoint_func *old; mutex_lock(&tracepoints_mutex); old = tracepoint_add_probe(name, probe, data); if (IS_ERR(old)) { mutex_unlock(&tracepoints_mutex); return PTR_ERR(old); } tracepoint_update_probes(); /* may update entry */ mutex_unlock(&tracepoints_mutex); release_probes(old); return 0; } EXPORT_SYMBOL_GPL(tracepoint_probe_register); static struct tracepoint_func * tracepoint_remove_probe(const char *name, void *probe, void *data) { struct tracepoint_entry *entry; struct tracepoint_func *old; entry = get_tracepoint(name); if (!entry) return ERR_PTR(-ENOENT); old = tracepoint_entry_remove_probe(entry, probe, data); if (IS_ERR(old)) return old; if (!entry->refcount) remove_tracepoint(entry); return old; } /** * tracepoint_probe_unregister - Disconnect a probe from a tracepoint * @name: tracepoint name * @probe: probe function pointer * * We do not need to call a synchronize_sched to make sure the probes have * finished running before doing a module unload, because the module unload * itself uses stop_machine(), which insures that every preempt disabled section * have finished. */ int tracepoint_probe_unregister(const char *name, void *probe, void *data) { struct tracepoint_func *old; mutex_lock(&tracepoints_mutex); old = tracepoint_remove_probe(name, probe, data); if (IS_ERR(old)) { mutex_unlock(&tracepoints_mutex); return PTR_ERR(old); } tracepoint_update_probes(); /* may update entry */ mutex_unlock(&tracepoints_mutex); release_probes(old); return 0; } EXPORT_SYMBOL_GPL(tracepoint_probe_unregister); static LIST_HEAD(old_probes); static int need_update; static void tracepoint_add_old_probes(void *old) { need_update = 1; if (old) { struct tp_probes *tp_probes = container_of(old, struct tp_probes, probes[0]); list_add(&tp_probes->u.list, &old_probes); } } /** * tracepoint_probe_register_noupdate - register a probe but not connect * @name: tracepoint name * @probe: probe handler * * caller must call tracepoint_probe_update_all() */ int tracepoint_probe_register_noupdate(const char *name, void *probe, void *data) { struct tracepoint_func *old; mutex_lock(&tracepoints_mutex); old = tracepoint_add_probe(name, probe, data); if (IS_ERR(old)) { mutex_unlock(&tracepoints_mutex); return PTR_ERR(old); } tracepoint_add_old_probes(old); mutex_unlock(&tracepoints_mutex); return 0; } EXPORT_SYMBOL_GPL(tracepoint_probe_register_noupdate); /** * tracepoint_probe_unregister_noupdate - remove a probe but not disconnect * @name: tracepoint name * @probe: probe function pointer * * caller must call tracepoint_probe_update_all() */ int tracepoint_probe_unregister_noupdate(const char *name, void *probe, void *data) { struct tracepoint_func *old; mutex_lock(&tracepoints_mutex); old = tracepoint_remove_probe(name, probe, data); if (IS_ERR(old)) { mutex_unlock(&tracepoints_mutex); return PTR_ERR(old); } tracepoint_add_old_probes(old); mutex_unlock(&tracepoints_mutex); return 0; } EXPORT_SYMBOL_GPL(tracepoint_probe_unregister_noupdate); /** * tracepoint_probe_update_all - update tracepoints */ void tracepoint_probe_update_all(void) { LIST_HEAD(release_probes); struct tp_probes *pos, *next; mutex_lock(&tracepoints_mutex); if (!need_update) { mutex_unlock(&tracepoints_mutex); return; } if (!list_empty(&old_probes)) list_replace_init(&old_probes, &release_probes); need_update = 0; tracepoint_update_probes(); mutex_unlock(&tracepoints_mutex); list_for_each_entry_safe(pos, next, &release_probes, u.list) { list_del(&pos->u.list); call_rcu_sched(&pos->u.rcu, rcu_free_old_probes); } } EXPORT_SYMBOL_GPL(tracepoint_probe_update_all); /** * tracepoint_get_iter_range - Get a next tracepoint iterator given a range. * @tracepoint: current tracepoints (in), next tracepoint (out) * @begin: beginning of the range * @end: end of the range * * Returns whether a next tracepoint has been found (1) or not (0). * Will return the first tracepoint in the range if the input tracepoint is * NULL. */ static int tracepoint_get_iter_range(struct tracepoint * const **tracepoint, struct tracepoint * const *begin, struct tracepoint * const *end) { if (!*tracepoint && begin != end) { *tracepoint = begin; return 1; } if (*tracepoint >= begin && *tracepoint < end) return 1; return 0; } #ifdef CONFIG_MODULES static void tracepoint_get_iter(struct tracepoint_iter *iter) { int found = 0; struct tp_module *iter_mod; /* Core kernel tracepoints */ if (!iter->module) { found = tracepoint_get_iter_range(&iter->tracepoint, __start___tracepoints_ptrs, __stop___tracepoints_ptrs); if (found) goto end; } /* Tracepoints in modules */ mutex_lock(&tracepoints_mutex); list_for_each_entry(iter_mod, &tracepoint_module_list, list) { /* * Sorted module list */ if (iter_mod < iter->module) continue; else if (iter_mod > iter->module) iter->tracepoint = NULL; found = tracepoint_get_iter_range(&iter->tracepoint, iter_mod->tracepoints_ptrs, iter_mod->tracepoints_ptrs + iter_mod->num_tracepoints); if (found) { iter->module = iter_mod; break; } } mutex_unlock(&tracepoints_mutex); end: if (!found) tracepoint_iter_reset(iter); } #else /* CONFIG_MODULES */ static void tracepoint_get_iter(struct tracepoint_iter *iter) { int found = 0; /* Core kernel tracepoints */ found = tracepoint_get_iter_range(&iter->tracepoint, __start___tracepoints_ptrs, __stop___tracepoints_ptrs); if (!found) tracepoint_iter_reset(iter); } #endif /* CONFIG_MODULES */ void tracepoint_iter_start(struct tracepoint_iter *iter) { tracepoint_get_iter(iter); } EXPORT_SYMBOL_GPL(tracepoint_iter_start); void tracepoint_iter_next(struct tracepoint_iter *iter) { iter->tracepoint++; /* * iter->tracepoint may be invalid because we blindly incremented it. * Make sure it is valid by marshalling on the tracepoints, getting the * tracepoints from following modules if necessary. */ tracepoint_get_iter(iter); } EXPORT_SYMBOL_GPL(tracepoint_iter_next); void tracepoint_iter_stop(struct tracepoint_iter *iter) { } EXPORT_SYMBOL_GPL(tracepoint_iter_stop); void tracepoint_iter_reset(struct tracepoint_iter *iter) { #ifdef CONFIG_MODULES iter->module = NULL; #endif /* CONFIG_MODULES */ iter->tracepoint = NULL; } EXPORT_SYMBOL_GPL(tracepoint_iter_reset); #ifdef CONFIG_MODULES static int tracepoint_module_coming(struct module *mod) { struct tp_module *tp_mod, *iter; int ret = 0; /* * We skip modules that taint the kernel, especially those with different * module headers (for forced load), to make sure we don't cause a crash. * Staging and out-of-tree GPL modules are fine. */ if (mod->taints & ~((1 << TAINT_OOT_MODULE) | (1 << TAINT_CRAP))) return 0; mutex_lock(&tracepoints_mutex); tp_mod = kmalloc(sizeof(struct tp_module), GFP_KERNEL); if (!tp_mod) { ret = -ENOMEM; goto end; } tp_mod->num_tracepoints = mod->num_tracepoints; tp_mod->tracepoints_ptrs = mod->tracepoints_ptrs; /* * tracepoint_module_list is kept sorted by struct module pointer * address for iteration on tracepoints from a seq_file that can release * the mutex between calls. */ list_for_each_entry_reverse(iter, &tracepoint_module_list, list) { BUG_ON(iter == tp_mod); /* Should never be in the list twice */ if (iter < tp_mod) { /* We belong to the location right after iter. */ list_add(&tp_mod->list, &iter->list); goto module_added; } } /* We belong to the beginning of the list */ list_add(&tp_mod->list, &tracepoint_module_list); module_added: tracepoint_update_probe_range(mod->tracepoints_ptrs, mod->tracepoints_ptrs + mod->num_tracepoints); end: mutex_unlock(&tracepoints_mutex); return ret; } static int tracepoint_module_going(struct module *mod) { struct tp_module *pos; mutex_lock(&tracepoints_mutex); tracepoint_update_probe_range(mod->tracepoints_ptrs, mod->tracepoints_ptrs + mod->num_tracepoints); list_for_each_entry(pos, &tracepoint_module_list, list) { if (pos->tracepoints_ptrs == mod->tracepoints_ptrs) { list_del(&pos->list); kfree(pos); break; } } /* * In the case of modules that were tainted at "coming", we'll simply * walk through the list without finding it. We cannot use the "tainted" * flag on "going", in case a module taints the kernel only after being * loaded. */ mutex_unlock(&tracepoints_mutex); return 0; } int tracepoint_module_notify(struct notifier_block *self, unsigned long val, void *data) { struct module *mod = data; int ret = 0; switch (val) { case MODULE_STATE_COMING: ret = tracepoint_module_coming(mod); break; case MODULE_STATE_LIVE: break; case MODULE_STATE_GOING: ret = tracepoint_module_going(mod); break; } return ret; } struct notifier_block tracepoint_module_nb = { .notifier_call = tracepoint_module_notify, .priority = 0, }; static int init_tracepoints(void) { return register_module_notifier(&tracepoint_module_nb); } __initcall(init_tracepoints); #endif /* CONFIG_MODULES */ #ifdef CONFIG_HAVE_SYSCALL_TRACEPOINTS /* NB: reg/unreg are called while guarded with the tracepoints_mutex */ static int sys_tracepoint_refcount; void syscall_regfunc(void) { unsigned long flags; struct task_struct *g, *t; if (!sys_tracepoint_refcount) { read_lock_irqsave(&tasklist_lock, flags); do_each_thread(g, t) { /* Skip kernel threads. */ if (t->mm) set_tsk_thread_flag(t, TIF_SYSCALL_TRACEPOINT); } while_each_thread(g, t); read_unlock_irqrestore(&tasklist_lock, flags); } sys_tracepoint_refcount++; } void syscall_unregfunc(void) { unsigned long flags; struct task_struct *g, *t; sys_tracepoint_refcount--; if (!sys_tracepoint_refcount) { read_lock_irqsave(&tasklist_lock, flags); do_each_thread(g, t) { clear_tsk_thread_flag(t, TIF_SYSCALL_TRACEPOINT); } while_each_thread(g, t); read_unlock_irqrestore(&tasklist_lock, flags); } } #endif