Merge branch 'core/locking' of git://git.kernel.org/pub/scm/linux/kernel/git/tip...
authorLinus Torvalds <torvalds@linux-foundation.org>
Mon, 14 Jul 2008 21:55:13 +0000 (14:55 -0700)
committerLinus Torvalds <torvalds@linux-foundation.org>
Mon, 14 Jul 2008 21:55:13 +0000 (14:55 -0700)
* 'core/locking' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip:
  lockdep: fix kernel/fork.c warning
  lockdep: fix ftrace irq tracing false positive
  lockdep: remove duplicate definition of STATIC_LOCKDEP_MAP_INIT
  lockdep: add lock_class information to lock_chain and output it
  lockdep: add lock_class information to lock_chain and output it
  lockdep: output lock_class key instead of address for forward dependency output
  __mutex_lock_common: use signal_pending_state()
  mutex-debug: check mutex magic before owner

Fixed up conflict in kernel/fork.c manually

include/linux/lockdep.h
kernel/fork.c
kernel/lockdep.c
kernel/lockdep_internals.h
kernel/lockdep_proc.c
kernel/mutex-debug.c
kernel/mutex.c

index 4c4d236ded1885415066c774877abc9fc39592f7..2486eb4edbf146b77effbfe27c14689ea70db1de 100644 (file)
@@ -182,6 +182,9 @@ struct lock_list {
  * We record lock dependency chains, so that we can cache them:
  */
 struct lock_chain {
+       u8                              irq_context;
+       u8                              depth;
+       u16                             base;
        struct list_head                entry;
        u64                             chain_key;
 };
@@ -275,14 +278,6 @@ extern void lockdep_init_map(struct lockdep_map *lock, const char *name,
                lockdep_init_map(&(lock)->dep_map, #lock, \
                                 (lock)->dep_map.key, sub)
 
-/*
- * To initialize a lockdep_map statically use this macro.
- * Note that _name must not be NULL.
- */
-#define STATIC_LOCKDEP_MAP_INIT(_name, _key) \
-       { .name = (_name), .key = (void *)(_key), }
-
-
 /*
  * Acquire a lock.
  *
index 16fd412c756d708c244545b4a264f1891840f764..4bd2f516401f5409c76772a314e014c1357c47f7 100644 (file)
@@ -910,7 +910,7 @@ static struct task_struct *copy_process(unsigned long clone_flags,
 
        rt_mutex_init_task(p);
 
-#if defined(CONFIG_TRACE_IRQFLAGS) && defined(CONFIG_LOCKDEP)
+#ifdef CONFIG_PROVE_LOCKING
        DEBUG_LOCKS_WARN_ON(!p->hardirqs_enabled);
        DEBUG_LOCKS_WARN_ON(!p->softirqs_enabled);
 #endif
index 65548eff029e4d61360d5a29626bce3f64ffaae1..d38a643629735b533c2e0ddff300a9881e1bf674 100644 (file)
@@ -1462,7 +1462,14 @@ out_bug:
 }
 
 unsigned long nr_lock_chains;
-static struct lock_chain lock_chains[MAX_LOCKDEP_CHAINS];
+struct lock_chain lock_chains[MAX_LOCKDEP_CHAINS];
+int nr_chain_hlocks;
+static u16 chain_hlocks[MAX_LOCKDEP_CHAIN_HLOCKS];
+
+struct lock_class *lock_chain_get_class(struct lock_chain *chain, int i)
+{
+       return lock_classes + chain_hlocks[chain->base + i];
+}
 
 /*
  * Look up a dependency chain. If the key is not present yet then
@@ -1470,10 +1477,15 @@ static struct lock_chain lock_chains[MAX_LOCKDEP_CHAINS];
  * validated. If the key is already hashed, return 0.
  * (On return with 1 graph_lock is held.)
  */
-static inline int lookup_chain_cache(u64 chain_key, struct lock_class *class)
+static inline int lookup_chain_cache(struct task_struct *curr,
+                                    struct held_lock *hlock,
+                                    u64 chain_key)
 {
+       struct lock_class *class = hlock->class;
        struct list_head *hash_head = chainhashentry(chain_key);
        struct lock_chain *chain;
+       struct held_lock *hlock_curr, *hlock_next;
+       int i, j, n, cn;
 
        if (DEBUG_LOCKS_WARN_ON(!irqs_disabled()))
                return 0;
@@ -1521,6 +1533,32 @@ cache_hit:
        }
        chain = lock_chains + nr_lock_chains++;
        chain->chain_key = chain_key;
+       chain->irq_context = hlock->irq_context;
+       /* Find the first held_lock of current chain */
+       hlock_next = hlock;
+       for (i = curr->lockdep_depth - 1; i >= 0; i--) {
+               hlock_curr = curr->held_locks + i;
+               if (hlock_curr->irq_context != hlock_next->irq_context)
+                       break;
+               hlock_next = hlock;
+       }
+       i++;
+       chain->depth = curr->lockdep_depth + 1 - i;
+       cn = nr_chain_hlocks;
+       while (cn + chain->depth <= MAX_LOCKDEP_CHAIN_HLOCKS) {
+               n = cmpxchg(&nr_chain_hlocks, cn, cn + chain->depth);
+               if (n == cn)
+                       break;
+               cn = n;
+       }
+       if (likely(cn + chain->depth <= MAX_LOCKDEP_CHAIN_HLOCKS)) {
+               chain->base = cn;
+               for (j = 0; j < chain->depth - 1; j++, i++) {
+                       int lock_id = curr->held_locks[i].class - lock_classes;
+                       chain_hlocks[chain->base + j] = lock_id;
+               }
+               chain_hlocks[chain->base + j] = class - lock_classes;
+       }
        list_add_tail_rcu(&chain->entry, hash_head);
        debug_atomic_inc(&chain_lookup_misses);
        inc_chains();
@@ -1542,7 +1580,7 @@ static int validate_chain(struct task_struct *curr, struct lockdep_map *lock,
         * graph_lock for us)
         */
        if (!hlock->trylock && (hlock->check == 2) &&
-                       lookup_chain_cache(chain_key, hlock->class)) {
+           lookup_chain_cache(curr, hlock, chain_key)) {
                /*
                 * Check whether last held lock:
                 *
@@ -2668,7 +2706,8 @@ __lock_release(struct lockdep_map *lock, int nested, unsigned long ip)
  */
 static void check_flags(unsigned long flags)
 {
-#if defined(CONFIG_DEBUG_LOCKDEP) && defined(CONFIG_TRACE_IRQFLAGS)
+#if defined(CONFIG_PROVE_LOCKING) && defined(CONFIG_DEBUG_LOCKDEP) && \
+    defined(CONFIG_TRACE_IRQFLAGS)
        if (!debug_locks)
                return;
 
index 8ce09bc4613dc1ea72566bb0a2a7822ac802930f..c3600a091a2874952e97c149f21080ede90ded95 100644 (file)
@@ -23,6 +23,8 @@
 #define MAX_LOCKDEP_CHAINS_BITS        14
 #define MAX_LOCKDEP_CHAINS     (1UL << MAX_LOCKDEP_CHAINS_BITS)
 
+#define MAX_LOCKDEP_CHAIN_HLOCKS (MAX_LOCKDEP_CHAINS*5)
+
 /*
  * Stack-trace: tightly packed array of stack backtrace
  * addresses. Protected by the hash_lock.
 #define MAX_STACK_TRACE_ENTRIES        262144UL
 
 extern struct list_head all_lock_classes;
+extern struct lock_chain lock_chains[];
 
 extern void
 get_usage_chars(struct lock_class *class, char *c1, char *c2, char *c3, char *c4);
 
 extern const char * __get_key_name(struct lockdep_subclass_key *key, char *str);
 
+struct lock_class *lock_chain_get_class(struct lock_chain *chain, int i);
+
 extern unsigned long nr_lock_classes;
 extern unsigned long nr_list_entries;
 extern unsigned long nr_lock_chains;
+extern int nr_chain_hlocks;
 extern unsigned long nr_stack_trace_entries;
 
 extern unsigned int nr_hardirq_chains;
index dc5d29648d852ccadd88ea79a91decc93653fa8b..9b0e940e2545efe99012dd1526ef79a4739f02d2 100644 (file)
@@ -139,7 +139,7 @@ static int l_show(struct seq_file *m, void *v)
 
        list_for_each_entry(entry, &class->locks_after, entry) {
                if (entry->distance == 1) {
-                       seq_printf(m, " -> [%p] ", entry->class);
+                       seq_printf(m, " -> [%p] ", entry->class->key);
                        print_name(m, entry->class);
                        seq_puts(m, "\n");
                }
@@ -178,6 +178,95 @@ static const struct file_operations proc_lockdep_operations = {
        .release        = seq_release,
 };
 
+#ifdef CONFIG_PROVE_LOCKING
+static void *lc_next(struct seq_file *m, void *v, loff_t *pos)
+{
+       struct lock_chain *chain;
+
+       (*pos)++;
+
+       if (v == SEQ_START_TOKEN)
+               chain = m->private;
+       else {
+               chain = v;
+
+               if (*pos < nr_lock_chains)
+                       chain = lock_chains + *pos;
+               else
+                       chain = NULL;
+       }
+
+       return chain;
+}
+
+static void *lc_start(struct seq_file *m, loff_t *pos)
+{
+       if (*pos == 0)
+               return SEQ_START_TOKEN;
+
+       if (*pos < nr_lock_chains)
+               return lock_chains + *pos;
+
+       return NULL;
+}
+
+static void lc_stop(struct seq_file *m, void *v)
+{
+}
+
+static int lc_show(struct seq_file *m, void *v)
+{
+       struct lock_chain *chain = v;
+       struct lock_class *class;
+       int i;
+
+       if (v == SEQ_START_TOKEN) {
+               seq_printf(m, "all lock chains:\n");
+               return 0;
+       }
+
+       seq_printf(m, "irq_context: %d\n", chain->irq_context);
+
+       for (i = 0; i < chain->depth; i++) {
+               class = lock_chain_get_class(chain, i);
+               seq_printf(m, "[%p] ", class->key);
+               print_name(m, class);
+               seq_puts(m, "\n");
+       }
+       seq_puts(m, "\n");
+
+       return 0;
+}
+
+static const struct seq_operations lockdep_chains_ops = {
+       .start  = lc_start,
+       .next   = lc_next,
+       .stop   = lc_stop,
+       .show   = lc_show,
+};
+
+static int lockdep_chains_open(struct inode *inode, struct file *file)
+{
+       int res = seq_open(file, &lockdep_chains_ops);
+       if (!res) {
+               struct seq_file *m = file->private_data;
+
+               if (nr_lock_chains)
+                       m->private = lock_chains;
+               else
+                       m->private = NULL;
+       }
+       return res;
+}
+
+static const struct file_operations proc_lockdep_chains_operations = {
+       .open           = lockdep_chains_open,
+       .read           = seq_read,
+       .llseek         = seq_lseek,
+       .release        = seq_release,
+};
+#endif /* CONFIG_PROVE_LOCKING */
+
 static void lockdep_stats_debug_show(struct seq_file *m)
 {
 #ifdef CONFIG_DEBUG_LOCKDEP
@@ -294,6 +383,8 @@ static int lockdep_stats_show(struct seq_file *m, void *v)
 #ifdef CONFIG_PROVE_LOCKING
        seq_printf(m, " dependency chains:             %11lu [max: %lu]\n",
                        nr_lock_chains, MAX_LOCKDEP_CHAINS);
+       seq_printf(m, " dependency chain hlocks:       %11d [max: %lu]\n",
+                       nr_chain_hlocks, MAX_LOCKDEP_CHAIN_HLOCKS);
 #endif
 
 #ifdef CONFIG_TRACE_IRQFLAGS
@@ -661,6 +752,10 @@ static const struct file_operations proc_lock_stat_operations = {
 static int __init lockdep_proc_init(void)
 {
        proc_create("lockdep", S_IRUSR, NULL, &proc_lockdep_operations);
+#ifdef CONFIG_PROVE_LOCKING
+       proc_create("lockdep_chains", S_IRUSR, NULL,
+                   &proc_lockdep_chains_operations);
+#endif
        proc_create("lockdep_stats", S_IRUSR, NULL,
                    &proc_lockdep_stats_operations);
 
index 3aaa06c561deba97838d6fab2cb66229e2a3e686..1d94160eb5326e94c0222da947b8c1ee4ee44875 100644 (file)
@@ -79,8 +79,8 @@ void debug_mutex_unlock(struct mutex *lock)
        if (unlikely(!debug_locks))
                return;
 
-       DEBUG_LOCKS_WARN_ON(lock->owner != current_thread_info());
        DEBUG_LOCKS_WARN_ON(lock->magic != lock);
+       DEBUG_LOCKS_WARN_ON(lock->owner != current_thread_info());
        DEBUG_LOCKS_WARN_ON(!lock->wait_list.prev && !lock->wait_list.next);
        DEBUG_LOCKS_WARN_ON(lock->owner != current_thread_info());
 }
index d046a345d365793e84076e2517ec02cdd11bc40d..bcdc9ac8ef60da5a2937ce8205e58ab2a2350deb 100644 (file)
@@ -165,10 +165,7 @@ __mutex_lock_common(struct mutex *lock, long state, unsigned int subclass,
                 * got a signal? (This code gets eliminated in the
                 * TASK_UNINTERRUPTIBLE case.)
                 */
-               if (unlikely((state == TASK_INTERRUPTIBLE &&
-                                       signal_pending(task)) ||
-                             (state == TASK_KILLABLE &&
-                                       fatal_signal_pending(task)))) {
+               if (unlikely(signal_pending_state(state, task))) {
                        mutex_remove_waiter(lock, &waiter,
                                            task_thread_info(task));
                        mutex_release(&lock->dep_map, 1, ip);