From: Ingo Molnar Temporary workaround for the lock validator: make all uses of sk_callback_lock softirq-safe. (The real solution will be to express to the lock validator that sk_callback_lock rules are to be generated per-address-family.) Signed-off-by: Ingo Molnar Signed-off-by: Arjan van de Ven Signed-off-by: Andrew Morton --- net/core/sock.c | 24 ++++++++++++------------ 1 file changed, 12 insertions(+), 12 deletions(-) diff -puN net/core/sock.c~lock-validator-sk_callback_lock-workaround net/core/sock.c --- devel/net/core/sock.c~lock-validator-sk_callback_lock-workaround 2006-05-29 18:12:44.000000000 -0700 +++ devel-akpm/net/core/sock.c 2006-05-29 18:12:44.000000000 -0700 @@ -937,9 +937,9 @@ int sock_i_uid(struct sock *sk) { int uid; - read_lock(&sk->sk_callback_lock); + read_lock_bh(&sk->sk_callback_lock); uid = sk->sk_socket ? SOCK_INODE(sk->sk_socket)->i_uid : 0; - read_unlock(&sk->sk_callback_lock); + read_unlock_bh(&sk->sk_callback_lock); return uid; } @@ -947,9 +947,9 @@ unsigned long sock_i_ino(struct sock *sk { unsigned long ino; - read_lock(&sk->sk_callback_lock); + read_lock_bh(&sk->sk_callback_lock); ino = sk->sk_socket ? SOCK_INODE(sk->sk_socket)->i_ino : 0; - read_unlock(&sk->sk_callback_lock); + read_unlock_bh(&sk->sk_callback_lock); return ino; } @@ -1309,33 +1309,33 @@ ssize_t sock_no_sendpage(struct socket * static void sock_def_wakeup(struct sock *sk) { - read_lock(&sk->sk_callback_lock); + read_lock_bh(&sk->sk_callback_lock); if (sk->sk_sleep && waitqueue_active(sk->sk_sleep)) wake_up_interruptible_all(sk->sk_sleep); - read_unlock(&sk->sk_callback_lock); + read_unlock_bh(&sk->sk_callback_lock); } static void sock_def_error_report(struct sock *sk) { - read_lock(&sk->sk_callback_lock); + read_lock_bh(&sk->sk_callback_lock); if (sk->sk_sleep && waitqueue_active(sk->sk_sleep)) wake_up_interruptible(sk->sk_sleep); sk_wake_async(sk,0,POLL_ERR); - read_unlock(&sk->sk_callback_lock); + read_unlock_bh(&sk->sk_callback_lock); } static void sock_def_readable(struct sock *sk, int len) { - read_lock(&sk->sk_callback_lock); + read_lock_bh(&sk->sk_callback_lock); if (sk->sk_sleep && waitqueue_active(sk->sk_sleep)) wake_up_interruptible(sk->sk_sleep); sk_wake_async(sk,1,POLL_IN); - read_unlock(&sk->sk_callback_lock); + read_unlock_bh(&sk->sk_callback_lock); } static void sock_def_write_space(struct sock *sk) { - read_lock(&sk->sk_callback_lock); + read_lock_bh(&sk->sk_callback_lock); /* Do not wake up a writer until he can make "significant" * progress. --DaveM @@ -1349,7 +1349,7 @@ static void sock_def_write_space(struct sk_wake_async(sk, 2, POLL_OUT); } - read_unlock(&sk->sk_callback_lock); + read_unlock_bh(&sk->sk_callback_lock); } static void sock_def_destruct(struct sock *sk) _