x86: some lock annotations for user copy paths
authorNick Piggin <npiggin@suse.de>
Wed, 10 Sep 2008 11:37:17 +0000 (13:37 +0200)
committerIngo Molnar <mingo@elte.hu>
Wed, 10 Sep 2008 11:48:49 +0000 (13:48 +0200)
copy_to/from_user and all its variants (except the atomic ones) can take a
page fault and perform non-trivial work like taking mmap_sem and entering
the filesyste/pagecache.

Unfortunately, this often escapes lockdep because a common pattern is to
use it to read in some arguments just set up from userspace, or write data
back to a hot buffer. In those cases, it will be unlikely for page reclaim
to get a window in to cause copy_*_user to fault.

With the new might_lock primitives, add some annotations to x86. I don't
know if I caught all possible faulting points (it's a bit of a maze, and I
didn't really look at 32-bit). But this is a starting point.

Boots and runs OK so far.

Signed-off-by: Nick Piggin <npiggin@suse.de>
Acked-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
arch/x86/lib/usercopy_32.c
arch/x86/lib/usercopy_64.c
include/asm-x86/uaccess.h
include/asm-x86/uaccess_32.h
include/asm-x86/uaccess_64.h

index 24e60944971a60522d615bf9dc912851beb32b86..8eedde2a9cac614f97580d61e03b97662ad3e762 100644 (file)
@@ -33,6 +33,8 @@ static inline int __movsl_is_ok(unsigned long a1, unsigned long a2, unsigned lon
 do {                                                                      \
        int __d0, __d1, __d2;                                              \
        might_sleep();                                                     \
+       if (current->mm)                                                   \
+               might_lock_read(&current->mm->mmap_sem);                   \
        __asm__ __volatile__(                                              \
                "       testl %1,%1\n"                                     \
                "       jz 2f\n"                                           \
@@ -120,6 +122,8 @@ EXPORT_SYMBOL(strncpy_from_user);
 do {                                                                   \
        int __d0;                                                       \
        might_sleep();                                                  \
+       if (current->mm)                                                \
+               might_lock_read(&current->mm->mmap_sem);                \
        __asm__ __volatile__(                                           \
                "0:     rep; stosl\n"                                   \
                "       movl %2,%0\n"                                   \
@@ -148,7 +152,6 @@ do {                                                                        \
 unsigned long
 clear_user(void __user *to, unsigned long n)
 {
-       might_sleep();
        if (access_ok(VERIFY_WRITE, to, n))
                __do_clear_user(to, n);
        return n;
@@ -191,6 +194,8 @@ long strnlen_user(const char __user *s, long n)
        unsigned long res, tmp;
 
        might_sleep();
+       if (current->mm)
+               might_lock_read(&current->mm->mmap_sem);
 
        __asm__ __volatile__(
                "       testl %0, %0\n"
index f4df6e7c718be506a59ef157bda9591dfb1790a6..847d12945998a99c3bca081aa89a56940b911621 100644 (file)
@@ -16,6 +16,8 @@
 do {                                                                      \
        long __d0, __d1, __d2;                                             \
        might_sleep();                                                     \
+       if (current->mm)                                                   \
+               might_lock_read(&current->mm->mmap_sem);                   \
        __asm__ __volatile__(                                              \
                "       testq %1,%1\n"                                     \
                "       jz 2f\n"                                           \
@@ -65,6 +67,8 @@ unsigned long __clear_user(void __user *addr, unsigned long size)
 {
        long __d0;
        might_sleep();
+       if (current->mm)
+               might_lock_read(&current->mm->mmap_sem);
        /* no memory constraint because it doesn't change any memory gcc knows
           about */
        asm volatile(
index 5f702d1d52184424941d3238ee82597c9f05ce37..ad29752a171338a94a6e57a437ebb7bfd9e62ecd 100644 (file)
@@ -8,6 +8,8 @@
 #include <linux/thread_info.h>
 #include <linux/prefetch.h>
 #include <linux/string.h>
+#include <linux/lockdep.h>
+#include <linux/sched.h>
 #include <asm/asm.h>
 #include <asm/page.h>
 
@@ -157,6 +159,9 @@ extern int __get_user_bad(void);
        int __ret_gu;                                                   \
        unsigned long __val_gu;                                         \
        __chk_user_ptr(ptr);                                            \
+       might_sleep();                                                  \
+       if (current->mm)                                                \
+               might_lock_read(&current->mm->mmap_sem);                \
        switch (sizeof(*(ptr))) {                                       \
        case 1:                                                         \
                __get_user_x(1, __ret_gu, __val_gu, ptr);               \
@@ -241,6 +246,9 @@ extern void __put_user_8(void);
        int __ret_pu;                                           \
        __typeof__(*(ptr)) __pu_val;                            \
        __chk_user_ptr(ptr);                                    \
+       might_sleep();                                          \
+       if (current->mm)                                        \
+               might_lock_read(&current->mm->mmap_sem);        \
        __pu_val = x;                                           \
        switch (sizeof(*(ptr))) {                               \
        case 1:                                                 \
@@ -265,6 +273,9 @@ extern void __put_user_8(void);
 #define __put_user_size(x, ptr, size, retval, errret)                  \
 do {                                                                   \
        retval = 0;                                                     \
+       might_sleep();                                                  \
+       if (current->mm)                                                \
+               might_lock_read(&current->mm->mmap_sem);                \
        __chk_user_ptr(ptr);                                            \
        switch (size) {                                                 \
        case 1:                                                         \
@@ -317,6 +328,9 @@ do {                                                                        \
 #define __get_user_size(x, ptr, size, retval, errret)                  \
 do {                                                                   \
        retval = 0;                                                     \
+       might_sleep();                                                  \
+       if (current->mm)                                                \
+               might_lock_read(&current->mm->mmap_sem);                \
        __chk_user_ptr(ptr);                                            \
        switch (size) {                                                 \
        case 1:                                                         \
index 6fdef39a0bcb849d86e3bb0e1ade448694d21cdb..d725e2d703f7dfa706ed9da3f31053253cb88ffc 100644 (file)
@@ -82,8 +82,10 @@ __copy_to_user_inatomic(void __user *to, const void *from, unsigned long n)
 static __always_inline unsigned long __must_check
 __copy_to_user(void __user *to, const void *from, unsigned long n)
 {
-       might_sleep();
-       return __copy_to_user_inatomic(to, from, n);
+       might_sleep();
+       if (current->mm)
+               might_lock_read(&current->mm->mmap_sem);
+       return __copy_to_user_inatomic(to, from, n);
 }
 
 static __always_inline unsigned long
@@ -138,6 +140,8 @@ static __always_inline unsigned long
 __copy_from_user(void *to, const void __user *from, unsigned long n)
 {
        might_sleep();
+       if (current->mm)
+               might_lock_read(&current->mm->mmap_sem);
        if (__builtin_constant_p(n)) {
                unsigned long ret;
 
@@ -160,6 +164,8 @@ static __always_inline unsigned long __copy_from_user_nocache(void *to,
                                const void __user *from, unsigned long n)
 {
        might_sleep();
+       if (current->mm)
+               might_lock_read(&current->mm->mmap_sem);
        if (__builtin_constant_p(n)) {
                unsigned long ret;
 
index 515d4dce96b598bc6e9d07dba21332a44924948c..40a7205fe576744135cb4210b4dcce5634f9874b 100644 (file)
@@ -28,6 +28,10 @@ static __always_inline __must_check
 int __copy_from_user(void *dst, const void __user *src, unsigned size)
 {
        int ret = 0;
+
+       might_sleep();
+       if (current->mm)
+               might_lock_read(&current->mm->mmap_sem);
        if (!__builtin_constant_p(size))
                return copy_user_generic(dst, (__force void *)src, size);
        switch (size) {
@@ -70,6 +74,10 @@ static __always_inline __must_check
 int __copy_to_user(void __user *dst, const void *src, unsigned size)
 {
        int ret = 0;
+
+       might_sleep();
+       if (current->mm)
+               might_lock_read(&current->mm->mmap_sem);
        if (!__builtin_constant_p(size))
                return copy_user_generic((__force void *)dst, src, size);
        switch (size) {
@@ -112,6 +120,10 @@ static __always_inline __must_check
 int __copy_in_user(void __user *dst, const void __user *src, unsigned size)
 {
        int ret = 0;
+
+       might_sleep();
+       if (current->mm)
+               might_lock_read(&current->mm->mmap_sem);
        if (!__builtin_constant_p(size))
                return copy_user_generic((__force void *)dst,
                                         (__force void *)src, size);