Merge to Fedora kernel-2.6.18-1.2224_FC5 patched with stable patch-2.6.18.1-vs2.0...
[linux-2.6.git] / arch / i386 / lib / usercopy.c
index e1bcec2..efc7e7d 100644 (file)
@@ -5,7 +5,6 @@
  * Copyright 1997 Andi Kleen <ak@muc.de>
  * Copyright 1997 Linus Torvalds
  */
-#include <linux/config.h>
 #include <linux/mm.h>
 #include <linux/highmem.h>
 #include <linux/blkdev.h>
@@ -31,6 +30,7 @@ static inline int __movsl_is_ok(unsigned long a1, unsigned long a2, unsigned lon
 #define __do_strncpy_from_user(dst,src,count,res)                         \
 do {                                                                      \
        int __d0, __d1, __d2;                                              \
+       might_sleep();                                                     \
        __asm__ __volatile__(                                              \
                "       testl %1,%1\n"                                     \
                "       jz 2f\n"                                           \
@@ -83,6 +83,7 @@ __strncpy_from_user(char *dst, const char __user *src, long count)
        __do_strncpy_from_user(dst, src, count, res);
        return res;
 }
+EXPORT_SYMBOL(__strncpy_from_user);
 
 /**
  * strncpy_from_user: - Copy a NUL terminated string from userspace.
@@ -110,7 +111,7 @@ strncpy_from_user(char *dst, const char __user *src, long count)
                __do_strncpy_from_user(dst, src, count, res);
        return res;
 }
-
+EXPORT_SYMBOL(strncpy_from_user);
 
 /*
  * Zero Userspace
@@ -119,6 +120,7 @@ strncpy_from_user(char *dst, const char __user *src, long count)
 #define __do_clear_user(addr,size)                                     \
 do {                                                                   \
        int __d0;                                                       \
+       might_sleep();                                                  \
        __asm__ __volatile__(                                           \
                "0:     rep; stosl\n"                                   \
                "       movl %2,%0\n"                                   \
@@ -155,6 +157,7 @@ clear_user(void __user *to, unsigned long n)
                __do_clear_user(to, n);
        return n;
 }
+EXPORT_SYMBOL(clear_user);
 
 /**
  * __clear_user: - Zero a block of memory in user space, with less checking.
@@ -173,6 +176,7 @@ __clear_user(void __user *to, unsigned long n)
        __do_clear_user(to, n);
        return n;
 }
+EXPORT_SYMBOL(__clear_user);
 
 /**
  * strlen_user: - Get the size of a string in user space.
@@ -216,6 +220,7 @@ long strnlen_user(const char __user *s, long n)
                :"cc");
        return res & mask;
 }
+EXPORT_SYMBOL(strnlen_user);
 
 #ifdef CONFIG_X86_INTEL_USERCOPY
 static unsigned long
@@ -419,15 +424,212 @@ __copy_user_zeroing_intel(void *to, const void __user *from, unsigned long size)
                       : "eax", "edx", "memory");
        return size;
 }
+
+/*
+ * Non Temporal Hint version of __copy_user_zeroing_intel.  It is cache aware.
+ * hyoshiok@miraclelinux.com
+ */
+
+static unsigned long __copy_user_zeroing_intel_nocache(void *to,
+                               const void __user *from, unsigned long size)
+{
+        int d0, d1;
+
+       __asm__ __volatile__(
+              "        .align 2,0x90\n"
+              "0:      movl 32(%4), %%eax\n"
+              "        cmpl $67, %0\n"
+              "        jbe 2f\n"
+              "1:      movl 64(%4), %%eax\n"
+              "        .align 2,0x90\n"
+              "2:      movl 0(%4), %%eax\n"
+              "21:     movl 4(%4), %%edx\n"
+              "        movnti %%eax, 0(%3)\n"
+              "        movnti %%edx, 4(%3)\n"
+              "3:      movl 8(%4), %%eax\n"
+              "31:     movl 12(%4),%%edx\n"
+              "        movnti %%eax, 8(%3)\n"
+              "        movnti %%edx, 12(%3)\n"
+              "4:      movl 16(%4), %%eax\n"
+              "41:     movl 20(%4), %%edx\n"
+              "        movnti %%eax, 16(%3)\n"
+              "        movnti %%edx, 20(%3)\n"
+              "10:     movl 24(%4), %%eax\n"
+              "51:     movl 28(%4), %%edx\n"
+              "        movnti %%eax, 24(%3)\n"
+              "        movnti %%edx, 28(%3)\n"
+              "11:     movl 32(%4), %%eax\n"
+              "61:     movl 36(%4), %%edx\n"
+              "        movnti %%eax, 32(%3)\n"
+              "        movnti %%edx, 36(%3)\n"
+              "12:     movl 40(%4), %%eax\n"
+              "71:     movl 44(%4), %%edx\n"
+              "        movnti %%eax, 40(%3)\n"
+              "        movnti %%edx, 44(%3)\n"
+              "13:     movl 48(%4), %%eax\n"
+              "81:     movl 52(%4), %%edx\n"
+              "        movnti %%eax, 48(%3)\n"
+              "        movnti %%edx, 52(%3)\n"
+              "14:     movl 56(%4), %%eax\n"
+              "91:     movl 60(%4), %%edx\n"
+              "        movnti %%eax, 56(%3)\n"
+              "        movnti %%edx, 60(%3)\n"
+              "        addl $-64, %0\n"
+              "        addl $64, %4\n"
+              "        addl $64, %3\n"
+              "        cmpl $63, %0\n"
+              "        ja  0b\n"
+              "        sfence \n"
+              "5:      movl  %0, %%eax\n"
+              "        shrl  $2, %0\n"
+              "        andl $3, %%eax\n"
+              "        cld\n"
+              "6:      rep; movsl\n"
+              "        movl %%eax,%0\n"
+              "7:      rep; movsb\n"
+              "8:\n"
+              ".section .fixup,\"ax\"\n"
+              "9:      lea 0(%%eax,%0,4),%0\n"
+              "16:     pushl %0\n"
+              "        pushl %%eax\n"
+              "        xorl %%eax,%%eax\n"
+              "        rep; stosb\n"
+              "        popl %%eax\n"
+              "        popl %0\n"
+              "        jmp 8b\n"
+              ".previous\n"
+              ".section __ex_table,\"a\"\n"
+              "        .align 4\n"
+              "        .long 0b,16b\n"
+              "        .long 1b,16b\n"
+              "        .long 2b,16b\n"
+              "        .long 21b,16b\n"
+              "        .long 3b,16b\n"
+              "        .long 31b,16b\n"
+              "        .long 4b,16b\n"
+              "        .long 41b,16b\n"
+              "        .long 10b,16b\n"
+              "        .long 51b,16b\n"
+              "        .long 11b,16b\n"
+              "        .long 61b,16b\n"
+              "        .long 12b,16b\n"
+              "        .long 71b,16b\n"
+              "        .long 13b,16b\n"
+              "        .long 81b,16b\n"
+              "        .long 14b,16b\n"
+              "        .long 91b,16b\n"
+              "        .long 6b,9b\n"
+              "        .long 7b,16b\n"
+              ".previous"
+              : "=&c"(size), "=&D" (d0), "=&S" (d1)
+              :  "1"(to), "2"(from), "0"(size)
+              : "eax", "edx", "memory");
+       return size;
+}
+
+static unsigned long __copy_user_intel_nocache(void *to,
+                               const void __user *from, unsigned long size)
+{
+        int d0, d1;
+
+       __asm__ __volatile__(
+              "        .align 2,0x90\n"
+              "0:      movl 32(%4), %%eax\n"
+              "        cmpl $67, %0\n"
+              "        jbe 2f\n"
+              "1:      movl 64(%4), %%eax\n"
+              "        .align 2,0x90\n"
+              "2:      movl 0(%4), %%eax\n"
+              "21:     movl 4(%4), %%edx\n"
+              "        movnti %%eax, 0(%3)\n"
+              "        movnti %%edx, 4(%3)\n"
+              "3:      movl 8(%4), %%eax\n"
+              "31:     movl 12(%4),%%edx\n"
+              "        movnti %%eax, 8(%3)\n"
+              "        movnti %%edx, 12(%3)\n"
+              "4:      movl 16(%4), %%eax\n"
+              "41:     movl 20(%4), %%edx\n"
+              "        movnti %%eax, 16(%3)\n"
+              "        movnti %%edx, 20(%3)\n"
+              "10:     movl 24(%4), %%eax\n"
+              "51:     movl 28(%4), %%edx\n"
+              "        movnti %%eax, 24(%3)\n"
+              "        movnti %%edx, 28(%3)\n"
+              "11:     movl 32(%4), %%eax\n"
+              "61:     movl 36(%4), %%edx\n"
+              "        movnti %%eax, 32(%3)\n"
+              "        movnti %%edx, 36(%3)\n"
+              "12:     movl 40(%4), %%eax\n"
+              "71:     movl 44(%4), %%edx\n"
+              "        movnti %%eax, 40(%3)\n"
+              "        movnti %%edx, 44(%3)\n"
+              "13:     movl 48(%4), %%eax\n"
+              "81:     movl 52(%4), %%edx\n"
+              "        movnti %%eax, 48(%3)\n"
+              "        movnti %%edx, 52(%3)\n"
+              "14:     movl 56(%4), %%eax\n"
+              "91:     movl 60(%4), %%edx\n"
+              "        movnti %%eax, 56(%3)\n"
+              "        movnti %%edx, 60(%3)\n"
+              "        addl $-64, %0\n"
+              "        addl $64, %4\n"
+              "        addl $64, %3\n"
+              "        cmpl $63, %0\n"
+              "        ja  0b\n"
+              "        sfence \n"
+              "5:      movl  %0, %%eax\n"
+              "        shrl  $2, %0\n"
+              "        andl $3, %%eax\n"
+              "        cld\n"
+              "6:      rep; movsl\n"
+              "        movl %%eax,%0\n"
+              "7:      rep; movsb\n"
+              "8:\n"
+              ".section .fixup,\"ax\"\n"
+              "9:      lea 0(%%eax,%0,4),%0\n"
+              "16:     jmp 8b\n"
+              ".previous\n"
+              ".section __ex_table,\"a\"\n"
+              "        .align 4\n"
+              "        .long 0b,16b\n"
+              "        .long 1b,16b\n"
+              "        .long 2b,16b\n"
+              "        .long 21b,16b\n"
+              "        .long 3b,16b\n"
+              "        .long 31b,16b\n"
+              "        .long 4b,16b\n"
+              "        .long 41b,16b\n"
+              "        .long 10b,16b\n"
+              "        .long 51b,16b\n"
+              "        .long 11b,16b\n"
+              "        .long 61b,16b\n"
+              "        .long 12b,16b\n"
+              "        .long 71b,16b\n"
+              "        .long 13b,16b\n"
+              "        .long 81b,16b\n"
+              "        .long 14b,16b\n"
+              "        .long 91b,16b\n"
+              "        .long 6b,9b\n"
+              "        .long 7b,16b\n"
+              ".previous"
+              : "=&c"(size), "=&D" (d0), "=&S" (d1)
+              :  "1"(to), "2"(from), "0"(size)
+              : "eax", "edx", "memory");
+       return size;
+}
+
 #else
+
 /*
  * Leave these declared but undefined.  They should not be any references to
  * them
  */
-unsigned long
-__copy_user_zeroing_intel(void *to, const void __user *from, unsigned long size);
-unsigned long
-__copy_user_intel(void __user *to, const void *from, unsigned long size);
+unsigned long __copy_user_zeroing_intel(void *to, const void __user *from,
+                                       unsigned long size);
+unsigned long __copy_user_intel(void __user *to, const void *from,
+                                       unsigned long size);
+unsigned long __copy_user_zeroing_intel_nocache(void *to,
+                               const void __user *from, unsigned long size);
 #endif /* CONFIG_X86_INTEL_USERCOPY */
 
 /* Generic arbitrary sized copy.  */
@@ -509,9 +711,10 @@ do {                                                                       \
                : "memory");                                            \
 } while (0)
 
-
-unsigned long __copy_to_user_ll(void __user *to, const void *from, unsigned long n)
+unsigned long __copy_to_user_ll(void __user *to, const void *from,
+                               unsigned long n)
 {
+       BUG_ON((long) n < 0);
 #ifndef CONFIG_X86_WP_WORKS_OK
        if (unlikely(boot_cpu_data.wp_works_ok == 0) &&
                        ((unsigned long )to) < TASK_SIZE) {
@@ -567,16 +770,62 @@ survive:
                n = __copy_user_intel(to, from, n);
        return n;
 }
+EXPORT_SYMBOL(__copy_to_user_ll);
 
-unsigned long
-__copy_from_user_ll(void *to, const void __user *from, unsigned long n)
+unsigned long __copy_from_user_ll(void *to, const void __user *from,
+                                       unsigned long n)
 {
+       BUG_ON((long)n < 0);
        if (movsl_is_ok(to, from, n))
                __copy_user_zeroing(to, from, n);
        else
                n = __copy_user_zeroing_intel(to, from, n);
        return n;
 }
+EXPORT_SYMBOL(__copy_from_user_ll);
+
+unsigned long __copy_from_user_ll_nozero(void *to, const void __user *from,
+                                        unsigned long n)
+{
+       BUG_ON((long)n < 0);
+       if (movsl_is_ok(to, from, n))
+               __copy_user(to, from, n);
+       else
+               n = __copy_user_intel((void __user *)to,
+                                     (const void *)from, n);
+       return n;
+}
+EXPORT_SYMBOL(__copy_from_user_ll_nozero);
+
+unsigned long __copy_from_user_ll_nocache(void *to, const void __user *from,
+                                       unsigned long n)
+{
+       BUG_ON((long)n < 0);
+#ifdef CONFIG_X86_INTEL_USERCOPY
+       if ( n > 64 && cpu_has_xmm2)
+                n = __copy_user_zeroing_intel_nocache(to, from, n);
+       else
+               __copy_user_zeroing(to, from, n);
+#else
+        __copy_user_zeroing(to, from, n);
+#endif
+       return n;
+}
+
+unsigned long __copy_from_user_ll_nocache_nozero(void *to, const void __user *from,
+                                       unsigned long n)
+{
+       BUG_ON((long)n < 0);
+#ifdef CONFIG_X86_INTEL_USERCOPY
+       if ( n > 64 && cpu_has_xmm2)
+                n = __copy_user_intel_nocache(to, from, n);
+       else
+               __copy_user(to, from, n);
+#else
+        __copy_user(to, from, n);
+#endif
+       return n;
+}
 
 /**
  * copy_to_user: - Copy a block of data into user space.
@@ -594,7 +843,7 @@ __copy_from_user_ll(void *to, const void __user *from, unsigned long n)
 unsigned long
 copy_to_user(void __user *to, const void *from, unsigned long n)
 {
-       might_sleep();
+       BUG_ON((long) n < 0);
        if (access_ok(VERIFY_WRITE, to, n))
                n = __copy_to_user(to, from, n);
        return n;
@@ -620,7 +869,7 @@ EXPORT_SYMBOL(copy_to_user);
 unsigned long
 copy_from_user(void *to, const void __user *from, unsigned long n)
 {
-       might_sleep();
+       BUG_ON((long) n < 0);
        if (access_ok(VERIFY_READ, from, n))
                n = __copy_from_user(to, from, n);
        else