Loading arch/c6x/Kconfig +1 −0 Original line number Diff line number Diff line Loading @@ -18,6 +18,7 @@ config C6X select GENERIC_CLOCKEVENTS select MODULES_USE_ELF_RELA select ARCH_NO_COHERENT_DMA_MMAP select ARCH_HAS_RAW_COPY_USER config MMU def_bool n Loading arch/c6x/include/asm/uaccess.h +6 −6 Original line number Diff line number Diff line Loading @@ -14,12 +14,10 @@ #include <linux/string.h> /* * __copy_from_user/copy_to_user are based on ones in asm-generic/uaccess.h * * C6X supports unaligned 32 and 64 bit loads and stores. */ static inline __must_check long __copy_from_user(void *to, const void __user *from, unsigned long n) static inline __must_check unsigned long raw_copy_from_user(void *to, const void __user *from, unsigned long n) { u32 tmp32; u64 tmp64; Loading Loading @@ -54,8 +52,8 @@ static inline __must_check long __copy_from_user(void *to, return 0; } static inline __must_check long __copy_to_user(void __user *to, const void *from, unsigned long n) static inline __must_check unsigned long raw_copy_to_user(void __user *to, const void *from, unsigned long n) { u32 tmp32; u64 tmp64; Loading Loading @@ -89,6 +87,8 @@ static inline __must_check long __copy_to_user(void __user *to, memcpy((void __force *)to, from, n); return 0; } #define INLINE_COPY_FROM_USER #define INLINE_COPY_TO_USER extern int _access_ok(unsigned long addr, unsigned long size); #ifdef CONFIG_ACCESS_CHECK Loading Loading
arch/c6x/Kconfig +1 −0 Original line number Diff line number Diff line Loading @@ -18,6 +18,7 @@ config C6X select GENERIC_CLOCKEVENTS select MODULES_USE_ELF_RELA select ARCH_NO_COHERENT_DMA_MMAP select ARCH_HAS_RAW_COPY_USER config MMU def_bool n Loading
arch/c6x/include/asm/uaccess.h +6 −6 Original line number Diff line number Diff line Loading @@ -14,12 +14,10 @@ #include <linux/string.h> /* * __copy_from_user/copy_to_user are based on ones in asm-generic/uaccess.h * * C6X supports unaligned 32 and 64 bit loads and stores. */ static inline __must_check long __copy_from_user(void *to, const void __user *from, unsigned long n) static inline __must_check unsigned long raw_copy_from_user(void *to, const void __user *from, unsigned long n) { u32 tmp32; u64 tmp64; Loading Loading @@ -54,8 +52,8 @@ static inline __must_check long __copy_from_user(void *to, return 0; } static inline __must_check long __copy_to_user(void __user *to, const void *from, unsigned long n) static inline __must_check unsigned long raw_copy_to_user(void __user *to, const void *from, unsigned long n) { u32 tmp32; u64 tmp64; Loading Loading @@ -89,6 +87,8 @@ static inline __must_check long __copy_to_user(void __user *to, memcpy((void __force *)to, from, n); return 0; } #define INLINE_COPY_FROM_USER #define INLINE_COPY_TO_USER extern int _access_ok(unsigned long addr, unsigned long size); #ifdef CONFIG_ACCESS_CHECK Loading