7bcb63b213
For 64bits arguments gcc caused that put_user macro works with wrong value because of optimalization. Adding volatile caused that gcc not optimized it. It is possible to use (as Blackfin do) two put_user macros with 32bits arguments but there is one more instruction which is due to duplication zero return value which is called put_user_asm macro. Signed-off-by: Michal Simek <monstr@monstr.eu>
328 lines
8.7 KiB
C
328 lines
8.7 KiB
C
/*
|
|
* Copyright (C) 2008-2009 Michal Simek <monstr@monstr.eu>
|
|
* Copyright (C) 2008-2009 PetaLogix
|
|
* Copyright (C) 2006 Atmark Techno, Inc.
|
|
*
|
|
* This file is subject to the terms and conditions of the GNU General Public
|
|
* License. See the file "COPYING" in the main directory of this archive
|
|
* for more details.
|
|
*/
|
|
|
|
#ifndef _ASM_MICROBLAZE_UACCESS_H
|
|
#define _ASM_MICROBLAZE_UACCESS_H
|
|
|
|
#ifdef __KERNEL__
|
|
#ifndef __ASSEMBLY__
|
|
|
|
#include <linux/kernel.h>
|
|
#include <linux/errno.h>
|
|
#include <linux/sched.h> /* RLIMIT_FSIZE */
|
|
#include <linux/mm.h>
|
|
|
|
#include <asm/mmu.h>
|
|
#include <asm/page.h>
|
|
#include <asm/pgtable.h>
|
|
#include <asm/segment.h>
|
|
#include <linux/string.h>
|
|
|
|
#define VERIFY_READ 0
|
|
#define VERIFY_WRITE 1
|
|
|
|
#define __clear_user(addr, n) (memset((void *)(addr), 0, (n)), 0)
|
|
|
|
#ifndef CONFIG_MMU
|
|
|
|
extern int ___range_ok(unsigned long addr, unsigned long size);
|
|
|
|
#define __range_ok(addr, size) \
|
|
___range_ok((unsigned long)(addr), (unsigned long)(size))
|
|
|
|
#define access_ok(type, addr, size) (__range_ok((addr), (size)) == 0)
|
|
#define __access_ok(add, size) (__range_ok((addr), (size)) == 0)
|
|
|
|
/* Undefined function to trigger linker error */
|
|
extern int bad_user_access_length(void);
|
|
|
|
/* FIXME this is function for optimalization -> memcpy */
|
|
#define __get_user(var, ptr) \
|
|
({ \
|
|
int __gu_err = 0; \
|
|
switch (sizeof(*(ptr))) { \
|
|
case 1: \
|
|
case 2: \
|
|
case 4: \
|
|
(var) = *(ptr); \
|
|
break; \
|
|
case 8: \
|
|
memcpy((void *) &(var), (ptr), 8); \
|
|
break; \
|
|
default: \
|
|
(var) = 0; \
|
|
__gu_err = __get_user_bad(); \
|
|
break; \
|
|
} \
|
|
__gu_err; \
|
|
})
|
|
|
|
#define __get_user_bad() (bad_user_access_length(), (-EFAULT))
|
|
|
|
/* FIXME is not there defined __pu_val */
|
|
#define __put_user(var, ptr) \
|
|
({ \
|
|
int __pu_err = 0; \
|
|
switch (sizeof(*(ptr))) { \
|
|
case 1: \
|
|
case 2: \
|
|
case 4: \
|
|
*(ptr) = (var); \
|
|
break; \
|
|
case 8: { \
|
|
typeof(*(ptr)) __pu_val = (var); \
|
|
memcpy(ptr, &__pu_val, sizeof(__pu_val)); \
|
|
} \
|
|
break; \
|
|
default: \
|
|
__pu_err = __put_user_bad(); \
|
|
break; \
|
|
} \
|
|
__pu_err; \
|
|
})
|
|
|
|
#define __put_user_bad() (bad_user_access_length(), (-EFAULT))
|
|
|
|
#define put_user(x, ptr) __put_user((x), (ptr))
|
|
#define get_user(x, ptr) __get_user((x), (ptr))
|
|
|
|
#define copy_to_user(to, from, n) (memcpy((to), (from), (n)), 0)
|
|
#define copy_from_user(to, from, n) (memcpy((to), (from), (n)), 0)
|
|
|
|
#define __copy_to_user(to, from, n) (copy_to_user((to), (from), (n)))
|
|
#define __copy_from_user(to, from, n) (copy_from_user((to), (from), (n)))
|
|
#define __copy_to_user_inatomic(to, from, n) \
|
|
(__copy_to_user((to), (from), (n)))
|
|
#define __copy_from_user_inatomic(to, from, n) \
|
|
(__copy_from_user((to), (from), (n)))
|
|
|
|
static inline unsigned long clear_user(void *addr, unsigned long size)
|
|
{
|
|
if (access_ok(VERIFY_WRITE, addr, size))
|
|
size = __clear_user(addr, size);
|
|
return size;
|
|
}
|
|
|
|
/* Returns 0 if exception not found and fixup otherwise. */
|
|
extern unsigned long search_exception_table(unsigned long);
|
|
|
|
extern long strncpy_from_user(char *dst, const char *src, long count);
|
|
extern long strnlen_user(const char *src, long count);
|
|
|
|
#else /* CONFIG_MMU */
|
|
|
|
/*
|
|
* Address is valid if:
|
|
* - "addr", "addr + size" and "size" are all below the limit
|
|
*/
|
|
#define access_ok(type, addr, size) \
|
|
(get_fs().seg > (((unsigned long)(addr)) | \
|
|
(size) | ((unsigned long)(addr) + (size))))
|
|
|
|
/* || printk("access_ok failed for %s at 0x%08lx (size %d), seg 0x%08x\n",
|
|
type?"WRITE":"READ",addr,size,get_fs().seg)) */
|
|
|
|
/*
|
|
* All the __XXX versions macros/functions below do not perform
|
|
* access checking. It is assumed that the necessary checks have been
|
|
* already performed before the finction (macro) is called.
|
|
*/
|
|
|
|
#define get_user(x, ptr) \
|
|
({ \
|
|
access_ok(VERIFY_READ, (ptr), sizeof(*(ptr))) \
|
|
? __get_user((x), (ptr)) : -EFAULT; \
|
|
})
|
|
|
|
#define put_user(x, ptr) \
|
|
({ \
|
|
access_ok(VERIFY_WRITE, (ptr), sizeof(*(ptr))) \
|
|
? __put_user((x), (ptr)) : -EFAULT; \
|
|
})
|
|
|
|
#define __get_user(x, ptr) \
|
|
({ \
|
|
unsigned long __gu_val; \
|
|
/*unsigned long __gu_ptr = (unsigned long)(ptr);*/ \
|
|
long __gu_err; \
|
|
switch (sizeof(*(ptr))) { \
|
|
case 1: \
|
|
__get_user_asm("lbu", (ptr), __gu_val, __gu_err); \
|
|
break; \
|
|
case 2: \
|
|
__get_user_asm("lhu", (ptr), __gu_val, __gu_err); \
|
|
break; \
|
|
case 4: \
|
|
__get_user_asm("lw", (ptr), __gu_val, __gu_err); \
|
|
break; \
|
|
default: \
|
|
__gu_val = 0; __gu_err = -EINVAL; \
|
|
} \
|
|
x = (__typeof__(*(ptr))) __gu_val; \
|
|
__gu_err; \
|
|
})
|
|
|
|
#define __get_user_asm(insn, __gu_ptr, __gu_val, __gu_err) \
|
|
({ \
|
|
__asm__ __volatile__ ( \
|
|
"1:" insn " %1, %2, r0; \
|
|
addk %0, r0, r0; \
|
|
2: \
|
|
.section .fixup,\"ax\"; \
|
|
3: brid 2b; \
|
|
addik %0, r0, %3; \
|
|
.previous; \
|
|
.section __ex_table,\"a\"; \
|
|
.word 1b,3b; \
|
|
.previous;" \
|
|
: "=r"(__gu_err), "=r"(__gu_val) \
|
|
: "r"(__gu_ptr), "i"(-EFAULT) \
|
|
); \
|
|
})
|
|
|
|
#define __put_user(x, ptr) \
|
|
({ \
|
|
__typeof__(*(ptr)) volatile __gu_val = (x); \
|
|
long __gu_err = 0; \
|
|
switch (sizeof(__gu_val)) { \
|
|
case 1: \
|
|
__put_user_asm("sb", (ptr), __gu_val, __gu_err); \
|
|
break; \
|
|
case 2: \
|
|
__put_user_asm("sh", (ptr), __gu_val, __gu_err); \
|
|
break; \
|
|
case 4: \
|
|
__put_user_asm("sw", (ptr), __gu_val, __gu_err); \
|
|
break; \
|
|
case 8: \
|
|
__put_user_asm_8((ptr), __gu_val, __gu_err); \
|
|
break; \
|
|
default: \
|
|
__gu_err = -EINVAL; \
|
|
} \
|
|
__gu_err; \
|
|
})
|
|
|
|
#define __put_user_asm_8(__gu_ptr, __gu_val, __gu_err) \
|
|
({ \
|
|
__asm__ __volatile__ (" lwi %0, %1, 0; \
|
|
1: swi %0, %2, 0; \
|
|
lwi %0, %1, 4; \
|
|
2: swi %0, %2, 4; \
|
|
addk %0,r0,r0; \
|
|
3: \
|
|
.section .fixup,\"ax\"; \
|
|
4: brid 3b; \
|
|
addik %0, r0, %3; \
|
|
.previous; \
|
|
.section __ex_table,\"a\"; \
|
|
.word 1b,4b,2b,4b; \
|
|
.previous;" \
|
|
: "=&r"(__gu_err) \
|
|
: "r"(&__gu_val), \
|
|
"r"(__gu_ptr), "i"(-EFAULT) \
|
|
); \
|
|
})
|
|
|
|
#define __put_user_asm(insn, __gu_ptr, __gu_val, __gu_err) \
|
|
({ \
|
|
__asm__ __volatile__ ( \
|
|
"1:" insn " %1, %2, r0; \
|
|
addk %0, r0, r0; \
|
|
2: \
|
|
.section .fixup,\"ax\"; \
|
|
3: brid 2b; \
|
|
addik %0, r0, %3; \
|
|
.previous; \
|
|
.section __ex_table,\"a\"; \
|
|
.word 1b,3b; \
|
|
.previous;" \
|
|
: "=r"(__gu_err) \
|
|
: "r"(__gu_val), "r"(__gu_ptr), "i"(-EFAULT) \
|
|
); \
|
|
})
|
|
|
|
/*
|
|
* Return: number of not copied bytes, i.e. 0 if OK or non-zero if fail.
|
|
*/
|
|
static inline int clear_user(char *to, int size)
|
|
{
|
|
if (size && access_ok(VERIFY_WRITE, to, size)) {
|
|
__asm__ __volatile__ (" \
|
|
1: \
|
|
sb r0, %2, r0; \
|
|
addik %0, %0, -1; \
|
|
bneid %0, 1b; \
|
|
addik %2, %2, 1; \
|
|
2: \
|
|
.section __ex_table,\"a\"; \
|
|
.word 1b,2b; \
|
|
.section .text;" \
|
|
: "=r"(size) \
|
|
: "0"(size), "r"(to)
|
|
);
|
|
}
|
|
return size;
|
|
}
|
|
|
|
extern unsigned long __copy_tofrom_user(void __user *to,
|
|
const void __user *from, unsigned long size);
|
|
|
|
#define copy_to_user(to, from, n) \
|
|
(access_ok(VERIFY_WRITE, (to), (n)) ? \
|
|
__copy_tofrom_user((void __user *)(to), \
|
|
(__force const void __user *)(from), (n)) \
|
|
: -EFAULT)
|
|
|
|
#define __copy_to_user(to, from, n) copy_to_user((to), (from), (n))
|
|
#define __copy_to_user_inatomic(to, from, n) copy_to_user((to), (from), (n))
|
|
|
|
#define copy_from_user(to, from, n) \
|
|
(access_ok(VERIFY_READ, (from), (n)) ? \
|
|
__copy_tofrom_user((__force void __user *)(to), \
|
|
(void __user *)(from), (n)) \
|
|
: -EFAULT)
|
|
|
|
#define __copy_from_user(to, from, n) copy_from_user((to), (from), (n))
|
|
#define __copy_from_user_inatomic(to, from, n) \
|
|
copy_from_user((to), (from), (n))
|
|
|
|
extern int __strncpy_user(char *to, const char __user *from, int len);
|
|
extern int __strnlen_user(const char __user *sstr, int len);
|
|
|
|
#define strncpy_from_user(to, from, len) \
|
|
(access_ok(VERIFY_READ, from, 1) ? \
|
|
__strncpy_user(to, from, len) : -EFAULT)
|
|
#define strnlen_user(str, len) \
|
|
(access_ok(VERIFY_READ, str, 1) ? __strnlen_user(str, len) : 0)
|
|
|
|
#endif /* CONFIG_MMU */
|
|
|
|
/*
|
|
* The exception table consists of pairs of addresses: the first is the
|
|
* address of an instruction that is allowed to fault, and the second is
|
|
* the address at which the program should continue. No registers are
|
|
* modified, so it is entirely up to the continuation code to figure out
|
|
* what to do.
|
|
*
|
|
* All the routines below use bits of fixup code that are out of line
|
|
* with the main instruction path. This means when everything is well,
|
|
* we don't even have to jump over them. Further, they do not intrude
|
|
* on our cache or tlb entries.
|
|
*/
|
|
struct exception_table_entry {
|
|
unsigned long insn, fixup;
|
|
};
|
|
|
|
#endif /* __ASSEMBLY__ */
|
|
#endif /* __KERNEL__ */
|
|
|
|
#endif /* _ASM_MICROBLAZE_UACCESS_H */
|