|
@@ -96,7 +96,7 @@ static bool write_ok_or_segv(unsigned long ptr, size_t size)
|
|
|
{
|
|
|
/*
|
|
|
* XXX: if access_ok, get_user, and put_user handled
|
|
|
- * sig_on_uaccess_error, this could go away.
|
|
|
+ * sig_on_uaccess_err, this could go away.
|
|
|
*/
|
|
|
|
|
|
if (!access_ok(VERIFY_WRITE, (void __user *)ptr, size)) {
|
|
@@ -125,7 +125,7 @@ bool emulate_vsyscall(struct pt_regs *regs, unsigned long address)
|
|
|
struct task_struct *tsk;
|
|
|
unsigned long caller;
|
|
|
int vsyscall_nr, syscall_nr, tmp;
|
|
|
- int prev_sig_on_uaccess_error;
|
|
|
+ int prev_sig_on_uaccess_err;
|
|
|
long ret;
|
|
|
|
|
|
/*
|
|
@@ -221,8 +221,8 @@ bool emulate_vsyscall(struct pt_regs *regs, unsigned long address)
|
|
|
* With a real vsyscall, page faults cause SIGSEGV. We want to
|
|
|
* preserve that behavior to make writing exploits harder.
|
|
|
*/
|
|
|
- prev_sig_on_uaccess_error = current->thread.sig_on_uaccess_error;
|
|
|
- current->thread.sig_on_uaccess_error = 1;
|
|
|
+ prev_sig_on_uaccess_err = current->thread.sig_on_uaccess_err;
|
|
|
+ current->thread.sig_on_uaccess_err = 1;
|
|
|
|
|
|
ret = -EFAULT;
|
|
|
switch (vsyscall_nr) {
|
|
@@ -243,7 +243,7 @@ bool emulate_vsyscall(struct pt_regs *regs, unsigned long address)
|
|
|
break;
|
|
|
}
|
|
|
|
|
|
- current->thread.sig_on_uaccess_error = prev_sig_on_uaccess_error;
|
|
|
+ current->thread.sig_on_uaccess_err = prev_sig_on_uaccess_err;
|
|
|
|
|
|
check_fault:
|
|
|
if (ret == -EFAULT) {
|