diff options
Diffstat (limited to 'arch/x86/mm/fault.c')
| -rw-r--r-- | arch/x86/mm/fault.c | 60 | 
1 files changed, 35 insertions, 25 deletions
| diff --git a/arch/x86/mm/fault.c b/arch/x86/mm/fault.c index b836a7274e12..b0ff378650a9 100644 --- a/arch/x86/mm/fault.c +++ b/arch/x86/mm/fault.c @@ -1,3 +1,4 @@ +// SPDX-License-Identifier: GPL-2.0  /*   *  Copyright (C) 1995  Linus Torvalds   *  Copyright (C) 2001, 2002 Andi Kleen, SuSE Labs. @@ -192,8 +193,7 @@ is_prefetch(struct pt_regs *regs, unsigned long error_code, unsigned long addr)   * 6. T1   : reaches here, sees vma_pkey(vma)=5, when we really   *	     faulted on a pte with its pkey=4.   */ -static void fill_sig_info_pkey(int si_code, siginfo_t *info, -		struct vm_area_struct *vma) +static void fill_sig_info_pkey(int si_code, siginfo_t *info, u32 *pkey)  {  	/* This is effectively an #ifdef */  	if (!boot_cpu_has(X86_FEATURE_OSPKE)) @@ -209,7 +209,7 @@ static void fill_sig_info_pkey(int si_code, siginfo_t *info,  	 * valid VMA, so we should never reach this without a  	 * valid VMA.  	 */ -	if (!vma) { +	if (!pkey) {  		WARN_ONCE(1, "PKU fault with no VMA passed in");  		info->si_pkey = 0;  		return; @@ -219,13 +219,12 @@ static void fill_sig_info_pkey(int si_code, siginfo_t *info,  	 * absolutely guranteed to be 100% accurate because of  	 * the race explained above.  	 */ -	info->si_pkey = vma_pkey(vma); +	info->si_pkey = *pkey;  }  static void  force_sig_info_fault(int si_signo, int si_code, unsigned long address, -		     struct task_struct *tsk, struct vm_area_struct *vma, -		     int fault) +		     struct task_struct *tsk, u32 *pkey, int fault)  {  	unsigned lsb = 0;  	siginfo_t info; @@ -240,7 +239,7 @@ force_sig_info_fault(int si_signo, int si_code, unsigned long address,  		lsb = PAGE_SHIFT;  	info.si_addr_lsb = lsb; -	fill_sig_info_pkey(si_code, &info, vma); +	fill_sig_info_pkey(si_code, &info, pkey);  	force_sig_info(si_signo, &info, tsk);  } @@ -762,8 +761,6 @@ no_context(struct pt_regs *regs, unsigned long error_code,  	struct task_struct *tsk = current;  	unsigned long flags;  	int sig; -	/* No context means no VMA to pass down */ -	struct vm_area_struct *vma = NULL;  	/* Are we prepared to handle this kernel fault? */  	if (fixup_exception(regs, X86_TRAP_PF)) { @@ -788,7 +785,7 @@ no_context(struct pt_regs *regs, unsigned long error_code,  			/* XXX: hwpoison faults will set the wrong code. */  			force_sig_info_fault(signal, si_code, address, -					     tsk, vma, 0); +					     tsk, NULL, 0);  		}  		/* @@ -806,7 +803,6 @@ no_context(struct pt_regs *regs, unsigned long error_code,  	if (is_vmalloc_addr((void *)address) &&  	    (((unsigned long)tsk->stack - 1 - address < PAGE_SIZE) ||  	     address - ((unsigned long)tsk->stack + THREAD_SIZE) < PAGE_SIZE)) { -		register void *__sp asm("rsp");  		unsigned long stack = this_cpu_read(orig_ist.ist[DOUBLEFAULT_STACK]) - sizeof(void *);  		/*  		 * We're likely to be running with very little stack space @@ -821,7 +817,7 @@ no_context(struct pt_regs *regs, unsigned long error_code,  		asm volatile ("movq %[stack], %%rsp\n\t"  			      "call handle_stack_overflow\n\t"  			      "1: jmp 1b" -			      : "+r" (__sp) +			      : ASM_CALL_CONSTRAINT  			      : "D" ("kernel stack overflow (page fault)"),  				"S" (regs), "d" (address),  				[stack] "rm" (stack)); @@ -897,8 +893,7 @@ show_signal_msg(struct pt_regs *regs, unsigned long error_code,  static void  __bad_area_nosemaphore(struct pt_regs *regs, unsigned long error_code, -		       unsigned long address, struct vm_area_struct *vma, -		       int si_code) +		       unsigned long address, u32 *pkey, int si_code)  {  	struct task_struct *tsk = current; @@ -946,7 +941,7 @@ __bad_area_nosemaphore(struct pt_regs *regs, unsigned long error_code,  		tsk->thread.error_code	= error_code;  		tsk->thread.trap_nr	= X86_TRAP_PF; -		force_sig_info_fault(SIGSEGV, si_code, address, tsk, vma, 0); +		force_sig_info_fault(SIGSEGV, si_code, address, tsk, pkey, 0);  		return;  	} @@ -959,9 +954,9 @@ __bad_area_nosemaphore(struct pt_regs *regs, unsigned long error_code,  static noinline void  bad_area_nosemaphore(struct pt_regs *regs, unsigned long error_code, -		     unsigned long address, struct vm_area_struct *vma) +		     unsigned long address, u32 *pkey)  { -	__bad_area_nosemaphore(regs, error_code, address, vma, SEGV_MAPERR); +	__bad_area_nosemaphore(regs, error_code, address, pkey, SEGV_MAPERR);  }  static void @@ -969,6 +964,10 @@ __bad_area(struct pt_regs *regs, unsigned long error_code,  	   unsigned long address,  struct vm_area_struct *vma, int si_code)  {  	struct mm_struct *mm = current->mm; +	u32 pkey; + +	if (vma) +		pkey = vma_pkey(vma);  	/*  	 * Something tried to access memory that isn't in our memory map.. @@ -976,7 +975,8 @@ __bad_area(struct pt_regs *regs, unsigned long error_code,  	 */  	up_read(&mm->mmap_sem); -	__bad_area_nosemaphore(regs, error_code, address, vma, si_code); +	__bad_area_nosemaphore(regs, error_code, address, +			       (vma) ? &pkey : NULL, si_code);  }  static noinline void @@ -1019,7 +1019,7 @@ bad_area_access_error(struct pt_regs *regs, unsigned long error_code,  static void  do_sigbus(struct pt_regs *regs, unsigned long error_code, unsigned long address, -	  struct vm_area_struct *vma, unsigned int fault) +	  u32 *pkey, unsigned int fault)  {  	struct task_struct *tsk = current;  	int code = BUS_ADRERR; @@ -1046,13 +1046,12 @@ do_sigbus(struct pt_regs *regs, unsigned long error_code, unsigned long address,  		code = BUS_MCEERR_AR;  	}  #endif -	force_sig_info_fault(SIGBUS, code, address, tsk, vma, fault); +	force_sig_info_fault(SIGBUS, code, address, tsk, pkey, fault);  }  static noinline void  mm_fault_error(struct pt_regs *regs, unsigned long error_code, -	       unsigned long address, struct vm_area_struct *vma, -	       unsigned int fault) +	       unsigned long address, u32 *pkey, unsigned int fault)  {  	if (fatal_signal_pending(current) && !(error_code & PF_USER)) {  		no_context(regs, error_code, address, 0, 0); @@ -1076,9 +1075,9 @@ mm_fault_error(struct pt_regs *regs, unsigned long error_code,  	} else {  		if (fault & (VM_FAULT_SIGBUS|VM_FAULT_HWPOISON|  			     VM_FAULT_HWPOISON_LARGE)) -			do_sigbus(regs, error_code, address, vma, fault); +			do_sigbus(regs, error_code, address, pkey, fault);  		else if (fault & VM_FAULT_SIGSEGV) -			bad_area_nosemaphore(regs, error_code, address, vma); +			bad_area_nosemaphore(regs, error_code, address, pkey);  		else  			BUG();  	} @@ -1268,6 +1267,7 @@ __do_page_fault(struct pt_regs *regs, unsigned long error_code,  	struct mm_struct *mm;  	int fault, major = 0;  	unsigned int flags = FAULT_FLAG_ALLOW_RETRY | FAULT_FLAG_KILLABLE; +	u32 pkey;  	tsk = current;  	mm = tsk->mm; @@ -1441,7 +1441,17 @@ good_area:  	 * make sure we exit gracefully rather than endlessly redo  	 * the fault.  Since we never set FAULT_FLAG_RETRY_NOWAIT, if  	 * we get VM_FAULT_RETRY back, the mmap_sem has been unlocked. +	 * +	 * Note that handle_userfault() may also release and reacquire mmap_sem +	 * (and not return with VM_FAULT_RETRY), when returning to userland to +	 * repeat the page fault later with a VM_FAULT_NOPAGE retval +	 * (potentially after handling any pending signal during the return to +	 * userland). The return to userland is identified whenever +	 * FAULT_FLAG_USER|FAULT_FLAG_KILLABLE are both set in flags. +	 * Thus we have to be careful about not touching vma after handling the +	 * fault, so we read the pkey beforehand.  	 */ +	pkey = vma_pkey(vma);  	fault = handle_mm_fault(vma, address, flags);  	major |= fault & VM_FAULT_MAJOR; @@ -1470,7 +1480,7 @@ good_area:  	up_read(&mm->mmap_sem);  	if (unlikely(fault & VM_FAULT_ERROR)) { -		mm_fault_error(regs, error_code, address, vma, fault); +		mm_fault_error(regs, error_code, address, &pkey, fault);  		return;  	} |