| /* | 
 |  * Flexible mmap layout support | 
 |  * | 
 |  * Based on code by Ingo Molnar and Andi Kleen, copyrighted | 
 |  * as follows: | 
 |  * | 
 |  * Copyright 2003-2009 Red Hat Inc. | 
 |  * All Rights Reserved. | 
 |  * Copyright 2005 Andi Kleen, SUSE Labs. | 
 |  * Copyright 2007 Jiri Kosina, SUSE Labs. | 
 |  * | 
 |  * This program is free software; you can redistribute it and/or modify | 
 |  * it under the terms of the GNU General Public License as published by | 
 |  * the Free Software Foundation; either version 2 of the License, or | 
 |  * (at your option) any later version. | 
 |  * | 
 |  * This program is distributed in the hope that it will be useful, | 
 |  * but WITHOUT ANY WARRANTY; without even the implied warranty of | 
 |  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the | 
 |  * GNU General Public License for more details. | 
 |  * | 
 |  * You should have received a copy of the GNU General Public License | 
 |  * along with this program; if not, write to the Free Software | 
 |  * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307  USA | 
 |  */ | 
 |  | 
 | #include <linux/personality.h> | 
 | #include <linux/mm.h> | 
 | #include <linux/random.h> | 
 | #include <linux/limits.h> | 
 | #include <linux/sched/signal.h> | 
 | #include <linux/sched/mm.h> | 
 | #include <linux/compat.h> | 
 | #include <asm/elf.h> | 
 |  | 
 | #include "physaddr.h" | 
 |  | 
 | struct va_alignment __read_mostly va_align = { | 
 | 	.flags = -1, | 
 | }; | 
 |  | 
 | unsigned long task_size_32bit(void) | 
 | { | 
 | 	return IA32_PAGE_OFFSET; | 
 | } | 
 |  | 
 | unsigned long task_size_64bit(int full_addr_space) | 
 | { | 
 | 	return full_addr_space ? TASK_SIZE_MAX : DEFAULT_MAP_WINDOW; | 
 | } | 
 |  | 
 | static unsigned long stack_maxrandom_size(unsigned long task_size) | 
 | { | 
 | 	unsigned long max = 0; | 
 | 	if (current->flags & PF_RANDOMIZE) { | 
 | 		max = (-1UL) & __STACK_RND_MASK(task_size == task_size_32bit()); | 
 | 		max <<= PAGE_SHIFT; | 
 | 	} | 
 |  | 
 | 	return max; | 
 | } | 
 |  | 
 | #ifdef CONFIG_COMPAT | 
 | # define mmap32_rnd_bits  mmap_rnd_compat_bits | 
 | # define mmap64_rnd_bits  mmap_rnd_bits | 
 | #else | 
 | # define mmap32_rnd_bits  mmap_rnd_bits | 
 | # define mmap64_rnd_bits  mmap_rnd_bits | 
 | #endif | 
 |  | 
 | #define SIZE_128M    (128 * 1024 * 1024UL) | 
 |  | 
 | static int mmap_is_legacy(void) | 
 | { | 
 | 	if (current->personality & ADDR_COMPAT_LAYOUT) | 
 | 		return 1; | 
 |  | 
 | 	return sysctl_legacy_va_layout; | 
 | } | 
 |  | 
 | static unsigned long arch_rnd(unsigned int rndbits) | 
 | { | 
 | 	if (!(current->flags & PF_RANDOMIZE)) | 
 | 		return 0; | 
 | 	return (get_random_long() & ((1UL << rndbits) - 1)) << PAGE_SHIFT; | 
 | } | 
 |  | 
 | unsigned long arch_mmap_rnd(void) | 
 | { | 
 | 	return arch_rnd(mmap_is_ia32() ? mmap32_rnd_bits : mmap64_rnd_bits); | 
 | } | 
 |  | 
 | static unsigned long mmap_base(unsigned long rnd, unsigned long task_size, | 
 | 			       struct rlimit *rlim_stack) | 
 | { | 
 | 	unsigned long gap = rlim_stack->rlim_cur; | 
 | 	unsigned long pad = stack_maxrandom_size(task_size) + stack_guard_gap; | 
 | 	unsigned long gap_min, gap_max; | 
 |  | 
 | 	/* Values close to RLIM_INFINITY can overflow. */ | 
 | 	if (gap + pad > gap) | 
 | 		gap += pad; | 
 |  | 
 | 	/* | 
 | 	 * Top of mmap area (just below the process stack). | 
 | 	 * Leave an at least ~128 MB hole with possible stack randomization. | 
 | 	 */ | 
 | 	gap_min = SIZE_128M; | 
 | 	gap_max = (task_size / 6) * 5; | 
 |  | 
 | 	if (gap < gap_min) | 
 | 		gap = gap_min; | 
 | 	else if (gap > gap_max) | 
 | 		gap = gap_max; | 
 |  | 
 | 	return PAGE_ALIGN(task_size - gap - rnd); | 
 | } | 
 |  | 
 | static unsigned long mmap_legacy_base(unsigned long rnd, | 
 | 				      unsigned long task_size) | 
 | { | 
 | 	return __TASK_UNMAPPED_BASE(task_size) + rnd; | 
 | } | 
 |  | 
 | /* | 
 |  * This function, called very early during the creation of a new | 
 |  * process VM image, sets up which VM layout function to use: | 
 |  */ | 
 | static void arch_pick_mmap_base(unsigned long *base, unsigned long *legacy_base, | 
 | 		unsigned long random_factor, unsigned long task_size, | 
 | 		struct rlimit *rlim_stack) | 
 | { | 
 | 	*legacy_base = mmap_legacy_base(random_factor, task_size); | 
 | 	if (mmap_is_legacy()) | 
 | 		*base = *legacy_base; | 
 | 	else | 
 | 		*base = mmap_base(random_factor, task_size, rlim_stack); | 
 | } | 
 |  | 
 | void arch_pick_mmap_layout(struct mm_struct *mm, struct rlimit *rlim_stack) | 
 | { | 
 | 	if (mmap_is_legacy()) | 
 | 		mm->get_unmapped_area = arch_get_unmapped_area; | 
 | 	else | 
 | 		mm->get_unmapped_area = arch_get_unmapped_area_topdown; | 
 |  | 
 | 	arch_pick_mmap_base(&mm->mmap_base, &mm->mmap_legacy_base, | 
 | 			arch_rnd(mmap64_rnd_bits), task_size_64bit(0), | 
 | 			rlim_stack); | 
 |  | 
 | #ifdef CONFIG_HAVE_ARCH_COMPAT_MMAP_BASES | 
 | 	/* | 
 | 	 * The mmap syscall mapping base decision depends solely on the | 
 | 	 * syscall type (64-bit or compat). This applies for 64bit | 
 | 	 * applications and 32bit applications. The 64bit syscall uses | 
 | 	 * mmap_base, the compat syscall uses mmap_compat_base. | 
 | 	 */ | 
 | 	arch_pick_mmap_base(&mm->mmap_compat_base, &mm->mmap_compat_legacy_base, | 
 | 			arch_rnd(mmap32_rnd_bits), task_size_32bit(), | 
 | 			rlim_stack); | 
 | #endif | 
 | } | 
 |  | 
 | unsigned long get_mmap_base(int is_legacy) | 
 | { | 
 | 	struct mm_struct *mm = current->mm; | 
 |  | 
 | #ifdef CONFIG_HAVE_ARCH_COMPAT_MMAP_BASES | 
 | 	if (in_compat_syscall()) { | 
 | 		return is_legacy ? mm->mmap_compat_legacy_base | 
 | 				 : mm->mmap_compat_base; | 
 | 	} | 
 | #endif | 
 | 	return is_legacy ? mm->mmap_legacy_base : mm->mmap_base; | 
 | } | 
 |  | 
 | const char *arch_vma_name(struct vm_area_struct *vma) | 
 | { | 
 | 	if (vma->vm_flags & VM_MPX) | 
 | 		return "[mpx]"; | 
 | 	return NULL; | 
 | } | 
 |  | 
 | /** | 
 |  * mmap_address_hint_valid - Validate the address hint of mmap | 
 |  * @addr:	Address hint | 
 |  * @len:	Mapping length | 
 |  * | 
 |  * Check whether @addr and @addr + @len result in a valid mapping. | 
 |  * | 
 |  * On 32bit this only checks whether @addr + @len is <= TASK_SIZE. | 
 |  * | 
 |  * On 64bit with 5-level page tables another sanity check is required | 
 |  * because mappings requested by mmap(@addr, 0) which cross the 47-bit | 
 |  * virtual address boundary can cause the following theoretical issue: | 
 |  * | 
 |  *  An application calls mmap(addr, 0), i.e. without MAP_FIXED, where @addr | 
 |  *  is below the border of the 47-bit address space and @addr + @len is | 
 |  *  above the border. | 
 |  * | 
 |  *  With 4-level paging this request succeeds, but the resulting mapping | 
 |  *  address will always be within the 47-bit virtual address space, because | 
 |  *  the hint address does not result in a valid mapping and is | 
 |  *  ignored. Hence applications which are not prepared to handle virtual | 
 |  *  addresses above 47-bit work correctly. | 
 |  * | 
 |  *  With 5-level paging this request would be granted and result in a | 
 |  *  mapping which crosses the border of the 47-bit virtual address | 
 |  *  space. If the application cannot handle addresses above 47-bit this | 
 |  *  will lead to misbehaviour and hard to diagnose failures. | 
 |  * | 
 |  * Therefore ignore address hints which would result in a mapping crossing | 
 |  * the 47-bit virtual address boundary. | 
 |  * | 
 |  * Note, that in the same scenario with MAP_FIXED the behaviour is | 
 |  * different. The request with @addr < 47-bit and @addr + @len > 47-bit | 
 |  * fails on a 4-level paging machine but succeeds on a 5-level paging | 
 |  * machine. It is reasonable to expect that an application does not rely on | 
 |  * the failure of such a fixed mapping request, so the restriction is not | 
 |  * applied. | 
 |  */ | 
 | bool mmap_address_hint_valid(unsigned long addr, unsigned long len) | 
 | { | 
 | 	if (TASK_SIZE - len < addr) | 
 | 		return false; | 
 |  | 
 | 	return (addr > DEFAULT_MAP_WINDOW) == (addr + len > DEFAULT_MAP_WINDOW); | 
 | } | 
 |  | 
 | /* Can we access it for direct reading/writing? Must be RAM: */ | 
 | int valid_phys_addr_range(phys_addr_t addr, size_t count) | 
 | { | 
 | 	return addr + count - 1 <= __pa(high_memory - 1); | 
 | } | 
 |  | 
 | /* Can we access it through mmap? Must be a valid physical address: */ | 
 | int valid_mmap_phys_addr_range(unsigned long pfn, size_t count) | 
 | { | 
 | 	phys_addr_t addr = (phys_addr_t)pfn << PAGE_SHIFT; | 
 |  | 
 | 	return phys_addr_valid(addr + count - 1); | 
 | } | 
 |  | 
 | /* | 
 |  * Only allow root to set high MMIO mappings to PROT_NONE. | 
 |  * This prevents an unpriv. user to set them to PROT_NONE and invert | 
 |  * them, then pointing to valid memory for L1TF speculation. | 
 |  * | 
 |  * Note: for locked down kernels may want to disable the root override. | 
 |  */ | 
 | bool pfn_modify_allowed(unsigned long pfn, pgprot_t prot) | 
 | { | 
 | 	if (!boot_cpu_has_bug(X86_BUG_L1TF)) | 
 | 		return true; | 
 | 	if (!__pte_needs_invert(pgprot_val(prot))) | 
 | 		return true; | 
 | 	/* If it's real memory always allow */ | 
 | 	if (pfn_valid(pfn)) | 
 | 		return true; | 
 | 	if (pfn >= l1tf_pfn_limit() && !capable(CAP_SYS_ADMIN)) | 
 | 		return false; | 
 | 	return true; | 
 | } |