Searched refs:task_size (Results 1 – 16 of 16) sorted by relevance
42 static unsigned long stack_maxrandom_size(unsigned long task_size) in stack_maxrandom_size() argument46 max = (-1UL) & __STACK_RND_MASK(task_size == task_size_32bit()); in stack_maxrandom_size()83 static unsigned long mmap_base(unsigned long rnd, unsigned long task_size, in mmap_base() argument87 unsigned long pad = stack_maxrandom_size(task_size) + stack_guard_gap; in mmap_base()99 gap_max = (task_size / 6) * 5; in mmap_base()106 return PAGE_ALIGN(task_size - gap - rnd); in mmap_base()110 unsigned long task_size) in mmap_legacy_base() argument112 return __TASK_UNMAPPED_BASE(task_size) + rnd; in mmap_legacy_base()120 unsigned long random_factor, unsigned long task_size, in arch_pick_mmap_base() argument123 *legacy_base = mmap_legacy_base(random_factor, task_size); in arch_pick_mmap_base()[all …]
94 unsigned long task_size = TASK_SIZE; in arch_get_unmapped_area() local109 task_size = STACK_TOP32; in arch_get_unmapped_area()110 if (unlikely(len > task_size || len >= VA_EXCLUDE_START)) in arch_get_unmapped_area()124 if (task_size - len >= addr && in arch_get_unmapped_area()132 info.high_limit = min(task_size, VA_EXCLUDE_START); in arch_get_unmapped_area()137 if ((addr & ~PAGE_MASK) && task_size > VA_EXCLUDE_END) { in arch_get_unmapped_area()140 info.high_limit = task_size; in arch_get_unmapped_area()154 unsigned long task_size = STACK_TOP32; in arch_get_unmapped_area_topdown() local172 if (unlikely(len > task_size)) in arch_get_unmapped_area_topdown()187 if (task_size - len >= addr && in arch_get_unmapped_area_topdown()[all …]
151 int task_size = sizeof(struct task_struct); in fpu__init_task_struct_size() local157 task_size -= sizeof(current->thread.fpu.__fpstate.regs); in fpu__init_task_struct_size()163 task_size += fpu_kernel_cfg.default_size; in fpu__init_task_struct_size()176 arch_task_struct_size = task_size; in fpu__init_task_struct_size()
19 unsigned long task_size = STACK_TOP; in prepare_hugepage_range() local26 if (len > task_size) in prepare_hugepage_range()28 if (task_size - len < addr) in prepare_hugepage_range()
18 unsigned long task_size = STACK_TOP; in prepare_hugepage_range() local25 if (len > task_size) in prepare_hugepage_range()27 if (task_size - len < addr) in prepare_hugepage_range()
16 extern unsigned long task_size;27 um_vdso_addr = task_size - PAGE_SIZE; in init_vdso()
33 unsigned long task_size = TASK_SIZE; in hugetlb_get_unmapped_area_bottomup() local37 task_size = STACK_TOP32; in hugetlb_get_unmapped_area_bottomup()42 info.high_limit = min(task_size, VA_EXCLUDE_START); in hugetlb_get_unmapped_area_bottomup()47 if ((addr & ~PAGE_MASK) && task_size > VA_EXCLUDE_END) { in hugetlb_get_unmapped_area_bottomup()50 info.high_limit = task_size; in hugetlb_get_unmapped_area_bottomup()103 unsigned long task_size = TASK_SIZE; in hugetlb_get_unmapped_area() local106 task_size = STACK_TOP32; in hugetlb_get_unmapped_area()110 if (len > task_size) in hugetlb_get_unmapped_area()122 if (task_size - len >= addr && in hugetlb_get_unmapped_area()
61 extern unsigned long task_size;63 #define TASK_SIZE (task_size)
27 #define TASK_SIZE_OF(tsk) ((tsk)->thread.task_size)107 unsigned long task_size; member144 .task_size = DEFAULT_TASK_SIZE, \
239 current->thread.task_size = DEFAULT_TASK_SIZE; \249 current->thread.task_size = DEFAULT_TASK_SIZE32; \
6 obj-y = registers.o task_size.o mcontext.o
200 addr >= mm->task_size || len >= mm->task_size || in SYSCALL_DEFINE3()201 addr + len > mm->task_size) in SYSCALL_DEFINE3()
467 BUG_ON(mm->task_size == 0); in slice_get_unmapped_area()
272 unsigned long task_size; variable273 EXPORT_SYMBOL(task_size);342 task_size = host_task_size & PGDIR_MASK; in linux_main()
836 int task_size = sizeof(struct task_struct); in setup_task_size() local839 task_size -= sizeof(__vector128) * __NUM_VXRS; in setup_task_size()840 task_size += sizeof(freg_t) * __NUM_FPRS; in setup_task_size()842 arch_task_struct_size = task_size; in setup_task_size()
666 #define __TASK_UNMAPPED_BASE(task_size) (PAGE_ALIGN(task_size / 3)) argument