/kernel/bpf/ |
D | cgroup_iter.c | 54 int order; member 77 if (p->order == BPF_CGROUP_ITER_DESCENDANTS_PRE) in cgroup_iter_seq_start() 79 else if (p->order == BPF_CGROUP_ITER_DESCENDANTS_POST) in cgroup_iter_seq_start() 110 if (p->order == BPF_CGROUP_ITER_DESCENDANTS_PRE) in cgroup_iter_seq_next() 112 else if (p->order == BPF_CGROUP_ITER_DESCENDANTS_POST) in cgroup_iter_seq_next() 114 else if (p->order == BPF_CGROUP_ITER_ANCESTORS_UP) in cgroup_iter_seq_next() 176 p->order = aux->cgroup.order; in BTF_ID_LIST_SINGLE() 200 int order = linfo->cgroup.order; in bpf_iter_attach_cgroup() local 203 if (order != BPF_CGROUP_ITER_DESCENDANTS_PRE && in bpf_iter_attach_cgroup() 204 order != BPF_CGROUP_ITER_DESCENDANTS_POST && in bpf_iter_attach_cgroup() [all …]
|
/kernel/locking/ |
D | test-ww_mutex.c | 391 int *order; in get_random_order() local 394 order = kmalloc_array(count, sizeof(*order), GFP_KERNEL); in get_random_order() 395 if (!order) in get_random_order() 396 return order; in get_random_order() 399 order[n] = n; in get_random_order() 404 tmp = order[n]; in get_random_order() 405 order[n] = order[r]; in get_random_order() 406 order[r] = tmp; in get_random_order() 410 return order; in get_random_order() 424 int *order; in stress_inorder_work() local [all …]
|
/kernel/dma/ |
D | pool.c | 82 unsigned int order; in atomic_pool_expand() local 88 order = min(get_order(pool_size), MAX_ORDER-1); in atomic_pool_expand() 91 pool_size = 1 << (PAGE_SHIFT + order); in atomic_pool_expand() 93 page = dma_alloc_from_contiguous(NULL, 1 << order, in atomic_pool_expand() 94 order, false); in atomic_pool_expand() 96 page = alloc_pages(gfp, order); in atomic_pool_expand() 97 } while (!page && order-- > 0); in atomic_pool_expand() 117 1 << order); in atomic_pool_expand() 130 1 << order); in atomic_pool_expand() 140 __free_pages(page, order); in atomic_pool_expand()
|
D | coherent.c | 145 int order = get_order(size); in __dma_alloc_from_coherent() local 155 pageno = bitmap_find_free_region(mem->bitmap, mem->size, order); in __dma_alloc_from_coherent() 200 int order, void *vaddr) in __dma_release_from_coherent() argument 208 bitmap_release_region(mem->bitmap, page, order); in __dma_release_from_coherent() 227 int dma_release_from_dev_coherent(struct device *dev, int order, void *vaddr) in dma_release_from_dev_coherent() argument 231 return __dma_release_from_coherent(mem, order, vaddr); in dma_release_from_dev_coherent() 292 int dma_release_from_global_coherent(int order, void *vaddr) in dma_release_from_global_coherent() argument 297 return __dma_release_from_coherent(dma_coherent_default_memory, order, in dma_release_from_global_coherent()
|
D | swiotlb.c | 434 unsigned int order, area_order; in swiotlb_init_late() local 445 order = get_order(nslabs << IO_TLB_SHIFT); in swiotlb_init_late() 446 nslabs = SLABS_PER_PAGE << order; in swiotlb_init_late() 448 while ((SLABS_PER_PAGE << order) > IO_TLB_MIN_SLABS) { in swiotlb_init_late() 450 order); in swiotlb_init_late() 453 order--; in swiotlb_init_late() 454 nslabs = SLABS_PER_PAGE << order; in swiotlb_init_late() 464 free_pages((unsigned long)vstart, order); in swiotlb_init_late() 475 (PAGE_SIZE << order) >> 20); in swiotlb_init_late() 501 free_pages((unsigned long)vstart, order); in swiotlb_init_late()
|
D | Kconfig | 187 int "Maximum PAGE_SIZE order of alignment for contiguous buffers" 192 PAGE_SIZE order which is greater than or equal to the requested buffer 195 specify the maximum PAGE_SIZE order for contiguous buffers. Larger 196 buffers will be aligned only to this specified order. The order is 199 For example, if your system defaults to 4KiB pages, the order value
|
/kernel/ |
D | kexec_core.c | 300 static struct page *kimage_alloc_pages(gfp_t gfp_mask, unsigned int order) in kimage_alloc_pages() argument 306 pages = alloc_pages(gfp_mask & ~__GFP_ZERO, order); in kimage_alloc_pages() 311 set_page_private(pages, order); in kimage_alloc_pages() 312 count = 1 << order; in kimage_alloc_pages() 329 unsigned int order, count, i; in kimage_free_pages() local 331 order = page_private(page); in kimage_free_pages() 332 count = 1 << order; in kimage_free_pages() 338 __free_pages(page, order); in kimage_free_pages() 352 unsigned int order) in kimage_alloc_normal_control_pages() argument 371 count = 1 << order; in kimage_alloc_normal_control_pages() [all …]
|
/kernel/events/ |
D | ring_buffer.c | 608 static struct page *rb_alloc_aux_page(int node, int order) in rb_alloc_aux_page() argument 612 if (order > MAX_ORDER) in rb_alloc_aux_page() 613 order = MAX_ORDER; in rb_alloc_aux_page() 616 page = alloc_pages_node(node, PERF_AUX_GFP, order); in rb_alloc_aux_page() 617 } while (!page && order--); in rb_alloc_aux_page() 619 if (page && order) { in rb_alloc_aux_page() 626 split_page(page, order); in rb_alloc_aux_page() 628 set_page_private(page, order); in rb_alloc_aux_page() 716 int last, order; in rb_alloc_aux() local 718 order = min(max_order, ilog2(nr_pages - rb->aux_nr_pages)); in rb_alloc_aux() [all …]
|
/kernel/trace/ |
D | ftrace.c | 1088 int order; member 3182 int order; in ftrace_allocate_records() local 3191 order = fls(pages) - 1; in ftrace_allocate_records() 3194 pg->records = (void *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, order); in ftrace_allocate_records() 3198 if (!order) in ftrace_allocate_records() 3200 order--; in ftrace_allocate_records() 3204 ftrace_number_of_pages += 1 << order; in ftrace_allocate_records() 3207 cnt = (PAGE_SIZE << order) / ENTRY_SIZE; in ftrace_allocate_records() 3208 pg->order = order; in ftrace_allocate_records() 3222 free_pages((unsigned long)pg->records, pg->order); in ftrace_free_pages() [all …]
|
D | trace.c | 2270 int order = get_order(sizeof(*s) + s->cmdline_num * TASK_COMM_LEN); in free_saved_cmdlines_buffer() local 2274 free_pages((unsigned long)s, order); in free_saved_cmdlines_buffer() 2282 int order; in allocate_cmdlines_buffer() local 2286 order = get_order(orig_size); in allocate_cmdlines_buffer() 2287 size = 1 << (order + PAGE_SHIFT); in allocate_cmdlines_buffer() 2288 page = alloc_pages(GFP_KERNEL, order); in allocate_cmdlines_buffer()
|
D | Kconfig | 755 separated out as a stand-alone facility in order to allow it
|
/kernel/time/ |
D | Kconfig | 165 other dependencies to provide in order to make the full
|
/kernel/power/ |
D | Kconfig | 288 In order to use APM, you will need supporting software. For location
|
/kernel/rcu/ |
D | Kconfig | 180 value to the maximum value possible in order to reduce the
|