Home
last modified time | relevance | path

Searched refs:alloc_flags (Results 1 – 16 of 16) sorted by relevance

/kernel/linux/linux-5.10/lib/
Dstackdepot.c105 u32 hash, void **prealloc, gfp_t alloc_flags) in depot_alloc_stack() argument
237 gfp_t alloc_flags) in stack_depot_save() argument
276 alloc_flags &= ~GFP_ZONEMASK; in stack_depot_save()
277 alloc_flags &= (GFP_ATOMIC | GFP_KERNEL); in stack_depot_save()
278 alloc_flags |= __GFP_NOWARN; in stack_depot_save()
279 page = alloc_pages(alloc_flags, STACK_ALLOC_ORDER); in stack_depot_save()
290 hash, &prealloc, alloc_flags); in stack_depot_save()
/kernel/linux/linux-5.10/mm/
Dpage_alloc.c2290 unsigned int alloc_flags) in prep_new_page() argument
2306 if (alloc_flags & ALLOC_NO_WATERMARKS) in prep_new_page()
2526 unsigned int alloc_flags, int start_type, bool whole_block) in steal_suitable_fallback() argument
2552 if (boost_watermark(zone) && (alloc_flags & ALLOC_KSWAPD)) in steal_suitable_fallback()
2768 unsigned int alloc_flags) in __rmqueue_fallback() argument
2782 if (alloc_flags & ALLOC_NOFRAGMENT) in __rmqueue_fallback()
2834 steal_suitable_fallback(zone, page, alloc_flags, start_migratetype, in __rmqueue_fallback()
2846 int migratetype, unsigned int alloc_flags) in __rmqueue_with_cma_reuse() argument
2854 alloc_flags &= ~ALLOC_CMA; in __rmqueue_with_cma_reuse()
2859 __rmqueue_fallback(zone, order, migratetype, alloc_flags)) in __rmqueue_with_cma_reuse()
[all …]
Dcompaction.c2079 unsigned int alloc_flags, in __compaction_suitable() argument
2088 watermark = wmark_pages(zone, alloc_flags & ALLOC_WMARK_MASK); in __compaction_suitable()
2094 alloc_flags)) in __compaction_suitable()
2122 unsigned int alloc_flags, in compaction_suitable() argument
2128 ret = __compaction_suitable(zone, order, alloc_flags, highest_zoneidx, in compaction_suitable()
2160 int alloc_flags) in compaction_zonelist_suitable() argument
2182 compact_result = __compaction_suitable(zone, order, alloc_flags, in compaction_zonelist_suitable()
2213 ret = compaction_suitable(cc->zone, cc->order, cc->alloc_flags, in compact_zone()
2405 unsigned int alloc_flags, int highest_zoneidx, in compact_zone_order() argument
2416 .alloc_flags = alloc_flags, in compact_zone_order()
[all …]
Dinternal.h366 const unsigned int alloc_flags; /* alloc flags of a direct compactor */ member
/kernel/linux/linux-5.10/include/linux/
Dcompaction.h94 unsigned int order, unsigned int alloc_flags,
99 unsigned int alloc_flags, int highest_zoneidx);
180 int alloc_flags);
192 int alloc_flags, int highest_zoneidx) in compaction_suitable() argument
Dmmzone.h881 int highest_zoneidx, unsigned int alloc_flags,
885 unsigned int alloc_flags);
/kernel/linux/linux-5.10/drivers/md/
Ddm-zoned-reclaim.c286 int alloc_flags = DMZ_ALLOC_SEQ; in dmz_reclaim_rnd_data() local
292 alloc_flags | DMZ_ALLOC_RECLAIM); in dmz_reclaim_rnd_data()
293 if (!szone && alloc_flags == DMZ_ALLOC_SEQ && dmz_nr_cache_zones(zmd)) { in dmz_reclaim_rnd_data()
294 alloc_flags = DMZ_ALLOC_RND; in dmz_reclaim_rnd_data()
Ddm-zoned-metadata.c2070 int alloc_flags = zmd->nr_cache ? DMZ_ALLOC_CACHE : DMZ_ALLOC_RND; in dmz_get_chunk_mapping() local
2085 dzone = dmz_alloc_zone(zmd, 0, alloc_flags); in dmz_get_chunk_mapping()
2182 int alloc_flags = zmd->nr_cache ? DMZ_ALLOC_CACHE : DMZ_ALLOC_RND; in dmz_get_chunk_buffer() local
2191 bzone = dmz_alloc_zone(zmd, 0, alloc_flags); in dmz_get_chunk_buffer()
/kernel/linux/linux-5.10/drivers/gpu/drm/amd/amdgpu/
Damdgpu_amdkfd_gpuvm.c393 bool coherent = mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_COHERENT; in get_pte_flags()
397 if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_WRITABLE) in get_pte_flags()
399 if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_EXECUTABLE) in get_pte_flags()
404 if (mem->alloc_flags & KFD_IOC_ALLOC_MEM_FLAGS_VRAM) { in get_pte_flags()
1149 u64 alloc_flags; in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu() local
1157 alloc_flags = AMDGPU_GEM_CREATE_VRAM_WIPE_ON_RELEASE; in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu()
1158 alloc_flags |= (flags & KFD_IOC_ALLOC_MEM_FLAGS_PUBLIC) ? in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu()
1163 alloc_flags = 0; in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu()
1167 alloc_flags = 0; in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu()
1176 alloc_flags = 0; in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu()
[all …]
Damdgpu_amdkfd.h61 uint32_t alloc_flags; member
/kernel/linux/linux-5.10/fs/btrfs/
Dblock-group.c2235 u64 alloc_flags; in btrfs_inc_block_group_ro() local
2266 alloc_flags = btrfs_get_alloc_profile(fs_info, cache->flags); in btrfs_inc_block_group_ro()
2267 if (alloc_flags != cache->flags) { in btrfs_inc_block_group_ro()
2268 ret = btrfs_chunk_alloc(trans, alloc_flags, in btrfs_inc_block_group_ro()
2286 alloc_flags = btrfs_get_alloc_profile(fs_info, cache->space_info->flags); in btrfs_inc_block_group_ro()
2287 ret = btrfs_chunk_alloc(trans, alloc_flags, CHUNK_ALLOC_FORCE); in btrfs_inc_block_group_ro()
2295 alloc_flags = btrfs_get_alloc_profile(fs_info, cache->flags); in btrfs_inc_block_group_ro()
2297 check_system_chunk(trans, alloc_flags); in btrfs_inc_block_group_ro()
3075 u64 alloc_flags = btrfs_get_alloc_profile(trans->fs_info, type); in btrfs_force_chunk_alloc() local
3077 return btrfs_chunk_alloc(trans, alloc_flags, CHUNK_ALLOC_FORCE); in btrfs_force_chunk_alloc()
/kernel/linux/linux-5.10/drivers/base/regmap/
Dinternal.h60 gfp_t alloc_flags; member
Dregmap.c785 map->alloc_flags = GFP_ATOMIC; in __regmap_init()
787 map->alloc_flags = GFP_KERNEL; in __regmap_init()
2254 wval = kmemdup(val, val_count * val_bytes, map->alloc_flags); in regmap_bulk_write()
/kernel/linux/linux-5.10/drivers/iommu/
Ddma-iommu.c546 gfp_t alloc_flags = gfp; in __iommu_dma_alloc_pages() local
550 alloc_flags |= __GFP_NORETRY; in __iommu_dma_alloc_pages()
551 page = alloc_pages_node(nid, alloc_flags, order); in __iommu_dma_alloc_pages()
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/
Dcmd.c77 gfp_t alloc_flags = cbk ? GFP_ATOMIC : GFP_KERNEL; in cmd_alloc_ent() local
80 ent = kzalloc(sizeof(*ent), alloc_flags); in cmd_alloc_ent()
/kernel/linux/linux-5.10/arch/s390/kvm/
Dkvm-s390.c2643 gfp_t alloc_flags = GFP_KERNEL; in kvm_arch_init_vm() local
2666 alloc_flags |= GFP_DMA; in kvm_arch_init_vm()
2669 kvm->arch.sca = (struct bsca_block *) get_zeroed_page(alloc_flags); in kvm_arch_init_vm()