Home
last modified time | relevance | path

Searched refs:range_end (Results 1 – 25 of 63) sorted by relevance

123

/kernel/linux/linux-5.10/arch/arm/mm/
Dcache-feroceon-l2.c144 unsigned long range_end; in calc_range_end() local
152 range_end = end; in calc_range_end()
159 if (range_end > start + MAX_RANGE_SIZE) in calc_range_end()
160 range_end = start + MAX_RANGE_SIZE; in calc_range_end()
165 if (range_end > (start | (PAGE_SIZE - 1)) + 1) in calc_range_end()
166 range_end = (start | (PAGE_SIZE - 1)) + 1; in calc_range_end()
168 return range_end; in calc_range_end()
193 unsigned long range_end = calc_range_end(start, end); in feroceon_l2_inv_range() local
194 l2_inv_pa_range(start, range_end - CACHE_LINE_SIZE); in feroceon_l2_inv_range()
195 start = range_end; in feroceon_l2_inv_range()
[all …]
Dpageattr.c29 unsigned long range_start, unsigned long range_end) in in_range() argument
31 return start >= range_start && start < range_end && in in_range()
32 size <= range_end - start; in in_range()
Dinit.c363 phys_addr_t range_start, range_end; in free_highpages() local
368 &range_start, &range_end, NULL) { in free_highpages()
370 unsigned long end = PFN_DOWN(range_end); in free_highpages()
/kernel/linux/linux-5.10/net/bridge/
Dbr_vlan_options.c33 const struct net_bridge_vlan *range_end) in __vlan_tun_can_enter_range() argument
35 return (!v_curr->tinfo.tunnel_dst && !range_end->tinfo.tunnel_dst) || in __vlan_tun_can_enter_range()
36 vlan_tunid_inrange(v_curr, range_end); in __vlan_tun_can_enter_range()
41 const struct net_bridge_vlan *range_end) in br_vlan_opts_eq_range() argument
43 return v_curr->state == range_end->state && in br_vlan_opts_eq_range()
44 __vlan_tun_can_enter_range(v_curr, range_end); in br_vlan_opts_eq_range()
190 struct net_bridge_vlan *range_end, in br_vlan_process_options() argument
208 if (!range_end || !br_vlan_should_use(range_end)) { in br_vlan_process_options()
214 for (vid = range_start->vid; vid <= range_end->vid; vid++) { in br_vlan_process_options()
Dbr_vlan.c1767 const struct net_bridge_vlan *range_end) in br_vlan_can_enter_range() argument
1769 return v_curr->vid - range_end->vid == 1 && in br_vlan_can_enter_range()
1770 range_end->flags == v_curr->flags && in br_vlan_can_enter_range()
1771 br_vlan_opts_eq_range(v_curr, range_end); in br_vlan_can_enter_range()
1779 struct net_bridge_vlan *v, *range_start = NULL, *range_end = NULL; in br_vlan_dump_dev() local
1829 range_end = v; in br_vlan_dump_dev()
1834 !br_vlan_can_enter_range(v, range_end)) { in br_vlan_dump_dev()
1838 range_end->vid, range_start, in br_vlan_dump_dev()
1844 idx += range_end->vid - range_start->vid + 1; in br_vlan_dump_dev()
1848 range_end = v; in br_vlan_dump_dev()
[all …]
Dbr_private.h1094 const struct net_bridge_vlan *range_end);
1310 const struct net_bridge_vlan *range_end) in br_vlan_can_enter_range() argument
1319 const struct net_bridge_vlan *range_end);
1325 struct net_bridge_vlan *range_end,
/kernel/linux/linux-5.10/drivers/infiniband/hw/hfi1/
Dfault.c156 unsigned long range_start, range_end, i; in fault_opcodes_write() local
174 if (kstrtoul(token, 0, &range_end)) in fault_opcodes_write()
177 range_end = range_start; in fault_opcodes_write()
179 if (range_start == range_end && range_start == -1UL) { in fault_opcodes_write()
185 if (range_start >= bound || range_end >= bound) in fault_opcodes_write()
188 for (i = range_start; i <= range_end; i++) { in fault_opcodes_write()
/kernel/linux/linux-5.10/arch/xtensa/mm/
Dinit.c86 phys_addr_t range_start, range_end; in free_highpages() local
91 &range_start, &range_end, NULL) { in free_highpages()
93 unsigned long end = PFN_DOWN(range_end); in free_highpages()
/kernel/linux/linux-5.10/drivers/gpu/drm/selftests/
Dtest-drm_mm.c765 u64 range_start, u64 range_end, in expect_insert_in_range() argument
772 range_start, range_end, in expect_insert_in_range()
777 range_start, range_end, err); in expect_insert_in_range()
792 u64 range_end) in expect_insert_in_range_fail() argument
799 range_start, range_end, in expect_insert_in_range_fail()
806 tmp.start, tmp.size, range_start, range_end); in expect_insert_in_range_fail()
810 err, -ENOSPC, size, range_start, range_end); in expect_insert_in_range_fail()
1448 u64 range_start, u64 range_end, in evict_something() argument
1464 range_start, range_end, in evict_something()
1482 if (tmp.start < range_start || tmp.start + tmp.size > range_end) { in evict_something()
[all …]
/kernel/linux/linux-5.10/drivers/gpu/drm/
Ddrm_mm.c521 u64 range_start, u64 range_end, in drm_mm_insert_node_in_range() argument
528 DRM_MM_BUG_ON(range_start > range_end); in drm_mm_insert_node_in_range()
530 if (unlikely(size == 0 || range_end - range_start < size)) in drm_mm_insert_node_in_range()
543 for (hole = first_hole(mm, range_start, range_end, size, mode); in drm_mm_insert_node_in_range()
551 if (mode == DRM_MM_INSERT_LOW && hole_start >= range_end) in drm_mm_insert_node_in_range()
563 adj_end = min(col_end, range_end); in drm_mm_insert_node_in_range()
584 min(col_end, range_end) - adj_start < size) in drm_mm_insert_node_in_range()
765 scan->range_end = end; in drm_mm_scan_init_with_range()
816 adj_end = min(col_end, scan->range_end); in drm_mm_scan_add_block()
835 min(col_end, scan->range_end) - adj_start < scan->size) in drm_mm_scan_add_block()
/kernel/linux/linux-5.10/arch/powerpc/platforms/powernv/
Dopal-prd.c57 uint64_t range_addr, range_size, range_end; in opal_prd_range_is_valid() local
64 range_end = range_addr + range_size; in opal_prd_range_is_valid()
72 if (range_end <= range_addr) in opal_prd_range_is_valid()
75 if (addr >= range_addr && addr + size <= range_end) { in opal_prd_range_is_valid()
/kernel/linux/linux-5.10/drivers/gpu/drm/msm/
Dmsm_gem_vma.c119 u64 range_start, u64 range_end) in msm_gem_init_vma() argument
128 0, range_start, range_end, 0); in msm_gem_init_vma()
Dmsm_drv.h244 u64 range_start, u64 range_end);
286 u64 range_start, u64 range_end);
Dmsm_gem.c382 u64 range_start, u64 range_end) in msm_gem_get_iova_locked() argument
398 range_start, range_end); in msm_gem_get_iova_locked()
446 u64 range_start, u64 range_end) in msm_gem_get_and_pin_iova_range() argument
455 range_start, range_end); in msm_gem_get_and_pin_iova_range()
/kernel/linux/linux-5.10/drivers/gpu/drm/amd/amdkfd/
Dkfd_device.c1146 (*mem_obj)->range_end = found; in kfd_gtt_sa_allocate()
1169 (*mem_obj)->range_end = in kfd_gtt_sa_allocate()
1177 if ((*mem_obj)->range_end != found) { in kfd_gtt_sa_allocate()
1197 (*mem_obj)->range_start, (*mem_obj)->range_end); in kfd_gtt_sa_allocate()
1201 found <= (*mem_obj)->range_end; in kfd_gtt_sa_allocate()
1225 mem_obj, mem_obj->range_start, mem_obj->range_end); in kfd_gtt_sa_free()
1231 bit <= mem_obj->range_end; in kfd_gtt_sa_free()
/kernel/linux/linux-5.10/fs/ocfs2/
Dfile.c866 u64 *range_start, u64 *range_end) in ocfs2_zero_extend_get_range() argument
894 *range_end = 0; in ocfs2_zero_extend_get_range()
926 *range_end = ocfs2_clusters_to_bytes(inode->i_sb, in ocfs2_zero_extend_get_range()
938 u64 range_end, struct buffer_head *di_bh) in ocfs2_zero_extend_range() argument
947 (unsigned long long)range_end); in ocfs2_zero_extend_range()
948 BUG_ON(range_start >= range_end); in ocfs2_zero_extend_range()
950 while (zero_pos < range_end) { in ocfs2_zero_extend_range()
952 if (next_pos > range_end) in ocfs2_zero_extend_range()
953 next_pos = range_end; in ocfs2_zero_extend_range()
975 u64 zero_start, range_start = 0, range_end = 0; in ocfs2_zero_extend() local
[all …]
/kernel/linux/linux-5.10/arch/ia64/kernel/
Dsetup.c146 u64 range_start, range_end, prev_start; in filter_rsvd_memory() local
165 range_end = min(end, rsvd_region[i].start); in filter_rsvd_memory()
167 if (range_start < range_end) in filter_rsvd_memory()
168 call_pernode_memory(__pa(range_start), range_end - range_start, func); in filter_rsvd_memory()
171 if (range_end == end) return 0; in filter_rsvd_memory()
/kernel/linux/linux-5.10/drivers/misc/habanalabs/goya/
Dgoya_coresight.c368 u64 range_start, range_end; in goya_etr_validate_address() local
378 range_end = prop->dmmu.end_addr; in goya_etr_validate_address()
381 range_end = prop->dram_end_address; in goya_etr_validate_address()
384 return hl_mem_area_inside_range(addr, size, range_start, range_end); in goya_etr_validate_address()
/kernel/linux/linux-5.10/fs/btrfs/
Dextent_map.c87 static u64 range_end(u64 start, u64 len) in range_end() function
100 u64 end = range_end(em->start, em->len); in tree_insert()
425 u64 end = range_end(start, len); in __lookup_extent_mapping()
Dreflink.c56 const u64 range_end = file_offset + block_size - 1; in copy_inline_to_page() local
85 clear_extent_bit(&inode->io_tree, file_offset, range_end, in copy_inline_to_page()
88 ret = btrfs_set_extent_delalloc(inode, file_offset, range_end, 0, NULL); in copy_inline_to_page()
/kernel/linux/linux-5.10/include/trace/events/
Dwriteback.h456 __field(long, range_end)
470 __entry->range_end = (long)wbc->range_end;
486 __entry->range_end,
/kernel/linux/linux-5.10/drivers/net/ethernet/netronome/nfp/bpf/
Djit.c2734 s16 range_end = meta->pkt_cache.range_end; in mem_ldx_data_init_pktcache() local
2741 len = range_end - range_start; in mem_ldx_data_init_pktcache()
2866 if (meta->pkt_cache.range_end) { in mem_ldx()
4273 s16 range_start = 0, range_end = 0; in nfp_bpf_opt_pkt_cache() local
4326 s16 new_end = range_end; in nfp_bpf_opt_pkt_cache()
4335 if (end > range_end) { in nfp_bpf_opt_pkt_cache()
4346 range_end = new_end; in nfp_bpf_opt_pkt_cache()
4353 range_node->pkt_cache.range_end = range_end; in nfp_bpf_opt_pkt_cache()
4360 range_end = insn->off + BPF_LDST_BYTES(insn); in nfp_bpf_opt_pkt_cache()
4365 range_node->pkt_cache.range_end = range_end; in nfp_bpf_opt_pkt_cache()
[all …]
/kernel/linux/linux-5.10/include/drm/
Ddrm_mm.h234 u64 range_end; member
/kernel/linux/linux-5.10/drivers/gpu/drm/msm/adreno/
Da6xx_gmu.c1132 u64 range_start, range_end; in a6xx_gmu_memory_alloc() local
1139 range_end = 0x80000000; in a6xx_gmu_memory_alloc()
1143 range_end = iova + size; in a6xx_gmu_memory_alloc()
1153 range_start >> PAGE_SHIFT, range_end >> PAGE_SHIFT); in a6xx_gmu_memory_alloc()
/kernel/linux/linux-5.10/drivers/net/ethernet/cisco/enic/
Denic.h79 u32 range_end; member

123