Home
last modified time | relevance | path

Searched refs:value_size (Results 1 – 16 of 16) sorted by relevance

/kernel/bpf/
Dqueue_stack_maps.c53 attr->value_size == 0 || in queue_stack_map_alloc_check()
58 if (attr->value_size > KMALLOC_MAX_SIZE) in queue_stack_map_alloc_check()
74 queue_size = sizeof(*qs) + size * attr->value_size; in queue_stack_map_alloc()
114 memset(value, 0, qs->map.value_size); in __queue_map_get()
119 ptr = &qs->elements[qs->tail * qs->map.value_size]; in __queue_map_get()
120 memcpy(value, ptr, qs->map.value_size); in __queue_map_get()
149 memset(value, 0, qs->map.value_size); in __stack_map_get()
158 ptr = &qs->elements[index * qs->map.value_size]; in __stack_map_get()
159 memcpy(value, ptr, qs->map.value_size); in __stack_map_get()
228 dst = &qs->elements[qs->head * qs->map.value_size]; in queue_stack_map_push_elem()
[all …]
Dmap_iter.c105 u32 key_acc_size, value_acc_size, key_size, value_size; in bpf_iter_attach_map() local
130 value_size = map->value_size; in bpf_iter_attach_map()
132 value_size = round_up(map->value_size, 8) * num_possible_cpus(); in bpf_iter_attach_map()
134 if (key_acc_size > key_size || value_acc_size > value_size) { in bpf_iter_attach_map()
Dstackmap.c67 (u64)smap->map.value_size; in prealloc_elems_and_freelist()
91 u32 value_size = attr->value_size; in stack_map_alloc() local
104 value_size < 8 || value_size % 8) in stack_map_alloc()
109 if (value_size % sizeof(struct bpf_stack_build_id) || in stack_map_alloc()
110 value_size / sizeof(struct bpf_stack_build_id) in stack_map_alloc()
113 } else if (value_size / 8 > sysctl_perf_event_max_stack) in stack_map_alloc()
127 smap->map.value_size = value_size; in stack_map_alloc()
330 u32 max_depth = map->value_size / stack_map_data_size(map); in BPF_CALL_3()
637 memset(value + trace_len, 0, map->value_size - trace_len); in bpf_stackmap_copy()
Dmap_in_map.c50 inner_map_meta->value_size = inner_map->value_size; in bpf_map_meta_alloc()
84 meta0->value_size == meta1->value_size && in bpf_map_meta_equal()
Dlocal_storage.c167 map->value_size, in cgroup_storage_update_elem()
173 memcpy(&new->data[0], value, map->value_size); in cgroup_storage_update_elem()
201 size = round_up(_map->value_size, 8); in bpf_percpu_cgroup_storage_copy()
235 size = round_up(_map->value_size, 8); in bpf_percpu_cgroup_storage_update()
302 if (attr->value_size == 0) in cgroup_storage_map_alloc()
305 if (attr->value_size > max_value_size) in cgroup_storage_map_alloc()
481 size = sizeof(struct bpf_storage_buffer) + map->value_size; in bpf_cgroup_storage_calculate_size()
485 size = map->value_size; in bpf_cgroup_storage_calculate_size()
Dbpf_struct_ops.c256 memset(value, 0, map->value_size); in bpf_struct_ops_map_sys_lookup_elem()
264 memcpy(uvalue, st_map->uvalue, map->value_size); in bpf_struct_ops_map_sys_lookup_elem()
361 memcpy(uvalue, value, map->value_size); in bpf_struct_ops_map_update_elem()
478 memset(uvalue, 0, map->value_size); in bpf_struct_ops_map_update_elem()
479 memset(kvalue, 0, map->value_size); in bpf_struct_ops_map_update_elem()
518 value = kmalloc(map->value_size, GFP_USER | __GFP_NOWARN); in bpf_struct_ops_map_seq_show_elem()
568 if (attr->value_size != vt->size) in bpf_struct_ops_map_alloc()
Dhashtab.c332 u32 size = round_up(htab->map.value_size, 8); in prealloc_init()
456 attr->value_size == 0) in htab_map_alloc_check()
459 if ((u64)attr->key_size + attr->value_size >= KMALLOC_MAX_SIZE - in htab_map_alloc_check()
515 htab->elem_size += round_up(htab->map.value_size, 8); in htab_map_alloc()
883 memcpy(this_cpu_ptr(pptr), value, htab->map.value_size); in pcpu_copy_value()
885 u32 size = round_up(htab->map.value_size, 8); in pcpu_copy_value()
906 u32 size = round_up(htab->map.value_size, 8); in pcpu_init_value()
933 u32 size = htab->map.value_size; in alloc_htab_elem()
1528 u32 roundup_value_size = round_up(map->value_size, 8); in __htab_map_lookup_and_delete_elem()
1603 u32 bucket_cnt, total, key_size, value_size, roundup_key_size; in __htab_map_lookup_and_delete_batch() local
[all …]
Darraymap.c58 attr->value_size == 0 || in array_map_alloc_check()
72 if (attr->value_size > KMALLOC_MAX_SIZE) in array_map_alloc_check()
90 elem_size = round_up(attr->value_size, 8); in array_map_alloc()
177 if (off >= map->value_size) in array_map_direct_value_addr()
205 u32 elem_size = round_up(map->value_size, 8); in array_map_gen_lookup()
260 size = round_up(map->value_size, 8); in bpf_percpu_array_copy()
322 value, map->value_size); in array_map_update_elem()
362 size = round_up(map->value_size, 8); in bpf_percpu_array_update()
581 size = round_up(map->value_size, 8); in __bpf_array_map_seq_show()
615 buf_size = round_up(map->value_size, 8) * num_possible_cpus(); in bpf_iter_init_array_map()
[all …]
Dreuseport_array.c45 if (attr->value_size != sizeof(u32) && in reuseport_array_alloc_check()
46 attr->value_size != sizeof(u64)) in reuseport_array_alloc_check()
180 if (map->value_size != sizeof(u64)) in bpf_fd_reuseport_array_lookup_elem()
258 if (map->value_size == sizeof(u64)) { in bpf_fd_reuseport_array_update_elem()
Dsyscall.c158 return round_up(map->value_size, 8) * num_possible_cpus(); in bpf_map_value_size()
162 return map->value_size; in bpf_map_value_size()
364 map->value_size = attr->value_size; in bpf_map_init_from_attr()
580 map->value_size, in bpf_map_show_fdinfo()
771 u32 key_size, value_size; in map_check_btf() local
785 value_type = btf_type_id_size(btf, &btf_value_id, &value_size); in map_check_btf()
786 if (!value_type || value_size != map->value_size) in map_check_btf()
802 map->value_size) { in map_check_btf()
805 map->spin_lock_off, map->value_size); in map_check_btf()
1058 u32 value_size; in map_lookup_elem() local
[all …]
Dcpumap.c87 u32 value_size = attr->value_size; in cpu_map_alloc() local
96 (value_size != offsetofend(struct bpf_cpumap_val, qsize) && in cpu_map_alloc()
97 value_size != offsetofend(struct bpf_cpumap_val, bpf_prog.fd)) || in cpu_map_alloc()
578 memcpy(&cpumap_value, value, map->value_size); in cpu_map_update_elem()
Dlpm_trie.c288 size += trie->map.value_size; in lpm_trie_node_alloc()
299 trie->map.value_size); in lpm_trie_node_alloc()
559 attr->value_size < LPM_VAL_SIZE_MIN || in trie_alloc()
560 attr->value_size > LPM_VAL_SIZE_MAX) in trie_alloc()
Dbpf_local_storage.c554 attr->key_size != sizeof(int) || !attr->value_size || in bpf_local_storage_map_alloc_check()
562 if (attr->value_size > BPF_LOCAL_STORAGE_MAX_VALUE_SIZE) in bpf_local_storage_map_alloc_check()
597 sizeof(struct bpf_local_storage_elem) + attr->value_size; in bpf_local_storage_map_alloc()
Ddevmap.c112 u32 valsize = attr->value_size; in dev_map_init_map()
912 memcpy(&val, value, map->value_size); in __dev_map_update_elem()
954 memcpy(&val, value, map->value_size); in __dev_map_hash_update_elem()
Dringbuf.c155 if (attr->key_size || attr->value_size || in ringbuf_map_alloc()
Dverifier.c668 reg->map_ptr->value_size); in print_verifier_state()
3357 map->value_size, off, size); in check_map_access_type()
3363 map->value_size, off, size); in check_map_access_type()
3475 err = check_mem_region_access(env, regno, off, size, map->value_size, in check_map_access()
5398 meta->map_ptr->value_size, false, in check_func_arg()
5512 map->value_size - reg->off, false); in check_func_arg()
5524 if (!strnchr(str_ptr + map_off, map->value_size - map_off, 0)) { in check_func_arg()
6951 max = ptr_reg->map_ptr->value_size; in retrieve_ptr_limit()
11878 map->value_size, off); in resolve_pseudo_ldimm64()