/kernel/bpf/ |
D | queue_stack_maps.c | 53 attr->value_size == 0 || in queue_stack_map_alloc_check() 58 if (attr->value_size > KMALLOC_MAX_SIZE) in queue_stack_map_alloc_check() 74 queue_size = sizeof(*qs) + size * attr->value_size; in queue_stack_map_alloc() 114 memset(value, 0, qs->map.value_size); in __queue_map_get() 119 ptr = &qs->elements[qs->tail * qs->map.value_size]; in __queue_map_get() 120 memcpy(value, ptr, qs->map.value_size); in __queue_map_get() 149 memset(value, 0, qs->map.value_size); in __stack_map_get() 158 ptr = &qs->elements[index * qs->map.value_size]; in __stack_map_get() 159 memcpy(value, ptr, qs->map.value_size); in __stack_map_get() 228 dst = &qs->elements[qs->head * qs->map.value_size]; in queue_stack_map_push_elem() [all …]
|
D | map_iter.c | 105 u32 key_acc_size, value_acc_size, key_size, value_size; in bpf_iter_attach_map() local 130 value_size = map->value_size; in bpf_iter_attach_map() 132 value_size = round_up(map->value_size, 8) * num_possible_cpus(); in bpf_iter_attach_map() 134 if (key_acc_size > key_size || value_acc_size > value_size) { in bpf_iter_attach_map()
|
D | stackmap.c | 67 (u64)smap->map.value_size; in prealloc_elems_and_freelist() 91 u32 value_size = attr->value_size; in stack_map_alloc() local 104 value_size < 8 || value_size % 8) in stack_map_alloc() 109 if (value_size % sizeof(struct bpf_stack_build_id) || in stack_map_alloc() 110 value_size / sizeof(struct bpf_stack_build_id) in stack_map_alloc() 113 } else if (value_size / 8 > sysctl_perf_event_max_stack) in stack_map_alloc() 127 smap->map.value_size = value_size; in stack_map_alloc() 330 u32 max_depth = map->value_size / stack_map_data_size(map); in BPF_CALL_3() 637 memset(value + trace_len, 0, map->value_size - trace_len); in bpf_stackmap_copy()
|
D | map_in_map.c | 50 inner_map_meta->value_size = inner_map->value_size; in bpf_map_meta_alloc() 84 meta0->value_size == meta1->value_size && in bpf_map_meta_equal()
|
D | local_storage.c | 167 map->value_size, in cgroup_storage_update_elem() 173 memcpy(&new->data[0], value, map->value_size); in cgroup_storage_update_elem() 201 size = round_up(_map->value_size, 8); in bpf_percpu_cgroup_storage_copy() 235 size = round_up(_map->value_size, 8); in bpf_percpu_cgroup_storage_update() 302 if (attr->value_size == 0) in cgroup_storage_map_alloc() 305 if (attr->value_size > max_value_size) in cgroup_storage_map_alloc() 481 size = sizeof(struct bpf_storage_buffer) + map->value_size; in bpf_cgroup_storage_calculate_size() 485 size = map->value_size; in bpf_cgroup_storage_calculate_size()
|
D | bpf_struct_ops.c | 256 memset(value, 0, map->value_size); in bpf_struct_ops_map_sys_lookup_elem() 264 memcpy(uvalue, st_map->uvalue, map->value_size); in bpf_struct_ops_map_sys_lookup_elem() 361 memcpy(uvalue, value, map->value_size); in bpf_struct_ops_map_update_elem() 478 memset(uvalue, 0, map->value_size); in bpf_struct_ops_map_update_elem() 479 memset(kvalue, 0, map->value_size); in bpf_struct_ops_map_update_elem() 518 value = kmalloc(map->value_size, GFP_USER | __GFP_NOWARN); in bpf_struct_ops_map_seq_show_elem() 568 if (attr->value_size != vt->size) in bpf_struct_ops_map_alloc()
|
D | hashtab.c | 332 u32 size = round_up(htab->map.value_size, 8); in prealloc_init() 456 attr->value_size == 0) in htab_map_alloc_check() 459 if ((u64)attr->key_size + attr->value_size >= KMALLOC_MAX_SIZE - in htab_map_alloc_check() 515 htab->elem_size += round_up(htab->map.value_size, 8); in htab_map_alloc() 883 memcpy(this_cpu_ptr(pptr), value, htab->map.value_size); in pcpu_copy_value() 885 u32 size = round_up(htab->map.value_size, 8); in pcpu_copy_value() 906 u32 size = round_up(htab->map.value_size, 8); in pcpu_init_value() 933 u32 size = htab->map.value_size; in alloc_htab_elem() 1528 u32 roundup_value_size = round_up(map->value_size, 8); in __htab_map_lookup_and_delete_elem() 1603 u32 bucket_cnt, total, key_size, value_size, roundup_key_size; in __htab_map_lookup_and_delete_batch() local [all …]
|
D | arraymap.c | 58 attr->value_size == 0 || in array_map_alloc_check() 72 if (attr->value_size > KMALLOC_MAX_SIZE) in array_map_alloc_check() 90 elem_size = round_up(attr->value_size, 8); in array_map_alloc() 177 if (off >= map->value_size) in array_map_direct_value_addr() 205 u32 elem_size = round_up(map->value_size, 8); in array_map_gen_lookup() 260 size = round_up(map->value_size, 8); in bpf_percpu_array_copy() 322 value, map->value_size); in array_map_update_elem() 362 size = round_up(map->value_size, 8); in bpf_percpu_array_update() 581 size = round_up(map->value_size, 8); in __bpf_array_map_seq_show() 615 buf_size = round_up(map->value_size, 8) * num_possible_cpus(); in bpf_iter_init_array_map() [all …]
|
D | reuseport_array.c | 45 if (attr->value_size != sizeof(u32) && in reuseport_array_alloc_check() 46 attr->value_size != sizeof(u64)) in reuseport_array_alloc_check() 180 if (map->value_size != sizeof(u64)) in bpf_fd_reuseport_array_lookup_elem() 258 if (map->value_size == sizeof(u64)) { in bpf_fd_reuseport_array_update_elem()
|
D | syscall.c | 158 return round_up(map->value_size, 8) * num_possible_cpus(); in bpf_map_value_size() 162 return map->value_size; in bpf_map_value_size() 364 map->value_size = attr->value_size; in bpf_map_init_from_attr() 580 map->value_size, in bpf_map_show_fdinfo() 771 u32 key_size, value_size; in map_check_btf() local 785 value_type = btf_type_id_size(btf, &btf_value_id, &value_size); in map_check_btf() 786 if (!value_type || value_size != map->value_size) in map_check_btf() 802 map->value_size) { in map_check_btf() 805 map->spin_lock_off, map->value_size); in map_check_btf() 1058 u32 value_size; in map_lookup_elem() local [all …]
|
D | cpumap.c | 87 u32 value_size = attr->value_size; in cpu_map_alloc() local 96 (value_size != offsetofend(struct bpf_cpumap_val, qsize) && in cpu_map_alloc() 97 value_size != offsetofend(struct bpf_cpumap_val, bpf_prog.fd)) || in cpu_map_alloc() 578 memcpy(&cpumap_value, value, map->value_size); in cpu_map_update_elem()
|
D | lpm_trie.c | 288 size += trie->map.value_size; in lpm_trie_node_alloc() 299 trie->map.value_size); in lpm_trie_node_alloc() 559 attr->value_size < LPM_VAL_SIZE_MIN || in trie_alloc() 560 attr->value_size > LPM_VAL_SIZE_MAX) in trie_alloc()
|
D | bpf_local_storage.c | 554 attr->key_size != sizeof(int) || !attr->value_size || in bpf_local_storage_map_alloc_check() 562 if (attr->value_size > BPF_LOCAL_STORAGE_MAX_VALUE_SIZE) in bpf_local_storage_map_alloc_check() 597 sizeof(struct bpf_local_storage_elem) + attr->value_size; in bpf_local_storage_map_alloc()
|
D | devmap.c | 112 u32 valsize = attr->value_size; in dev_map_init_map() 912 memcpy(&val, value, map->value_size); in __dev_map_update_elem() 954 memcpy(&val, value, map->value_size); in __dev_map_hash_update_elem()
|
D | ringbuf.c | 155 if (attr->key_size || attr->value_size || in ringbuf_map_alloc()
|
D | verifier.c | 668 reg->map_ptr->value_size); in print_verifier_state() 3357 map->value_size, off, size); in check_map_access_type() 3363 map->value_size, off, size); in check_map_access_type() 3475 err = check_mem_region_access(env, regno, off, size, map->value_size, in check_map_access() 5398 meta->map_ptr->value_size, false, in check_func_arg() 5512 map->value_size - reg->off, false); in check_func_arg() 5524 if (!strnchr(str_ptr + map_off, map->value_size - map_off, 0)) { in check_func_arg() 6951 max = ptr_reg->map_ptr->value_size; in retrieve_ptr_limit() 11878 map->value_size, off); in resolve_pseudo_ldimm64()
|