Searched refs:max_entries (Results 1 – 4 of 4) sorted by relevance
68 if (attr->max_entries == 0 || attr->key_size != 4 || in xsk_map_alloc()74 size = struct_size(m, xsk_map, attr->max_entries); in xsk_map_alloc()100 if (index >= m->map.max_entries) { in xsk_map_get_next_key()105 if (index == m->map.max_entries - 1) in xsk_map_get_next_key()117 *insn++ = BPF_JMP_IMM(BPF_JGE, ret, map->max_entries, 5); in xsk_map_gen_lookup()135 if (key >= map->max_entries) in __xsk_map_lookup_elem()164 if (unlikely(i >= m->map.max_entries)) in xsk_map_update_elem()219 if (k >= map->max_entries) in xsk_map_delete_elem()252 return meta0->max_entries == meta1->max_entries && in xsk_map_meta_equal()
361 static u32 xsk_tx_peek_release_fallback(struct xsk_buff_pool *pool, u32 max_entries) in xsk_tx_peek_release_fallback() argument366 while (nb_pkts < max_entries && xsk_tx_peek_desc(pool, &descs[nb_pkts])) in xsk_tx_peek_release_fallback()
37 if (attr->max_entries == 0 || in sock_map_alloc()51 stab->sks = bpf_map_area_alloc((u64) stab->map.max_entries * in sock_map_alloc()344 for (i = 0; i < stab->map.max_entries; i++) { in sock_map_free()378 if (unlikely(key >= map->max_entries)) in __sock_map_lookup_elem()445 if (unlikely(i >= map->max_entries)) in sock_map_delete_elem()458 if (i == stab->map.max_entries - 1) in sock_map_get_next_key()460 if (i >= stab->map.max_entries) in sock_map_get_next_key()479 if (unlikely(idx >= map->max_entries)) in sock_map_update_common()717 if (unlikely(info->index >= info->map->max_entries)) in DEFINE_BPF_ITER_FUNC()958 if (atomic_inc_return(&htab->count) > htab->map.max_entries) { in sock_hash_alloc_elem()[all …]
4468 if (unlikely(idx >= array->map.max_entries)) in BPF_CALL_3()