Home
last modified time | relevance | path

Searched refs:chunk_list (Results 1 – 19 of 19) sorted by relevance

/kernel/linux/linux-5.10/arch/powerpc/platforms/ps3/
Dmm.c404 list_for_each_entry(c, &r->chunk_list.head, link) { in dma_find_chunk()
433 list_for_each_entry(c, &r->chunk_list.head, link) { in dma_find_chunk_lpar()
539 list_add(&c->link, &r->chunk_list.head); in dma_sb_map_pages()
573 if (list_empty(&r->chunk_list.head)) { in dma_ioc0_map_pages()
578 last = list_entry(r->chunk_list.head.next, in dma_ioc0_map_pages()
609 list_add(&c->link, &r->chunk_list.head); in dma_ioc0_map_pages()
659 INIT_LIST_HEAD(&r->chunk_list.head); in dma_sb_region_create()
660 spin_lock_init(&r->chunk_list.lock); in dma_sb_region_create()
681 INIT_LIST_HEAD(&r->chunk_list.head); in dma_ioc0_region_create()
682 spin_lock_init(&r->chunk_list.lock); in dma_ioc0_region_create()
[all …]
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx4/
Dicm.h68 struct list_head chunk_list; member
100 iter->chunk = list_empty(&icm->chunk_list) ? in mlx4_icm_first()
101 NULL : list_entry(icm->chunk_list.next, in mlx4_icm_first()
114 if (iter->chunk->list.next == &iter->icm->chunk_list) { in mlx4_icm_next()
Dicm.c86 list_for_each_entry_safe(chunk, tmp, &icm->chunk_list, list) { in mlx4_free_icm()
155 INIT_LIST_HEAD(&icm->chunk_list); in mlx4_alloc_icm()
176 list_add_tail(&chunk->list, &icm->chunk_list); in mlx4_alloc_icm()
335 list_for_each_entry(chunk, &icm->chunk_list, list) { in mlx4_table_find()
/kernel/linux/linux-5.10/drivers/infiniband/hw/usnic/
Dusnic_uiom.c66 static void usnic_uiom_put_pages(struct list_head *chunk_list, int dirty) in usnic_uiom_put_pages() argument
74 list_for_each_entry_safe(chunk, tmp, chunk_list, list) { in usnic_uiom_put_pages()
88 struct list_head *chunk_list = &uiomr->chunk_list; in usnic_uiom_get_pages() local
117 INIT_LIST_HEAD(chunk_list); in usnic_uiom_get_pages()
177 list_add_tail(&chunk->list, chunk_list); in usnic_uiom_get_pages()
185 usnic_uiom_put_pages(chunk_list, 0); in usnic_uiom_get_pages()
240 usnic_uiom_put_pages(&uiomr->chunk_list, dirty & writable); in __usnic_uiom_reg_release()
260 chunk = list_first_entry(&uiomr->chunk_list, struct usnic_uiom_chunk, in usnic_uiom_map_sorted_intervals()
415 usnic_uiom_put_pages(&uiomr->chunk_list, 0); in usnic_uiom_reg_get()
Dusnic_uiom.h72 struct list_head chunk_list; member
/kernel/linux/linux-5.10/drivers/infiniband/hw/hns/
Dhns_roce_hem.h90 struct list_head chunk_list; member
147 iter->chunk = list_empty(&hem->chunk_list) ? NULL : in hns_roce_hem_first()
148 list_entry(hem->chunk_list.next, in hns_roce_hem_first()
161 if (iter->chunk->list.next == &iter->hem->chunk_list) { in hns_roce_hem_next()
Dhns_roce_hem.c264 INIT_LIST_HEAD(&hem->chunk_list); in hns_roce_alloc_hem()
279 list_add_tail(&chunk->list, &hem->chunk_list); in hns_roce_alloc_hem()
318 list_for_each_entry_safe(chunk, tmp, &hem->chunk_list, list) { in hns_roce_free_hem()
835 list_for_each_entry(chunk, &hem->chunk_list, list) { in hns_roce_table_find()
/kernel/linux/linux-5.10/drivers/infiniband/hw/mthca/
Dmthca_memfree.h59 struct list_head chunk_list; member
103 iter->chunk = list_empty(&icm->chunk_list) ? in mthca_icm_first()
104 NULL : list_entry(icm->chunk_list.next, in mthca_icm_first()
117 if (iter->chunk->list.next == &iter->icm->chunk_list) { in mthca_icm_next()
Dmthca_memfree.c95 list_for_each_entry_safe(chunk, tmp, &icm->chunk_list, list) { in mthca_free_icm()
153 INIT_LIST_HEAD(&icm->chunk_list); in mthca_alloc_icm()
167 list_add_tail(&chunk->list, &icm->chunk_list); in mthca_alloc_icm()
297 list_for_each_entry(chunk, &icm->chunk_list, list) { in mthca_table_find()
/kernel/linux/linux-5.10/drivers/infiniband/hw/efa/
Defa_verbs.c105 struct pbl_chunk_list chunk_list; member
1191 struct pbl_chunk_list *chunk_list = &pbl->phys.indirect.chunk_list; in pbl_chunk_list_create() local
1205 chunk_list->size = chunk_list_size; in pbl_chunk_list_create()
1206 chunk_list->chunks = kcalloc(chunk_list_size, in pbl_chunk_list_create()
1207 sizeof(*chunk_list->chunks), in pbl_chunk_list_create()
1209 if (!chunk_list->chunks) in pbl_chunk_list_create()
1218 chunk_list->chunks[i].buf = kzalloc(EFA_CHUNK_SIZE, GFP_KERNEL); in pbl_chunk_list_create()
1219 if (!chunk_list->chunks[i].buf) in pbl_chunk_list_create()
1222 chunk_list->chunks[i].length = EFA_CHUNK_USED_SIZE; in pbl_chunk_list_create()
1224 chunk_list->chunks[chunk_list_size - 1].length = in pbl_chunk_list_create()
[all …]
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/steering/
Ddr_icm_pool.c258 INIT_LIST_HEAD(&chunk->chunk_list); in dr_icm_chunks_create()
259 list_add(&chunk->chunk_list, &bucket->free_list); in dr_icm_chunks_create()
284 list_del(&chunk->chunk_list); in dr_icm_chunk_destroy()
319 list_for_each_entry_safe(chunk, next, &bucket->free_list, chunk_list) in dr_icm_bucket_cleanup()
325 list_for_each_entry_safe(chunk, next, &bucket->used_list, chunk_list) in dr_icm_bucket_cleanup()
481 chunk_list); in mlx5dr_icm_alloc_chunk()
483 list_del_init(&chunk->chunk_list); in mlx5dr_icm_alloc_chunk()
484 list_add_tail(&chunk->chunk_list, &bucket->used_list); in mlx5dr_icm_alloc_chunk()
506 list_del_init(&chunk->chunk_list); in mlx5dr_icm_free_chunk()
507 list_add_tail(&chunk->chunk_list, &bucket->hot_list); in mlx5dr_icm_free_chunk()
Ddr_types.h813 struct list_head chunk_list; member
/kernel/linux/linux-5.10/net/sctp/
Doutput.c151 INIT_LIST_HEAD(&packet->chunk_list); in sctp_packet_init()
165 list_for_each_entry_safe(chunk, tmp, &packet->chunk_list, list) { in sctp_packet_free()
347 list_add_tail(&chunk->list, &packet->chunk_list); in __sctp_packet_append_chunk()
429 list_for_each_entry_safe(chunk, tmp, &packet->chunk_list, in sctp_packet_pack()
450 list_for_each_entry_safe(chunk, tmp, &packet->chunk_list, list) { in sctp_packet_pack()
493 if (list_empty(&packet->chunk_list)) in sctp_packet_pack()
497 &packet->chunk_list); in sctp_packet_pack()
504 } while (!list_empty(&packet->chunk_list)); in sctp_packet_pack()
557 if (list_empty(&packet->chunk_list)) in sctp_packet_transmit()
559 chunk = list_entry(packet->chunk_list.next, struct sctp_chunk, list); in sctp_packet_transmit()
[all …]
/kernel/linux/linux-5.10/drivers/gpu/drm/qxl/
Dqxl_image.c51 list_add_tail(&chunk->head, &image->chunk_list); in qxl_allocate_chunk()
68 INIT_LIST_HEAD(&image->chunk_list); in qxl_image_alloc_objects()
90 list_for_each_entry_safe(chunk, tmp, &dimage->chunk_list, head) { in qxl_image_free_objects()
120 drv_chunk = list_first_entry(&dimage->chunk_list, struct qxl_drm_chunk, head); in qxl_image_init_helper()
Dqxl_drv.h167 struct list_head chunk_list; member
/kernel/linux/linux-5.10/drivers/s390/block/
Ddasd_int.h650 dasd_init_chunklist(struct list_head *chunk_list, void *mem, in dasd_init_chunklist() argument
655 INIT_LIST_HEAD(chunk_list); in dasd_init_chunklist()
658 list_add(&chunk->list, chunk_list); in dasd_init_chunklist()
662 dasd_alloc_chunk(struct list_head *chunk_list, unsigned long size) in dasd_alloc_chunk() argument
667 list_for_each_entry(chunk, chunk_list, list) { in dasd_alloc_chunk()
684 dasd_free_chunk(struct list_head *chunk_list, void *mem) in dasd_free_chunk() argument
692 left = chunk_list; in dasd_free_chunk()
693 list_for_each(p, chunk_list) { in dasd_free_chunk()
699 if (left->next != chunk_list) { in dasd_free_chunk()
707 if (left != chunk_list) { in dasd_free_chunk()
/kernel/linux/linux-5.10/drivers/crypto/
Dn2_core.c684 struct list_head chunk_list; member
861 INIT_LIST_HEAD(&rctx->chunk_list); in n2_compute_chunks()
894 &rctx->chunk_list); in n2_compute_chunks()
921 list_add_tail(&chunk->entry, &rctx->chunk_list); in n2_compute_chunks()
935 list_for_each_entry_safe(c, tmp, &rctx->chunk_list, entry) { in n2_chunk_complete()
962 list_for_each_entry_safe(c, tmp, &rctx->chunk_list, entry) { in n2_do_ecb()
1019 list_for_each_entry_safe(c, tmp, &rctx->chunk_list, in n2_do_chaining()
1032 list_for_each_entry_safe_reverse(c, tmp, &rctx->chunk_list, in n2_do_chaining()
/kernel/linux/linux-5.10/arch/powerpc/include/asm/
Dps3.h96 } chunk_list; member
/kernel/linux/linux-5.10/include/net/sctp/
Dstructs.h706 struct list_head chunk_list; member