/drivers/infiniband/hw/hns/ |
D | hns_roce_hem.h | 70 struct list_head chunk_list; member 105 iter->chunk = list_empty(&hem->chunk_list) ? NULL : in hns_roce_hem_first() 106 list_entry(hem->chunk_list.next, in hns_roce_hem_first() 119 if (iter->chunk->list.next == &iter->hem->chunk_list) { in hns_roce_hem_next()
|
D | hns_roce_hem.c | 62 INIT_LIST_HEAD(&hem->chunk_list); in hns_roce_alloc_hem() 76 list_add_tail(&chunk->list, &hem->chunk_list); in hns_roce_alloc_hem() 116 list_for_each_entry_safe(chunk, tmp, &hem->chunk_list, list) { in hns_roce_free_hem() 295 list_for_each_entry(chunk, &hem->chunk_list, list) { in hns_roce_table_find()
|
/drivers/net/ethernet/mellanox/mlx4/ |
D | icm.h | 58 struct list_head chunk_list; member 90 iter->chunk = list_empty(&icm->chunk_list) ? in mlx4_icm_first() 91 NULL : list_entry(icm->chunk_list.next, in mlx4_icm_first() 104 if (iter->chunk->list.next == &iter->icm->chunk_list) { in mlx4_icm_next()
|
D | icm.c | 85 list_for_each_entry_safe(chunk, tmp, &icm->chunk_list, list) { in mlx4_free_icm() 154 INIT_LIST_HEAD(&icm->chunk_list); in mlx4_alloc_icm() 175 list_add_tail(&chunk->list, &icm->chunk_list); in mlx4_alloc_icm() 331 list_for_each_entry(chunk, &icm->chunk_list, list) { in mlx4_table_find()
|
/drivers/infiniband/hw/usnic/ |
D | usnic_uiom.c | 80 static void usnic_uiom_put_pages(struct list_head *chunk_list, int dirty) in usnic_uiom_put_pages() argument 88 list_for_each_entry_safe(chunk, tmp, chunk_list, list) { in usnic_uiom_put_pages() 102 int dmasync, struct list_head *chunk_list) in usnic_uiom_get_pages() argument 131 INIT_LIST_HEAD(chunk_list); in usnic_uiom_get_pages() 190 list_add_tail(&chunk->list, chunk_list); in usnic_uiom_get_pages() 198 usnic_uiom_put_pages(chunk_list, 0); in usnic_uiom_get_pages() 252 usnic_uiom_put_pages(&uiomr->chunk_list, dirty & writable); in __usnic_uiom_reg_release() 272 chunk = list_first_entry(&uiomr->chunk_list, struct usnic_uiom_chunk, in usnic_uiom_map_sorted_intervals() 382 &uiomr->chunk_list); in usnic_uiom_reg_get() 427 usnic_uiom_put_pages(&uiomr->chunk_list, 0); in usnic_uiom_reg_get()
|
D | usnic_uiom.h | 72 struct list_head chunk_list; member
|
/drivers/infiniband/hw/mthca/ |
D | mthca_memfree.h | 59 struct list_head chunk_list; member 103 iter->chunk = list_empty(&icm->chunk_list) ? in mthca_icm_first() 104 NULL : list_entry(icm->chunk_list.next, in mthca_icm_first() 117 if (iter->chunk->list.next == &iter->icm->chunk_list) { in mthca_icm_next()
|
D | mthca_memfree.c | 95 list_for_each_entry_safe(chunk, tmp, &icm->chunk_list, list) { in mthca_free_icm() 153 INIT_LIST_HEAD(&icm->chunk_list); in mthca_alloc_icm() 167 list_add_tail(&chunk->list, &icm->chunk_list); in mthca_alloc_icm() 297 list_for_each_entry(chunk, &icm->chunk_list, list) { in mthca_table_find()
|
/drivers/gpu/drm/qxl/ |
D | qxl_image.c | 51 list_add_tail(&chunk->head, &image->chunk_list); in qxl_allocate_chunk() 68 INIT_LIST_HEAD(&image->chunk_list); in qxl_image_alloc_objects() 90 list_for_each_entry_safe(chunk, tmp, &dimage->chunk_list, head) { in qxl_image_free_objects() 120 drv_chunk = list_first_entry(&dimage->chunk_list, struct qxl_drm_chunk, head); in qxl_image_init_helper()
|
D | qxl_drv.h | 207 struct list_head chunk_list; member
|
/drivers/s390/block/ |
D | dasd_int.h | 642 dasd_init_chunklist(struct list_head *chunk_list, void *mem, in dasd_init_chunklist() argument 647 INIT_LIST_HEAD(chunk_list); in dasd_init_chunklist() 650 list_add(&chunk->list, chunk_list); in dasd_init_chunklist() 654 dasd_alloc_chunk(struct list_head *chunk_list, unsigned long size) in dasd_alloc_chunk() argument 659 list_for_each_entry(chunk, chunk_list, list) { in dasd_alloc_chunk() 676 dasd_free_chunk(struct list_head *chunk_list, void *mem) in dasd_free_chunk() argument 684 left = chunk_list; in dasd_free_chunk() 685 list_for_each(p, chunk_list) { in dasd_free_chunk() 691 if (left->next != chunk_list) { in dasd_free_chunk() 699 if (left != chunk_list) { in dasd_free_chunk()
|
/drivers/net/ethernet/mellanox/mlxsw/ |
D | spectrum_acl_tcam.c | 168 struct list_head chunk_list; /* List of chunks under this region */ member 309 if (list_empty(®ion->chunk_list)) in mlxsw_sp_acl_tcam_region_prio() 312 chunk = list_first_entry(®ion->chunk_list, typeof(*chunk), list); in mlxsw_sp_acl_tcam_region_prio() 321 if (list_empty(®ion->chunk_list)) in mlxsw_sp_acl_tcam_region_max_prio() 323 chunk = list_last_entry(®ion->chunk_list, typeof(*chunk), list); in mlxsw_sp_acl_tcam_region_max_prio() 707 INIT_LIST_HEAD(®ion->chunk_list); in mlxsw_sp_acl_tcam_region_create() 805 list_add_tail(&chunk->list, ®ion->chunk_list); in mlxsw_sp_acl_tcam_chunk_assoc() 828 if (list_empty(®ion->chunk_list)) { in mlxsw_sp_acl_tcam_chunk_deassoc()
|
/drivers/crypto/ |
D | n2_core.c | 678 struct list_head chunk_list; member 893 INIT_LIST_HEAD(&rctx->chunk_list); in n2_compute_chunks() 926 &rctx->chunk_list); in n2_compute_chunks() 953 list_add_tail(&chunk->entry, &rctx->chunk_list); in n2_compute_chunks() 968 list_for_each_entry_safe(c, tmp, &rctx->chunk_list, entry) { in n2_chunk_complete() 995 list_for_each_entry_safe(c, tmp, &rctx->chunk_list, entry) { in n2_do_ecb() 1052 list_for_each_entry_safe(c, tmp, &rctx->chunk_list, in n2_do_chaining() 1065 list_for_each_entry_safe_reverse(c, tmp, &rctx->chunk_list, in n2_do_chaining()
|