Home
last modified time | relevance | path

Searched refs:buffer_list (Results 1 – 24 of 24) sorted by relevance

/drivers/infiniband/hw/cxgb3/
Diwch_mem.c134 int build_phys_page_list(struct ib_phys_buf *buffer_list, in build_phys_page_list() argument
148 if (i != 0 && buffer_list[i].addr & ~PAGE_MASK) in build_phys_page_list()
151 (buffer_list[i].size & ~PAGE_MASK)) in build_phys_page_list()
153 *total_size += buffer_list[i].size; in build_phys_page_list()
155 mask |= buffer_list[i].addr; in build_phys_page_list()
157 mask |= buffer_list[i].addr & PAGE_MASK; in build_phys_page_list()
159 mask |= buffer_list[i].addr + buffer_list[i].size; in build_phys_page_list()
161 mask |= (buffer_list[i].addr + buffer_list[i].size + in build_phys_page_list()
173 buffer_list[0].size += buffer_list[0].addr & ((1ULL << *shift) - 1); in build_phys_page_list()
174 buffer_list[0].addr &= ~0ull << *shift; in build_phys_page_list()
[all …]
Diwch_provider.c483 struct ib_phys_buf *buffer_list, in iwch_register_phys_mem() argument
508 if ((*iova_start & ~PAGE_MASK) != (buffer_list[0].addr & ~PAGE_MASK)) { in iwch_register_phys_mem()
514 ((buffer_list[0].addr + buffer_list[0].size) & ~PAGE_MASK)) { in iwch_register_phys_mem()
519 ret = build_phys_page_list(buffer_list, num_phys_buf, iova_start, in iwch_register_phys_mem()
562 struct ib_phys_buf *buffer_list, in iwch_reregister_phys_mem() argument
597 ret = build_phys_page_list(buffer_list, num_phys_buf, in iwch_reregister_phys_mem()
Diwch_provider.h351 int build_phys_page_list(struct ib_phys_buf *buffer_list,
/drivers/infiniband/hw/cxgb4/
Dmem.c440 static int build_phys_page_list(struct ib_phys_buf *buffer_list, in build_phys_page_list() argument
451 if (i != 0 && buffer_list[i].addr & ~PAGE_MASK) in build_phys_page_list()
454 (buffer_list[i].size & ~PAGE_MASK)) in build_phys_page_list()
456 *total_size += buffer_list[i].size; in build_phys_page_list()
458 mask |= buffer_list[i].addr; in build_phys_page_list()
460 mask |= buffer_list[i].addr & PAGE_MASK; in build_phys_page_list()
462 mask |= buffer_list[i].addr + buffer_list[i].size; in build_phys_page_list()
464 mask |= (buffer_list[i].addr + buffer_list[i].size + in build_phys_page_list()
476 buffer_list[0].size += buffer_list[0].addr & ((1ULL << *shift) - 1); in build_phys_page_list()
477 buffer_list[0].addr &= ~0ull << *shift; in build_phys_page_list()
[all …]
Diw_cxgb4.h972 struct ib_phys_buf *buffer_list,
979 struct ib_phys_buf *buffer_list,
/drivers/iio/
Dindustrialio-buffer.c37 return !list_empty(&buf->buffer_list); in iio_buffer_is_active()
204 INIT_LIST_HEAD(&buffer->buffer_list); in iio_buffer_init()
566 list_add(&buffer->buffer_list, &indio_dev->buffer_list); in iio_buffer_activate()
571 list_del_init(&buffer->buffer_list); in iio_buffer_deactivate()
581 &indio_dev->buffer_list, buffer_list) in iio_buffer_deactivate_all()
651 list_is_singular(&indio_dev->buffer_list)) in iio_verify_update()
656 list_for_each_entry(buffer, &indio_dev->buffer_list, buffer_list) { in iio_verify_update()
673 if (insert_buffer && !list_empty(&indio_dev->buffer_list)) in iio_verify_update()
694 list_for_each_entry(buffer, &indio_dev->buffer_list, buffer_list) { in iio_verify_update()
790 if (list_empty(&indio_dev->buffer_list)) in iio_disable_buffers()
[all …]
Dindustrialio-core.c1022 INIT_LIST_HEAD(&dev->buffer_list); in iio_device_alloc()
/drivers/infiniband/hw/mthca/
Dmthca_mr.c255 int start_index, u64 *buffer_list, int list_len) in __mthca_write_mtt() argument
273 mtt_entry[i + 2] = cpu_to_be64(buffer_list[i] | in __mthca_write_mtt()
291 buffer_list += i; in __mthca_write_mtt()
317 u64 *buffer_list, int list_len) in mthca_tavor_write_mtt_seg() argument
325 mthca_write64_raw(cpu_to_be64(buffer_list[i] | MTHCA_MTT_FLAG_PRESENT), in mthca_tavor_write_mtt_seg()
331 u64 *buffer_list, int list_len) in mthca_arbel_write_mtt_seg() argument
352 mtts[i] = cpu_to_be64(buffer_list[i] | MTHCA_MTT_FLAG_PRESENT); in mthca_arbel_write_mtt_seg()
359 int start_index, u64 *buffer_list, int list_len) in mthca_write_mtt() argument
366 return __mthca_write_mtt(dev, mtt, start_index, buffer_list, list_len); in mthca_write_mtt()
372 buffer_list, chunk); in mthca_write_mtt()
[all …]
Dmthca_provider.c902 struct ib_phys_buf *buffer_list, in mthca_reg_phys_mr() argument
916 mask = buffer_list[0].addr ^ *iova_start; in mthca_reg_phys_mr()
920 mask |= buffer_list[i].addr; in mthca_reg_phys_mr()
922 mask |= buffer_list[i].addr + buffer_list[i].size; in mthca_reg_phys_mr()
924 total_size += buffer_list[i].size; in mthca_reg_phys_mr()
932 buffer_list[0].size += buffer_list[0].addr & ((1ULL << shift) - 1); in mthca_reg_phys_mr()
933 buffer_list[0].addr &= ~0ull << shift; in mthca_reg_phys_mr()
941 npages += (buffer_list[i].size + (1ULL << shift) - 1) >> shift; in mthca_reg_phys_mr()
955 j < (buffer_list[i].size + (1ULL << shift) - 1) >> shift; in mthca_reg_phys_mr()
957 page_list[n++] = buffer_list[i].addr + ((u64) j << shift); in mthca_reg_phys_mr()
[all …]
Dmthca_dev.h469 int start_index, u64 *buffer_list, int list_len);
475 u64 *buffer_list, int buffer_size_shift,
/drivers/staging/rdma/amso1100/
Dc2_provider.c341 struct ib_phys_buf *buffer_list, in c2_reg_phys_mr() argument
362 if (offset_in_page(buffer_list[i].addr)) { in c2_reg_phys_mr()
364 (unsigned int) buffer_list[i].addr); in c2_reg_phys_mr()
368 if (!buffer_list[i].size) { in c2_reg_phys_mr()
373 total_len += buffer_list[i].size; in c2_reg_phys_mr()
374 pbl_depth += ALIGN(buffer_list[i].size, in c2_reg_phys_mr()
389 naddrs = ALIGN(buffer_list[i].size, in c2_reg_phys_mr()
392 page_list[j++] = (buffer_list[i].addr + in c2_reg_phys_mr()
/drivers/staging/rdma/ipath/
Dipath_mr.c133 struct ib_phys_buf *buffer_list, in ipath_reg_phys_mr() argument
158 mr->mr.map[m]->segs[n].vaddr = (void *) buffer_list[i].addr; in ipath_reg_phys_mr()
159 mr->mr.map[m]->segs[n].length = buffer_list[i].size; in ipath_reg_phys_mr()
160 mr->mr.length += buffer_list[i].size; in ipath_reg_phys_mr()
Dipath_verbs.h832 struct ib_phys_buf *buffer_list,
/drivers/staging/rdma/hfi1/
Dmr.c200 struct ib_phys_buf *buffer_list, in hfi1_reg_phys_mr() argument
220 mr->mr.map[m]->segs[n].vaddr = (void *) buffer_list[i].addr; in hfi1_reg_phys_mr()
221 mr->mr.map[m]->segs[n].length = buffer_list[i].size; in hfi1_reg_phys_mr()
222 mr->mr.length += buffer_list[i].size; in hfi1_reg_phys_mr()
Dverbs.h1016 struct ib_phys_buf *buffer_list,
/drivers/infiniband/hw/qib/
Dqib_mr.c183 struct ib_phys_buf *buffer_list, in qib_reg_phys_mr() argument
203 mr->mr.map[m]->segs[n].vaddr = (void *) buffer_list[i].addr; in qib_reg_phys_mr()
204 mr->mr.map[m]->segs[n].length = buffer_list[i].size; in qib_reg_phys_mr()
205 mr->mr.length += buffer_list[i].size; in qib_reg_phys_mr()
Dqib_verbs.h1036 struct ib_phys_buf *buffer_list,
/drivers/media/usb/hackrf/
Dhackrf.c447 struct list_head *buffer_list) in hackrf_get_next_buffer() argument
453 if (list_empty(buffer_list)) in hackrf_get_next_buffer()
456 buffer = list_entry(buffer_list->next, struct hackrf_buffer, list); in hackrf_get_next_buffer()
742 struct list_head *buffer_list; in hackrf_return_all_buffers() local
748 buffer_list = &dev->rx_buffer_list; in hackrf_return_all_buffers()
750 buffer_list = &dev->tx_buffer_list; in hackrf_return_all_buffers()
753 list_for_each_entry_safe(buffer, node, buffer_list, list) { in hackrf_return_all_buffers()
785 struct list_head *buffer_list; in hackrf_buf_queue() local
791 buffer_list = &dev->rx_buffer_list; in hackrf_buf_queue()
793 buffer_list = &dev->tx_buffer_list; in hackrf_buf_queue()
[all …]
/drivers/media/pci/sta2x11/
Dsta2x11_vip.c146 struct list_head buffer_list; member
250 if (list_empty(&vip->buffer_list)) {/* No available buffer */ in vip_active_buf_next()
254 vip->active = list_first_entry(&vip->buffer_list, in vip_active_buf_next()
323 list_add_tail(&vip_buf->list, &vip->buffer_list); in buffer_queue()
325 vip->active = list_first_entry(&vip->buffer_list, in buffer_queue()
376 list_for_each_entry_safe(vip_buf, node, &vip->buffer_list, list) { in stop_streaming()
875 INIT_LIST_HEAD(&vip->buffer_list); in sta2x11_vip_init_buffer()
/drivers/infiniband/hw/ocrdma/
Docrdma_verbs.h121 struct ib_phys_buf *buffer_list,
/drivers/infiniband/hw/nes/
Dnes_verbs.c2078 struct ib_phys_buf *buffer_list, int num_phys_buf, int acc, in nes_reg_phys_mr() argument
2120 if ((buffer_list[0].addr ^ *iova_start) & ~PAGE_MASK) in nes_reg_phys_mr()
2189 mask = !buffer_list[i].size; in nes_reg_phys_mr()
2191 mask |= buffer_list[i].addr; in nes_reg_phys_mr()
2193 mask |= buffer_list[i].addr + buffer_list[i].size; in nes_reg_phys_mr()
2203 region_length += buffer_list[i].size; in nes_reg_phys_mr()
2205 if ((buffer_list[i-1].addr+PAGE_SIZE) != buffer_list[i].addr) in nes_reg_phys_mr()
2208 vpbl.pbl_vbase[cur_pbl_index].pa_low = cpu_to_le32((u32)buffer_list[i].addr & PAGE_MASK); in nes_reg_phys_mr()
2210 cpu_to_le32((u32)((((u64)buffer_list[i].addr) >> 32))); in nes_reg_phys_mr()
2232 buffer_list[0].addr, pbl_count, (u16)cur_pbl_index, acc, iova_start, in nes_reg_phys_mr()
/drivers/tty/
Dsynclink.c220 char *buffer_list; /* virtual address of Rx & Tx buffer lists */ member
3687 info->buffer_list = info->memory_base + info->last_mem_alloc; in mgsl_alloc_buffer_list_memory()
3697 …info->buffer_list = dma_alloc_coherent(NULL, BUFFERLISTSIZE, &info->buffer_list_dma_addr, GFP_KERN… in mgsl_alloc_buffer_list_memory()
3698 if (info->buffer_list == NULL) in mgsl_alloc_buffer_list_memory()
3705 memset( info->buffer_list, 0, BUFFERLISTSIZE ); in mgsl_alloc_buffer_list_memory()
3710 info->rx_buffer_list = (DMABUFFERENTRY *)info->buffer_list; in mgsl_alloc_buffer_list_memory()
3711 info->tx_buffer_list = (DMABUFFERENTRY *)info->buffer_list; in mgsl_alloc_buffer_list_memory()
3767 if (info->buffer_list && info->bus_type != MGSL_BUS_TYPE_PCI) in mgsl_free_buffer_list_memory()
3768 dma_free_coherent(NULL, BUFFERLISTSIZE, info->buffer_list, info->buffer_list_dma_addr); in mgsl_free_buffer_list_memory()
3770 info->buffer_list = NULL; in mgsl_free_buffer_list_memory()
[all …]
Dsynclinkmp.c194 char *buffer_list; /* virtual address of Rx & Tx buffer lists */ member
3447 info->buffer_list = info->memory_base + info->port_array[0]->last_mem_alloc; in alloc_buf_list()
3451 memset(info->buffer_list, 0, BUFFERLISTSIZE); in alloc_buf_list()
3456 info->rx_buf_list = (SCADESC *)info->buffer_list; in alloc_buf_list()
3458 info->tx_buf_list = (SCADESC *)info->buffer_list; in alloc_buf_list()
3520 info->buffer_list = NULL; in free_dma_bufs()
/drivers/firewire/
Dohci.c143 struct list_head buffer_list; member
1075 desc = list_entry(ctx->buffer_list.next, in context_tasklet()
1103 list_move_tail(&old_desc->list, &ctx->buffer_list); in context_tasklet()
1143 list_add_tail(&desc->list, &ctx->buffer_list); in context_add_buffer()
1156 INIT_LIST_HEAD(&ctx->buffer_list); in context_init()
1160 ctx->buffer_tail = list_entry(ctx->buffer_list.next, in context_init()
1187 list_for_each_entry_safe(desc, tmp, &ctx->buffer_list, list) in context_release()
1207 if (desc->list.next == &ctx->buffer_list) { in context_get_descriptors()