/drivers/s390/char/ |
D | sclp_rw.c | 43 struct sclp_buffer *buffer; in sclp_make_buffer() local 51 buffer = ((struct sclp_buffer *) ((addr_t) sccb + PAGE_SIZE)) - 1; in sclp_make_buffer() 52 buffer->sccb = sccb; in sclp_make_buffer() 53 buffer->retry_count = 0; in sclp_make_buffer() 54 buffer->messages = 0; in sclp_make_buffer() 55 buffer->char_sum = 0; in sclp_make_buffer() 56 buffer->current_line = NULL; in sclp_make_buffer() 57 buffer->current_length = 0; in sclp_make_buffer() 58 buffer->columns = columns; in sclp_make_buffer() 59 buffer->htab = htab; in sclp_make_buffer() [all …]
|
/drivers/gpu/drm/etnaviv/ |
D | etnaviv_buffer.c | 25 static inline void OUT(struct etnaviv_cmdbuf *buffer, u32 data) in OUT() argument 27 u32 *vaddr = (u32 *)buffer->vaddr; in OUT() 29 BUG_ON(buffer->user_size >= buffer->size); in OUT() 31 vaddr[buffer->user_size / 4] = data; in OUT() 32 buffer->user_size += 4; in OUT() 35 static inline void CMD_LOAD_STATE(struct etnaviv_cmdbuf *buffer, in CMD_LOAD_STATE() argument 40 buffer->user_size = ALIGN(buffer->user_size, 8); in CMD_LOAD_STATE() 43 OUT(buffer, VIV_FE_LOAD_STATE_HEADER_OP_LOAD_STATE | in CMD_LOAD_STATE() 46 OUT(buffer, value); in CMD_LOAD_STATE() 49 static inline void CMD_END(struct etnaviv_cmdbuf *buffer) in CMD_END() argument [all …]
|
/drivers/media/tuners/ |
D | tea5767.c | 133 unsigned char *buffer) in tea5767_status_dump() argument 137 if (TEA5767_READY_FLAG_MASK & buffer[0]) in tea5767_status_dump() 142 if (TEA5767_BAND_LIMIT_MASK & buffer[0]) in tea5767_status_dump() 147 div = ((buffer[0] & 0x3f) << 8) | buffer[1]; in tea5767_status_dump() 164 buffer[0] = (div >> 8) & 0x3f; in tea5767_status_dump() 165 buffer[1] = div & 0xff; in tea5767_status_dump() 170 if (TEA5767_STEREO_MASK & buffer[2]) in tea5767_status_dump() 175 tuner_info("IF Counter = %d\n", buffer[2] & TEA5767_IF_CNTR_MASK); in tea5767_status_dump() 178 (buffer[3] & TEA5767_ADC_LEVEL_MASK) >> 4); in tea5767_status_dump() 180 tuner_info("Chip ID = %d\n", (buffer[3] & TEA5767_CHIP_ID_MASK)); in tea5767_status_dump() [all …]
|
D | tea5761.c | 123 static void tea5761_status_dump(unsigned char *buffer) in tea5761_status_dump() argument 127 div = ((buffer[2] & 0x3f) << 8) | buffer[3]; in tea5761_status_dump() 142 unsigned char buffer[7] = {0, 0, 0, 0, 0, 0, 0 }; in __set_radio_freq() local 150 buffer[5] |= TEA5761_TNCTRL_MU; in __set_radio_freq() 152 buffer[4] |= TEA5761_TNCTRL_PUPD_0; in __set_radio_freq() 158 buffer[5] |= TEA5761_TNCTRL_MST; in __set_radio_freq() 164 buffer[1] = (div >> 8) & 0x3f; in __set_radio_freq() 165 buffer[2] = div & 0xff; in __set_radio_freq() 168 tea5761_status_dump(buffer); in __set_radio_freq() 170 if (7 != (rc = tuner_i2c_xfer_send(&priv->i2c_props, buffer, 7))) in __set_radio_freq() [all …]
|
/drivers/dma-buf/heaps/ |
D | cma_heap.c | 53 struct cma_heap_buffer *buffer = dmabuf->priv; in cma_heap_attach() local 61 ret = sg_alloc_table_from_pages(&a->table, buffer->pages, in cma_heap_attach() 62 buffer->pagecount, 0, in cma_heap_attach() 63 buffer->pagecount << PAGE_SHIFT, in cma_heap_attach() 76 mutex_lock(&buffer->lock); in cma_heap_attach() 77 list_add(&a->list, &buffer->attachments); in cma_heap_attach() 78 mutex_unlock(&buffer->lock); in cma_heap_attach() 86 struct cma_heap_buffer *buffer = dmabuf->priv; in cma_heap_detach() local 89 mutex_lock(&buffer->lock); in cma_heap_detach() 91 mutex_unlock(&buffer->lock); in cma_heap_detach() [all …]
|
D | system_heap.c | 90 struct system_heap_buffer *buffer = dmabuf->priv; in system_heap_attach() local 98 table = dup_sg_table(&buffer->sg_table); in system_heap_attach() 108 a->uncached = buffer->uncached; in system_heap_attach() 111 mutex_lock(&buffer->lock); in system_heap_attach() 112 list_add(&a->list, &buffer->attachments); in system_heap_attach() 113 mutex_unlock(&buffer->lock); in system_heap_attach() 121 struct system_heap_buffer *buffer = dmabuf->priv; in system_heap_detach() local 124 mutex_lock(&buffer->lock); in system_heap_detach() 126 mutex_unlock(&buffer->lock); in system_heap_detach() 168 struct system_heap_buffer *buffer = dmabuf->priv; in system_heap_dma_buf_begin_cpu_access() local [all …]
|
/drivers/pnp/ |
D | interface.c | 26 char *buffer; /* pointer to begin of buffer */ member 37 static int pnp_printf(pnp_info_buffer_t * buffer, char *fmt, ...) in pnp_printf() argument 42 if (buffer->stop || buffer->error) in pnp_printf() 45 res = vsnprintf(buffer->curr, buffer->len - buffer->size, fmt, args); in pnp_printf() 47 if (buffer->size + res >= buffer->len) { in pnp_printf() 48 buffer->stop = 1; in pnp_printf() 51 buffer->curr += res; in pnp_printf() 52 buffer->size += res; in pnp_printf() 56 static void pnp_print_port(pnp_info_buffer_t * buffer, char *space, in pnp_print_port() argument 59 pnp_printf(buffer, "%sport %#llx-%#llx, align %#llx, size %#llx, " in pnp_print_port() [all …]
|
/drivers/android/ |
D | binder_alloc.c | 51 static struct binder_buffer *binder_buffer_next(struct binder_buffer *buffer) in binder_buffer_next() argument 53 return list_entry(buffer->entry.next, struct binder_buffer, entry); in binder_buffer_next() 56 static struct binder_buffer *binder_buffer_prev(struct binder_buffer *buffer) in binder_buffer_prev() argument 58 return list_entry(buffer->entry.prev, struct binder_buffer, entry); in binder_buffer_prev() 62 struct binder_buffer *buffer) in binder_alloc_buffer_size() argument 64 if (list_is_last(&buffer->entry, &alloc->buffers)) in binder_alloc_buffer_size() 65 return alloc->buffer + alloc->buffer_size - buffer->user_data; in binder_alloc_buffer_size() 66 return binder_buffer_next(buffer)->user_data - buffer->user_data; in binder_alloc_buffer_size() 74 struct binder_buffer *buffer; in binder_insert_free_buffer() local 88 buffer = rb_entry(parent, struct binder_buffer, rb_node); in binder_insert_free_buffer() [all …]
|
/drivers/iio/ |
D | industrialio-buffer.c | 110 struct iio_buffer *rb = ib->buffer; in iio_buffer_read() 179 struct iio_buffer *rb = ib->buffer; in iio_buffer_write() 242 struct iio_buffer *rb = ib->buffer; in iio_buffer_poll() 268 struct iio_buffer *rb = ib->buffer; in iio_buffer_read_wrapper() 281 struct iio_buffer *rb = ib->buffer; in iio_buffer_write_wrapper() 294 struct iio_buffer *rb = ib->buffer; in iio_buffer_poll_wrapper() 313 struct iio_buffer *buffer; in iio_buffer_wakeup_poll() local 317 buffer = iio_dev_opaque->attached_buffers[i]; in iio_buffer_wakeup_poll() 318 wake_up(&buffer->pollq); in iio_buffer_wakeup_poll() 322 int iio_pop_from_buffer(struct iio_buffer *buffer, void *data) in iio_pop_from_buffer() argument [all …]
|
/drivers/s390/net/ |
D | qeth_core_mpc.h | 18 #define QETH_IPA_PDU_LEN_TOTAL(buffer) (buffer + 0x0e) argument 19 #define QETH_IPA_PDU_LEN_PDU1(buffer) (buffer + 0x26) argument 20 #define QETH_IPA_PDU_LEN_PDU2(buffer) (buffer + 0x29) argument 21 #define QETH_IPA_PDU_LEN_PDU3(buffer) (buffer + 0x3a) argument 23 #define QETH_IPA_CMD_DEST_ADDR(buffer) (buffer + 0x2c) argument 523 char *buffer; member 877 #define QETH_CM_ENABLE_ISSUER_RM_TOKEN(buffer) (buffer + 0x2c) argument 878 #define QETH_CM_ENABLE_FILTER_TOKEN(buffer) (buffer + 0x53) argument 879 #define QETH_CM_ENABLE_USER_DATA(buffer) (buffer + 0x5b) argument 881 #define QETH_CM_ENABLE_RESP_FILTER_TOKEN(buffer) \ argument [all …]
|
/drivers/gpu/drm/ |
D | drm_client.c | 265 static void drm_client_buffer_delete(struct drm_client_buffer *buffer) in drm_client_buffer_delete() argument 267 if (buffer->gem) { in drm_client_buffer_delete() 268 drm_gem_vunmap_unlocked(buffer->gem, &buffer->map); in drm_client_buffer_delete() 269 drm_gem_object_put(buffer->gem); in drm_client_buffer_delete() 272 kfree(buffer); in drm_client_buffer_delete() 282 struct drm_client_buffer *buffer; in drm_client_buffer_create() local 286 buffer = kzalloc(sizeof(*buffer), GFP_KERNEL); in drm_client_buffer_create() 287 if (!buffer) in drm_client_buffer_create() 290 buffer->client = client; in drm_client_buffer_create() 305 buffer->pitch = dumb_args.pitch; in drm_client_buffer_create() [all …]
|
/drivers/scsi/ |
D | sr_vendor.c | 126 unsigned char *buffer; /* the buffer for the ioctl */ in sr_set_blocklength() local 134 buffer = kmalloc(512, GFP_KERNEL); in sr_set_blocklength() 135 if (!buffer) in sr_set_blocklength() 145 modesel = (struct ccs_modesel_head *) buffer; in sr_set_blocklength() 151 cgc.buffer = buffer; in sr_set_blocklength() 164 kfree(buffer); in sr_set_blocklength() 175 unsigned char *buffer; /* the buffer for the ioctl */ in sr_cd_check() local 182 buffer = kmalloc(512, GFP_KERNEL); in sr_cd_check() 183 if (!buffer) in sr_cd_check() 198 cgc.buffer = buffer; in sr_cd_check() [all …]
|
/drivers/gpu/drm/nouveau/nvkm/subdev/fault/ |
D | tu102.c | 35 struct nvkm_fault_buffer *buffer = container_of(inth, typeof(*buffer), inth); in tu102_fault_buffer_notify() local 37 nvkm_event_ntfy(&buffer->fault->event, buffer->id, NVKM_FAULT_BUFFER_EVENT_PENDING); in tu102_fault_buffer_notify() 42 tu102_fault_buffer_intr(struct nvkm_fault_buffer *buffer, bool enable) in tu102_fault_buffer_intr() argument 45 nvkm_inth_allow(&buffer->inth); in tu102_fault_buffer_intr() 47 nvkm_inth_block(&buffer->inth); in tu102_fault_buffer_intr() 51 tu102_fault_buffer_fini(struct nvkm_fault_buffer *buffer) in tu102_fault_buffer_fini() argument 53 struct nvkm_device *device = buffer->fault->subdev.device; in tu102_fault_buffer_fini() 54 const u32 foff = buffer->id * 0x20; in tu102_fault_buffer_fini() 60 tu102_fault_buffer_init(struct nvkm_fault_buffer *buffer) in tu102_fault_buffer_init() argument 62 struct nvkm_device *device = buffer->fault->subdev.device; in tu102_fault_buffer_init() [all …]
|
D | gv100.c | 34 struct nvkm_fault_buffer *buffer = fault->buffer[0]; in gv100_fault_buffer_process() local 36 struct nvkm_memory *mem = buffer->mem; in gv100_fault_buffer_process() 37 u32 get = nvkm_rd32(device, buffer->get); in gv100_fault_buffer_process() 38 u32 put = nvkm_rd32(device, buffer->put); in gv100_fault_buffer_process() 44 const u32 base = get * buffer->fault->func->buffer.entry_size; in gv100_fault_buffer_process() 55 if (++get == buffer->entries) in gv100_fault_buffer_process() 57 nvkm_wr32(device, buffer->get, get); in gv100_fault_buffer_process() 76 gv100_fault_buffer_intr(struct nvkm_fault_buffer *buffer, bool enable) in gv100_fault_buffer_intr() argument 78 struct nvkm_device *device = buffer->fault->subdev.device; in gv100_fault_buffer_intr() 79 const u32 intr = buffer->id ? 0x08000000 : 0x20000000; in gv100_fault_buffer_intr() [all …]
|
D | gp100.c | 30 gp100_fault_buffer_intr(struct nvkm_fault_buffer *buffer, bool enable) in gp100_fault_buffer_intr() argument 32 struct nvkm_device *device = buffer->fault->subdev.device; in gp100_fault_buffer_intr() 37 gp100_fault_buffer_fini(struct nvkm_fault_buffer *buffer) in gp100_fault_buffer_fini() argument 39 struct nvkm_device *device = buffer->fault->subdev.device; in gp100_fault_buffer_fini() 44 gp100_fault_buffer_init(struct nvkm_fault_buffer *buffer) in gp100_fault_buffer_init() argument 46 struct nvkm_device *device = buffer->fault->subdev.device; in gp100_fault_buffer_init() 47 nvkm_wr32(device, 0x002a74, upper_32_bits(buffer->addr)); in gp100_fault_buffer_init() 48 nvkm_wr32(device, 0x002a70, lower_32_bits(buffer->addr)); in gp100_fault_buffer_init() 52 u64 gp100_fault_buffer_pin(struct nvkm_fault_buffer *buffer) in gp100_fault_buffer_pin() argument 54 return nvkm_memory_bar2(buffer->mem); in gp100_fault_buffer_pin() [all …]
|
D | user.c | 34 struct nvkm_fault_buffer *buffer = nvkm_fault_buffer(object); in nvkm_ufault_uevent() local 42 return nvkm_uevent_add(uevent, &buffer->fault->event, buffer->id, in nvkm_ufault_uevent() 50 struct nvkm_fault_buffer *buffer = nvkm_fault_buffer(object); in nvkm_ufault_map() local 51 struct nvkm_device *device = buffer->fault->subdev.device; in nvkm_ufault_map() 53 *addr = device->func->resource_addr(device, 3) + buffer->addr; in nvkm_ufault_map() 54 *size = nvkm_memory_size(buffer->mem); in nvkm_ufault_map() 61 struct nvkm_fault_buffer *buffer = nvkm_fault_buffer(object); in nvkm_ufault_fini() local 62 buffer->fault->func->buffer.fini(buffer); in nvkm_ufault_fini() 69 struct nvkm_fault_buffer *buffer = nvkm_fault_buffer(object); in nvkm_ufault_init() local 70 buffer->fault->func->buffer.init(buffer); in nvkm_ufault_init() [all …]
|
D | base.c | 30 fault->func->buffer.intr(fault->buffer[index], false); in nvkm_fault_ntfy_fini() 37 fault->func->buffer.intr(fault->buffer[index], true); in nvkm_fault_ntfy_init() 76 struct nvkm_fault_buffer *buffer; in nvkm_fault_oneinit_buffer() local 79 if (!(buffer = kzalloc(sizeof(*buffer), GFP_KERNEL))) in nvkm_fault_oneinit_buffer() 81 buffer->fault = fault; in nvkm_fault_oneinit_buffer() 82 buffer->id = id; in nvkm_fault_oneinit_buffer() 83 fault->func->buffer.info(buffer); in nvkm_fault_oneinit_buffer() 84 fault->buffer[id] = buffer; in nvkm_fault_oneinit_buffer() 86 nvkm_debug(subdev, "buffer %d: %d entries\n", id, buffer->entries); in nvkm_fault_oneinit_buffer() 88 ret = nvkm_memory_new(device, NVKM_MEM_TARGET_INST, buffer->entries * in nvkm_fault_oneinit_buffer() [all …]
|
/drivers/usb/misc/ |
D | cytherm.c | 81 unsigned char *buffer; in brightness_store() local 84 buffer = kmalloc(8, GFP_KERNEL); in brightness_store() 85 if (!buffer) in brightness_store() 97 cytherm->brightness, buffer, 8); in brightness_store() 102 0x01, buffer, 8); in brightness_store() 106 kfree(buffer); in brightness_store() 123 unsigned char *buffer; in temp_show() local 127 buffer = kmalloc(8, GFP_KERNEL); in temp_show() 128 if (!buffer) in temp_show() 132 retval = vendor_command(cytherm->udev, READ_RAM, TEMP, 0, buffer, 8); in temp_show() [all …]
|
/drivers/net/ethernet/sfc/siena/ |
D | tx_common.c | 39 tx_queue->buffer = kcalloc(entries, sizeof(*tx_queue->buffer), in efx_siena_probe_tx_queue() 41 if (!tx_queue->buffer) in efx_siena_probe_tx_queue() 63 kfree(tx_queue->buffer); in efx_siena_probe_tx_queue() 64 tx_queue->buffer = NULL; in efx_siena_probe_tx_queue() 102 if (!tx_queue->buffer) in efx_siena_remove_tx_queue() 117 kfree(tx_queue->buffer); in efx_siena_remove_tx_queue() 118 tx_queue->buffer = NULL; in efx_siena_remove_tx_queue() 123 struct efx_tx_buffer *buffer, in efx_dequeue_buffer() argument 127 if (buffer->unmap_len) { in efx_dequeue_buffer() 129 dma_addr_t unmap_addr = buffer->dma_addr - buffer->dma_offset; in efx_dequeue_buffer() [all …]
|
/drivers/acpi/acpica/ |
D | exstorob.c | 34 u8 *buffer; in acpi_ex_store_buffer_to_buffer() local 46 buffer = ACPI_CAST_PTR(u8, source_desc->buffer.pointer); in acpi_ex_store_buffer_to_buffer() 47 length = source_desc->buffer.length; in acpi_ex_store_buffer_to_buffer() 53 if ((target_desc->buffer.length == 0) || in acpi_ex_store_buffer_to_buffer() 55 target_desc->buffer.pointer = ACPI_ALLOCATE(length); in acpi_ex_store_buffer_to_buffer() 56 if (!target_desc->buffer.pointer) { in acpi_ex_store_buffer_to_buffer() 60 target_desc->buffer.length = length; in acpi_ex_store_buffer_to_buffer() 65 if (length <= target_desc->buffer.length) { in acpi_ex_store_buffer_to_buffer() 69 memset(target_desc->buffer.pointer, 0, in acpi_ex_store_buffer_to_buffer() 70 target_desc->buffer.length); in acpi_ex_store_buffer_to_buffer() [all …]
|
/drivers/net/ethernet/sfc/ |
D | tx_common.c | 39 tx_queue->buffer = kcalloc(entries, sizeof(*tx_queue->buffer), in efx_probe_tx_queue() 41 if (!tx_queue->buffer) in efx_probe_tx_queue() 63 kfree(tx_queue->buffer); in efx_probe_tx_queue() 64 tx_queue->buffer = NULL; in efx_probe_tx_queue() 100 struct efx_tx_buffer *buffer; in efx_fini_tx_queue() local 107 if (!tx_queue->buffer) in efx_fini_tx_queue() 115 buffer = &tx_queue->buffer[tx_queue->read_count & tx_queue->ptr_mask]; in efx_fini_tx_queue() 116 efx_dequeue_buffer(tx_queue, buffer, &pkts_compl, &bytes_compl, in efx_fini_tx_queue() 129 if (!tx_queue->buffer) in efx_remove_tx_queue() 144 kfree(tx_queue->buffer); in efx_remove_tx_queue() [all …]
|
/drivers/gpu/drm/nouveau/ |
D | nouveau_svm.c | 70 } buffer[1]; member 457 struct nouveau_svm_fault_buffer *buffer, u32 offset) in nouveau_svm_fault_cache() argument 459 struct nvif_object *memory = &buffer->object; in nouveau_svm_fault_cache() 480 if (!buffer->fault[buffer->fault_nr]) { in nouveau_svm_fault_cache() 486 buffer->fault[buffer->fault_nr] = fault; in nouveau_svm_fault_cache() 489 fault = buffer->fault[buffer->fault_nr++]; in nouveau_svm_fault_cache() 718 struct nouveau_svm_fault_buffer *buffer = container_of(work, typeof(*buffer), work); in nouveau_svm_fault() local 719 struct nouveau_svm *svm = container_of(buffer, typeof(*svm), buffer[buffer->id]); in nouveau_svm_fault() 735 if (buffer->get == buffer->put) { in nouveau_svm_fault() 736 buffer->put = nvif_rd32(device, buffer->putaddr); in nouveau_svm_fault() [all …]
|
/drivers/net/ethernet/sfc/falcon/ |
D | tx.c | 26 struct ef4_tx_buffer *buffer) in ef4_tx_get_copy_buffer() argument 38 buffer->dma_addr = page_buf->dma_addr + offset; in ef4_tx_get_copy_buffer() 39 buffer->unmap_len = 0; in ef4_tx_get_copy_buffer() 44 struct ef4_tx_buffer *buffer, size_t len) in ef4_tx_get_copy_buffer_limited() argument 48 return ef4_tx_get_copy_buffer(tx_queue, buffer); in ef4_tx_get_copy_buffer_limited() 52 struct ef4_tx_buffer *buffer, in ef4_dequeue_buffer() argument 56 if (buffer->unmap_len) { in ef4_dequeue_buffer() 58 dma_addr_t unmap_addr = buffer->dma_addr - buffer->dma_offset; in ef4_dequeue_buffer() 59 if (buffer->flags & EF4_TX_BUF_MAP_SINGLE) in ef4_dequeue_buffer() 60 dma_unmap_single(dma_dev, unmap_addr, buffer->unmap_len, in ef4_dequeue_buffer() [all …]
|
/drivers/firewire/ |
D | nosy.c | 50 } buffer[13]; member 107 struct packet_buffer buffer; member 115 packet_buffer_init(struct packet_buffer *buffer, size_t capacity) in packet_buffer_init() argument 117 buffer->data = kmalloc(capacity, GFP_KERNEL); in packet_buffer_init() 118 if (buffer->data == NULL) in packet_buffer_init() 120 buffer->head = (struct packet *) buffer->data; in packet_buffer_init() 121 buffer->tail = (struct packet *) buffer->data; in packet_buffer_init() 122 buffer->capacity = capacity; in packet_buffer_init() 123 buffer->lost_packet_count = 0; in packet_buffer_init() 124 atomic_set(&buffer->size, 0); in packet_buffer_init() [all …]
|
/drivers/net/ethernet/mellanox/mlx5/core/en/ |
D | port_buffer.c | 41 void *buffer; in mlx5e_port_query_buffer() local 55 buffer = MLX5_ADDR_OF(pbmc_reg, out, buffer[i]); in mlx5e_port_query_buffer() 56 port_buffer->buffer[i].lossy = in mlx5e_port_query_buffer() 57 MLX5_GET(bufferx_reg, buffer, lossy); in mlx5e_port_query_buffer() 58 port_buffer->buffer[i].epsb = in mlx5e_port_query_buffer() 59 MLX5_GET(bufferx_reg, buffer, epsb); in mlx5e_port_query_buffer() 60 port_buffer->buffer[i].size = in mlx5e_port_query_buffer() 61 MLX5_GET(bufferx_reg, buffer, size) * port_buff_cell_sz; in mlx5e_port_query_buffer() 62 port_buffer->buffer[i].xon = in mlx5e_port_query_buffer() 63 MLX5_GET(bufferx_reg, buffer, xon_threshold) * port_buff_cell_sz; in mlx5e_port_query_buffer() [all …]
|