/drivers/staging/media/atomisp/pci/base/circbuf/src/ |
D | circbuf.c | 36 ia_css_circbuf_read(ia_css_circbuf_t *cb); 48 static inline void ia_css_circbuf_shift_chunk(ia_css_circbuf_t *cb, 72 ia_css_circbuf_create(ia_css_circbuf_t *cb, in ia_css_circbuf_create() argument 80 cb->desc = desc; in ia_css_circbuf_create() 82 cb->desc->start = 0; in ia_css_circbuf_create() 83 cb->desc->end = 0; in ia_css_circbuf_create() 84 cb->desc->step = 0; in ia_css_circbuf_create() 86 for (i = 0; i < cb->desc->size; i++) in ia_css_circbuf_create() 89 cb->elems = elems; in ia_css_circbuf_create() 96 void ia_css_circbuf_destroy(ia_css_circbuf_t *cb) in ia_css_circbuf_destroy() argument [all …]
|
/drivers/staging/media/atomisp/pci/base/circbuf/interface/ |
D | ia_css_circbuf.h | 49 ia_css_circbuf_t *cb, 59 ia_css_circbuf_t *cb); 72 ia_css_circbuf_t *cb); 86 ia_css_circbuf_t *cb, 147 ia_css_circbuf_t *cb, in ia_css_circbuf_get_pos_at_offset() argument 153 OP___assert(cb); in ia_css_circbuf_get_pos_at_offset() 154 OP___assert(cb->desc); in ia_css_circbuf_get_pos_at_offset() 155 OP___assert(cb->desc->size > 0); in ia_css_circbuf_get_pos_at_offset() 159 offset += cb->desc->size; in ia_css_circbuf_get_pos_at_offset() 163 dest = OP_std_modadd(base, offset, cb->desc->size); in ia_css_circbuf_get_pos_at_offset() [all …]
|
/drivers/misc/habanalabs/common/ |
D | command_buffer.c | 16 static int cb_map_mem(struct hl_ctx *ctx, struct hl_cb *cb) in cb_map_mem() argument 39 INIT_LIST_HEAD(&cb->va_block_list); in cb_map_mem() 41 for (bus_addr = cb->bus_address; in cb_map_mem() 42 bus_addr < cb->bus_address + cb->size; in cb_map_mem() 63 list_add_tail(&va_block->node, &cb->va_block_list); in cb_map_mem() 68 bus_addr = cb->bus_address; in cb_map_mem() 70 list_for_each_entry(va_block, &cb->va_block_list, node) { in cb_map_mem() 73 &cb->va_block_list)); in cb_map_mem() 88 cb->is_mmu_mapped = true; in cb_map_mem() 93 list_for_each_entry(va_block, &cb->va_block_list, node) { in cb_map_mem() [all …]
|
/drivers/scsi/ |
D | myrb.c | 90 static bool myrb_create_mempools(struct pci_dev *pdev, struct myrb_hba *cb) in myrb_create_mempools() argument 95 elem_size = cb->host->sg_tablesize * elem_align; in myrb_create_mempools() 96 cb->sg_pool = dma_pool_create("myrb_sg", &pdev->dev, in myrb_create_mempools() 98 if (cb->sg_pool == NULL) { in myrb_create_mempools() 99 shost_printk(KERN_ERR, cb->host, in myrb_create_mempools() 104 cb->dcdb_pool = dma_pool_create("myrb_dcdb", &pdev->dev, in myrb_create_mempools() 107 if (!cb->dcdb_pool) { in myrb_create_mempools() 108 dma_pool_destroy(cb->sg_pool); in myrb_create_mempools() 109 cb->sg_pool = NULL; in myrb_create_mempools() 110 shost_printk(KERN_ERR, cb->host, in myrb_create_mempools() [all …]
|
/drivers/mfd/ |
D | lm3533-ctrlbank.c | 28 static inline u8 lm3533_ctrlbank_get_reg(struct lm3533_ctrlbank *cb, u8 base) in lm3533_ctrlbank_get_reg() argument 30 return base + cb->id; in lm3533_ctrlbank_get_reg() 33 int lm3533_ctrlbank_enable(struct lm3533_ctrlbank *cb) in lm3533_ctrlbank_enable() argument 38 dev_dbg(cb->dev, "%s - %d\n", __func__, cb->id); in lm3533_ctrlbank_enable() 40 mask = 1 << cb->id; in lm3533_ctrlbank_enable() 41 ret = lm3533_update(cb->lm3533, LM3533_REG_CTRLBANK_ENABLE, in lm3533_ctrlbank_enable() 44 dev_err(cb->dev, "failed to enable ctrlbank %d\n", cb->id); in lm3533_ctrlbank_enable() 50 int lm3533_ctrlbank_disable(struct lm3533_ctrlbank *cb) in lm3533_ctrlbank_disable() argument 55 dev_dbg(cb->dev, "%s - %d\n", __func__, cb->id); in lm3533_ctrlbank_disable() 57 mask = 1 << cb->id; in lm3533_ctrlbank_disable() [all …]
|
/drivers/irqchip/ |
D | irq-crossbar.c | 43 static struct crossbar_device *cb; variable 47 writel(cb_no, cb->crossbar_base + cb->register_offsets[irq_no]); in crossbar_writel() 52 writew(cb_no, cb->crossbar_base + cb->register_offsets[irq_no]); in crossbar_writew() 57 writeb(cb_no, cb->crossbar_base + cb->register_offsets[irq_no]); in crossbar_writeb() 84 raw_spin_lock(&cb->lock); in allocate_gic_irq() 85 for (i = cb->int_max - 1; i >= 0; i--) { in allocate_gic_irq() 86 if (cb->irq_map[i] == IRQ_FREE) { in allocate_gic_irq() 87 cb->irq_map[i] = hwirq; in allocate_gic_irq() 91 raw_spin_unlock(&cb->lock); in allocate_gic_irq() 104 cb->irq_map[i] = IRQ_FREE; in allocate_gic_irq() [all …]
|
/drivers/net/mdio/ |
D | mdio-mux.c | 40 struct mdio_mux_child_bus *cb = bus->priv; in mdio_mux_read() local 41 struct mdio_mux_parent_bus *pb = cb->parent; in mdio_mux_read() 45 r = pb->switch_fn(pb->current_child, cb->bus_number, pb->switch_data); in mdio_mux_read() 49 pb->current_child = cb->bus_number; in mdio_mux_read() 64 struct mdio_mux_child_bus *cb = bus->priv; in mdio_mux_write() local 65 struct mdio_mux_parent_bus *pb = cb->parent; in mdio_mux_write() 70 r = pb->switch_fn(pb->current_child, cb->bus_number, pb->switch_data); in mdio_mux_write() 74 pb->current_child = cb->bus_number; in mdio_mux_write() 87 struct mdio_mux_child_bus *cb = pb->children; in mdio_mux_uninit_children() local 89 while (cb) { in mdio_mux_uninit_children() [all …]
|
/drivers/misc/mei/ |
D | client.c | 317 void mei_io_cb_free(struct mei_cl_cb *cb) in mei_io_cb_free() argument 319 if (cb == NULL) in mei_io_cb_free() 322 list_del(&cb->list); in mei_io_cb_free() 323 kfree(cb->buf.data); in mei_io_cb_free() 324 kfree(cb); in mei_io_cb_free() 335 static inline void mei_tx_cb_enqueue(struct mei_cl_cb *cb, in mei_tx_cb_enqueue() argument 338 list_add_tail(&cb->list, head); in mei_tx_cb_enqueue() 339 cb->cl->tx_cb_queued++; in mei_tx_cb_enqueue() 349 static inline void mei_tx_cb_dequeue(struct mei_cl_cb *cb) in mei_tx_cb_dequeue() argument 351 if (!WARN_ON(cb->cl->tx_cb_queued == 0)) in mei_tx_cb_dequeue() [all …]
|
D | interrupt.c | 31 struct mei_cl_cb *cb, *next; in mei_irq_compl_handler() local 34 list_for_each_entry_safe(cb, next, cmpl_list, list) { in mei_irq_compl_handler() 35 cl = cb->cl; in mei_irq_compl_handler() 36 list_del_init(&cb->list); in mei_irq_compl_handler() 39 mei_cl_complete(cl, cb); in mei_irq_compl_handler() 99 struct mei_cl_cb *cb; in mei_cl_irq_read_msg() local 112 cb = list_first_entry_or_null(&cl->rd_pending, struct mei_cl_cb, list); in mei_cl_irq_read_msg() 113 if (!cb) { in mei_cl_irq_read_msg() 118 cb = mei_cl_alloc_cb(cl, mei_cl_mtu(cl), MEI_FOP_READ, cl->fp); in mei_cl_irq_read_msg() 119 if (!cb) in mei_cl_irq_read_msg() [all …]
|
/drivers/misc/sgi-gru/ |
D | gru_instructions.h | 22 extern int gru_check_status_proc(void *cb); 23 extern int gru_wait_proc(void *cb); 24 extern void gru_wait_abort_proc(void *cb); 79 unsigned long cb; member 359 static inline void gru_vload_phys(void *cb, unsigned long gpa, in gru_vload_phys() argument 362 struct gru_instruction *ins = (struct gru_instruction *)cb; in gru_vload_phys() 371 static inline void gru_vstore_phys(void *cb, unsigned long gpa, in gru_vstore_phys() argument 374 struct gru_instruction *ins = (struct gru_instruction *)cb; in gru_vstore_phys() 383 static inline void gru_vload(void *cb, unsigned long mem_addr, in gru_vload() argument 387 struct gru_instruction *ins = (struct gru_instruction *)cb; in gru_vload() [all …]
|
D | grukservices.c | 255 static int gru_get_cpu_resources(int dsr_bytes, void **cb, void **dsr) in gru_get_cpu_resources() argument 264 *cb = bs->kernel_cb + lcpu * GRU_HANDLE_STRIDE; in gru_get_cpu_resources() 272 static void gru_free_cpu_resources(void *cb, void *dsr) in gru_free_cpu_resources() argument 359 void gru_lock_async_resource(unsigned long han, void **cb, void **dsr) in gru_lock_async_resource() argument 367 if (cb) in gru_lock_async_resource() 368 *cb = bs->kernel_cb + ncpus * GRU_HANDLE_STRIDE; in gru_lock_async_resource() 387 int gru_get_cb_exception_detail(void *cb, in gru_get_cb_exception_detail() argument 406 off = cb - kgts->ts_gru->gs_gru_base_vaddr; in gru_get_cb_exception_detail() 412 cbrnum = thread_cbr_number(kgts, get_cb_number(cb)); in gru_get_cb_exception_detail() 413 cbe = get_cbe(GRUBASE(cb), cbrnum); in gru_get_cb_exception_detail() [all …]
|
/drivers/net/ethernet/netronome/nfp/ |
D | ccm_mbox.c | 64 struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb; in nfp_ccm_mbox_msg_init() local 66 cb->state = NFP_NET_MBOX_CMSG_STATE_QUEUED; in nfp_ccm_mbox_msg_init() 67 cb->err = 0; in nfp_ccm_mbox_msg_init() 68 cb->max_len = max_len; in nfp_ccm_mbox_msg_init() 69 cb->exp_reply = exp_reply; in nfp_ccm_mbox_msg_init() 70 cb->posted = false; in nfp_ccm_mbox_msg_init() 75 struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb; in nfp_ccm_mbox_maxlen() local 77 return cb->max_len; in nfp_ccm_mbox_maxlen() 82 struct nfp_ccm_mbox_cmsg_cb *cb = (void *)skb->cb; in nfp_ccm_mbox_done() local 84 return cb->state == NFP_NET_MBOX_CMSG_STATE_DONE; in nfp_ccm_mbox_done() [all …]
|
/drivers/dma-buf/ |
D | st-dma-fence.c | 37 struct dma_fence_cb cb; member 41 static void mock_wakeup(struct dma_fence *f, struct dma_fence_cb *cb) in mock_wakeup() argument 43 wake_up_process(container_of(cb, struct wait_cb, cb)->task); in mock_wakeup() 49 struct wait_cb cb = { .task = current }; in mock_wait() local 51 if (dma_fence_add_callback(f, &cb.cb, mock_wakeup)) in mock_wait() 67 if (!dma_fence_remove_callback(f, &cb.cb)) in mock_wait() 147 struct dma_fence_cb cb; member 151 static void simple_callback(struct dma_fence *f, struct dma_fence_cb *cb) in simple_callback() argument 153 smp_store_mb(container_of(cb, struct simple_cb, cb)->seen, true); in simple_callback() 158 struct simple_cb cb = {}; in test_add_callback() local [all …]
|
D | dma-fence.c | 612 int dma_fence_add_callback(struct dma_fence *fence, struct dma_fence_cb *cb, in dma_fence_add_callback() argument 622 INIT_LIST_HEAD(&cb->node); in dma_fence_add_callback() 629 cb->func = func; in dma_fence_add_callback() 630 list_add_tail(&cb->node, &fence->cb_list); in dma_fence_add_callback() 632 INIT_LIST_HEAD(&cb->node); in dma_fence_add_callback() 686 dma_fence_remove_callback(struct dma_fence *fence, struct dma_fence_cb *cb) in dma_fence_remove_callback() argument 693 ret = !list_empty(&cb->node); in dma_fence_remove_callback() 695 list_del_init(&cb->node); in dma_fence_remove_callback() 709 dma_fence_default_wait_cb(struct dma_fence *fence, struct dma_fence_cb *cb) in dma_fence_default_wait_cb() argument 712 container_of(cb, struct default_wait_cb, base); in dma_fence_default_wait_cb() [all …]
|
/drivers/dma/ |
D | dmaengine.h | 115 struct dmaengine_desc_callback *cb) in dmaengine_desc_get_callback() argument 117 cb->callback = tx->callback; in dmaengine_desc_get_callback() 118 cb->callback_result = tx->callback_result; in dmaengine_desc_get_callback() 119 cb->callback_param = tx->callback_param; in dmaengine_desc_get_callback() 132 dmaengine_desc_callback_invoke(struct dmaengine_desc_callback *cb, in dmaengine_desc_callback_invoke() argument 140 if (cb->callback_result) { in dmaengine_desc_callback_invoke() 143 cb->callback_result(cb->callback_param, result); in dmaengine_desc_callback_invoke() 144 } else if (cb->callback) { in dmaengine_desc_callback_invoke() 145 cb->callback(cb->callback_param); in dmaengine_desc_callback_invoke() 163 struct dmaengine_desc_callback cb; in dmaengine_desc_get_callback_invoke() local [all …]
|
/drivers/gpu/drm/i915/ |
D | i915_sw_fence.c | 412 struct i915_sw_dma_fence_cb *cb = container_of(data, typeof(*cb), base); in dma_i915_sw_fence_wake() local 414 i915_sw_fence_set_error_once(cb->fence, dma->error); in dma_i915_sw_fence_wake() 415 i915_sw_fence_complete(cb->fence); in dma_i915_sw_fence_wake() 416 kfree(cb); in dma_i915_sw_fence_wake() 421 struct i915_sw_dma_fence_cb_timer *cb = from_timer(cb, t, timer); in timer_i915_sw_fence_wake() local 424 fence = xchg(&cb->base.fence, NULL); in timer_i915_sw_fence_wake() 429 cb->dma->ops->get_driver_name(cb->dma), in timer_i915_sw_fence_wake() 430 cb->dma->ops->get_timeline_name(cb->dma), in timer_i915_sw_fence_wake() 431 cb->dma->seqno, in timer_i915_sw_fence_wake() 441 struct i915_sw_dma_fence_cb_timer *cb = in dma_i915_sw_fence_wake_timer() local [all …]
|
/drivers/net/wireless/marvell/mwifiex/ |
D | util.h | 55 struct mwifiex_cb *cb = (struct mwifiex_cb *)skb->cb; in MWIFIEX_SKB_RXCB() local 57 BUILD_BUG_ON(sizeof(struct mwifiex_cb) > sizeof(skb->cb)); in MWIFIEX_SKB_RXCB() 58 return &cb->rx_info; in MWIFIEX_SKB_RXCB() 63 struct mwifiex_cb *cb = (struct mwifiex_cb *)skb->cb; in MWIFIEX_SKB_TXCB() local 65 return &cb->tx_info; in MWIFIEX_SKB_TXCB() 71 struct mwifiex_cb *cb = (struct mwifiex_cb *)skb->cb; in mwifiex_store_mapping() local 73 memcpy(&cb->dma_mapping, mapping, sizeof(*mapping)); in mwifiex_store_mapping() 79 struct mwifiex_cb *cb = (struct mwifiex_cb *)skb->cb; in mwifiex_get_mapping() local 81 memcpy(mapping, &cb->dma_mapping, sizeof(*mapping)); in mwifiex_get_mapping()
|
/drivers/s390/net/ |
D | smsgiucv.c | 55 struct smsg_callback *cb; in smsg_message_pending() local 78 list_for_each_entry(cb, &smsg_list, list) in smsg_message_pending() 79 if (strncmp(buffer + 8, cb->prefix, cb->len) == 0) { in smsg_message_pending() 80 cb->callback(sender, buffer + 8); in smsg_message_pending() 91 struct smsg_callback *cb; in smsg_register_callback() local 93 cb = kmalloc(sizeof(struct smsg_callback), GFP_KERNEL); in smsg_register_callback() 94 if (!cb) in smsg_register_callback() 96 cb->prefix = prefix; in smsg_register_callback() 97 cb->len = strlen(prefix); in smsg_register_callback() 98 cb->callback = callback; in smsg_register_callback() [all …]
|
/drivers/platform/x86/intel_speed_select_if/ |
D | isst_if_common.c | 152 static void isst_mbox_resume_command(struct isst_if_cmd_cb *cb, in isst_mbox_resume_command() argument 163 (cb->cmd_callback)((u8 *)&mbox_cmd, &wr_only, 1); in isst_mbox_resume_command() 179 struct isst_if_cmd_cb *cb; in isst_resume_common() local 182 cb = &punit_callbacks[ISST_IF_DEV_MBOX]; in isst_resume_common() 183 if (cb->registered) in isst_resume_common() 184 isst_mbox_resume_command(cb, sst_cmd); in isst_resume_common() 442 static long isst_if_exec_multi_cmd(void __user *argp, struct isst_if_cmd_cb *cb) in isst_if_exec_multi_cmd() argument 457 cmd_ptr = kmalloc(cb->cmd_size, GFP_KERNEL); in isst_if_exec_multi_cmd() 462 ptr = argp + cb->offset; in isst_if_exec_multi_cmd() 472 if (copy_from_user(cmd_ptr, ptr, cb->cmd_size)) { in isst_if_exec_multi_cmd() [all …]
|
/drivers/media/test-drivers/vicodec/ |
D | codec-v4l2-fwht.c | 112 rf->cb = NULL; in prepare_raw_frame() 119 rf->cb = NULL; in prepare_raw_frame() 123 rf->cb = rf->luma + size; in prepare_raw_frame() 124 rf->cr = rf->cb + size / 4; in prepare_raw_frame() 128 rf->cb = rf->cr + size / 4; in prepare_raw_frame() 131 rf->cb = rf->luma + size; in prepare_raw_frame() 132 rf->cr = rf->cb + size / 2; in prepare_raw_frame() 137 rf->cb = rf->luma + size; in prepare_raw_frame() 138 rf->cr = rf->cb + 1; in prepare_raw_frame() 144 rf->cb = rf->cr + 1; in prepare_raw_frame() [all …]
|
/drivers/net/ethernet/hisilicon/hns/ |
D | hnae.c | 36 static int hnae_alloc_buffer(struct hnae_ring *ring, struct hnae_desc_cb *cb) in hnae_alloc_buffer() argument 44 cb->priv = p; in hnae_alloc_buffer() 45 cb->page_offset = 0; in hnae_alloc_buffer() 46 cb->reuse_flag = 0; in hnae_alloc_buffer() 47 cb->buf = page_address(p); in hnae_alloc_buffer() 48 cb->length = hnae_page_size(ring); in hnae_alloc_buffer() 49 cb->type = DESC_TYPE_PAGE; in hnae_alloc_buffer() 54 static void hnae_free_buffer(struct hnae_ring *ring, struct hnae_desc_cb *cb) in hnae_free_buffer() argument 56 if (unlikely(!cb->priv)) in hnae_free_buffer() 59 if (cb->type == DESC_TYPE_SKB) in hnae_free_buffer() [all …]
|
/drivers/misc/ |
D | enclosure.c | 116 struct enclosure_component_callbacks *cb) in enclosure_register() argument 122 BUG_ON(!cb); in enclosure_register() 131 edev->cb = cb; in enclosure_register() 176 edev->cb = &enclosure_null_callbacks; in enclosure_unregister() 439 if (edev->cb->show_id) in id_show() 440 return edev->cb->show_id(edev, buf); in id_show() 482 if (edev->cb->get_fault) in get_component_fault() 483 edev->cb->get_fault(edev, ecomp); in get_component_fault() 495 if (edev->cb->set_fault) in set_component_fault() 496 edev->cb->set_fault(edev, ecomp, val); in set_component_fault() [all …]
|
/drivers/infiniband/core/ |
D | iwpm_msg.c | 401 int iwpm_register_pid_cb(struct sk_buff *skb, struct netlink_callback *cb) in iwpm_register_pid_cb() argument 412 if (iwpm_parse_nlmsg(cb, IWPM_NLA_RREG_PID_MAX, in iwpm_register_pid_cb() 439 iwpm_user_pid = cb->nlh->nlmsg_pid; in iwpm_register_pid_cb() 444 atomic_set(&echo_nlmsg_seq, cb->nlh->nlmsg_seq); in iwpm_register_pid_cb() 474 int iwpm_add_mapping_cb(struct sk_buff *skb, struct netlink_callback *cb) in iwpm_add_mapping_cb() argument 485 if (iwpm_parse_nlmsg(cb, IWPM_NLA_RMANAGE_MAPPING_MAX, in iwpm_add_mapping_cb() 489 atomic_set(&echo_nlmsg_seq, cb->nlh->nlmsg_seq); in iwpm_add_mapping_cb() 552 struct netlink_callback *cb) in iwpm_add_and_query_mapping_cb() argument 564 if (iwpm_parse_nlmsg(cb, IWPM_NLA_RQUERY_MAPPING_MAX, in iwpm_add_and_query_mapping_cb() 567 atomic_set(&echo_nlmsg_seq, cb->nlh->nlmsg_seq); in iwpm_add_and_query_mapping_cb() [all …]
|
/drivers/media/tuners/ |
D | tuner-simple.c | 261 unsigned *frequency, u8 *config, u8 *cb) in simple_config_lookup() argument 277 *cb = t_params->ranges[i].cb; in simple_config_lookup() 281 i, *config, *cb); in simple_config_lookup() 289 u8 *config, u8 *cb, unsigned int rf) in simple_set_rf_input() argument 297 *cb |= 0x08; in simple_set_rf_input() 300 *cb &= ~0x08; in simple_set_rf_input() 307 *cb |= 0x01; in simple_set_rf_input() 310 *cb &= ~0x01; in simple_set_rf_input() 321 u8 *config, u8 *cb) in simple_std_setup() argument 332 *cb &= ~0x03; in simple_std_setup() [all …]
|
/drivers/mcb/ |
D | mcb-parse.c | 30 struct chameleon_bar *cb, in chameleon_parse_bdd() argument 37 struct chameleon_bar *cb, in chameleon_parse_gdd() argument 77 dev_mapbase = cb[mdev->bar].addr; in chameleon_parse_gdd() 115 struct chameleon_bar *cb, int bar_count) in chameleon_parse_bar() argument 124 cb[i].addr = readl(p); in chameleon_parse_bar() 125 cb[i].size = readl(p + 4); in chameleon_parse_bar() 132 struct chameleon_bar **cb) in chameleon_get_bar() argument 170 *cb = c; in chameleon_get_bar() 179 struct chameleon_bar *cb; in chameleon_parse_cells() local 212 bar_count = chameleon_get_bar(&p, mapbase, &cb); in chameleon_parse_cells() [all …]
|