/drivers/s390/cio/ |
D | vfio_ccw_ops.c | 22 struct vfio_ccw_private *private; in vfio_ccw_mdev_reset() local 26 private = dev_get_drvdata(mdev_parent_dev(mdev)); in vfio_ccw_mdev_reset() 27 sch = private->sch; in vfio_ccw_mdev_reset() 42 private->state = VFIO_CCW_STATE_IDLE; in vfio_ccw_mdev_reset() 51 struct vfio_ccw_private *private = in vfio_ccw_mdev_notifier() local 61 if (!cp_iova_pinned(&private->cp, unmap->iova)) in vfio_ccw_mdev_notifier() 64 if (vfio_ccw_mdev_reset(private->mdev)) in vfio_ccw_mdev_notifier() 67 cp_free(&private->cp); in vfio_ccw_mdev_notifier() 90 struct vfio_ccw_private *private = dev_get_drvdata(dev); in available_instances_show() local 92 return sprintf(buf, "%d\n", atomic_read(&private->avail)); in available_instances_show() [all …]
|
D | device_fsm.c | 41 struct io_subchannel_private *private; in ccw_timeout_log() local 46 private = to_io_private(sch); in ccw_timeout_log() 47 orb = &private->orb; in ccw_timeout_log() 71 &private->dma_area->sense_ccw || in ccw_timeout_log() 73 cdev->private->dma_area->iccws) in ccw_timeout_log() 84 cdev->private->state); in ccw_timeout_log() 91 &cdev->private->flags, sizeof(cdev->private->flags), 0); in ccw_timeout_log() 117 del_timer(&cdev->private->timer); in ccw_device_set_timeout() 120 if (timer_pending(&cdev->private->timer)) { in ccw_device_set_timeout() 121 if (mod_timer(&cdev->private->timer, jiffies + expires)) in ccw_device_set_timeout() [all …]
|
D | vfio_ccw_drv.c | 38 struct vfio_ccw_private *private = dev_get_drvdata(&sch->dev); in vfio_ccw_sch_quiesce() local 64 private->completion = &completion; in vfio_ccw_sch_quiesce() 70 private->completion = NULL; in vfio_ccw_sch_quiesce() 76 private->state = VFIO_CCW_STATE_NOT_OPER; in vfio_ccw_sch_quiesce() 83 struct vfio_ccw_private *private; in vfio_ccw_sch_io_todo() local 87 private = container_of(work, struct vfio_ccw_private, io_work); in vfio_ccw_sch_io_todo() 88 irb = &private->irb; in vfio_ccw_sch_io_todo() 93 cp_update_scsw(&private->cp, &irb->scsw); in vfio_ccw_sch_io_todo() 94 if (is_final && private->state == VFIO_CCW_STATE_CP_PENDING) in vfio_ccw_sch_io_todo() 95 cp_free(&private->cp); in vfio_ccw_sch_io_todo() [all …]
|
D | device_pgid.c | 36 struct ccw_dev_id *id = &cdev->private->dev_id; in verify_done() 37 int mpath = cdev->private->flags.mpath; in verify_done() 38 int pgroup = cdev->private->flags.pgroup; in verify_done() 59 struct ccw_request *req = &cdev->private->req; in nop_build_cp() 60 struct ccw1 *cp = cdev->private->dma_area->iccws; in nop_build_cp() 75 struct ccw_request *req = &cdev->private->req; in nop_do() 78 ~cdev->private->path_noirq_mask); in nop_do() 107 struct ccw_request *req = &cdev->private->req; in nop_callback() 114 cdev->private->path_noirq_mask |= req->lpm; in nop_callback() 117 cdev->private->path_notoper_mask |= req->lpm; in nop_callback() [all …]
|
D | vfio_ccw_fsm.c | 21 static int fsm_io_helper(struct vfio_ccw_private *private) in fsm_io_helper() argument 30 sch = private->sch; in fsm_io_helper() 34 orb = cp_get_orb(&private->cp, (u32)(addr_t)sch, sch->lpm); in fsm_io_helper() 55 private->state = VFIO_CCW_STATE_CP_PENDING; in fsm_io_helper() 83 static int fsm_do_halt(struct vfio_ccw_private *private) in fsm_do_halt() argument 90 sch = private->sch; in fsm_do_halt() 124 static int fsm_do_clear(struct vfio_ccw_private *private) in fsm_do_clear() argument 131 sch = private->sch; in fsm_do_clear() 162 static void fsm_notoper(struct vfio_ccw_private *private, in fsm_notoper() argument 165 struct subchannel *sch = private->sch; in fsm_notoper() [all …]
|
D | eadm_sch.c | 100 struct eadm_private *private = from_timer(private, t, timer); in eadm_subchannel_timeout() local 101 struct subchannel *sch = private->sch; in eadm_subchannel_timeout() 113 struct eadm_private *private = get_eadm_private(sch); in eadm_subchannel_set_timeout() local 116 del_timer(&private->timer); in eadm_subchannel_set_timeout() 119 if (timer_pending(&private->timer)) { in eadm_subchannel_set_timeout() 120 if (mod_timer(&private->timer, jiffies + expires)) in eadm_subchannel_set_timeout() 123 private->timer.expires = jiffies + expires; in eadm_subchannel_set_timeout() 124 add_timer(&private->timer); in eadm_subchannel_set_timeout() 129 struct eadm_private *private = get_eadm_private(sch); in eadm_subchannel_irq() local 148 if (private->state != EADM_BUSY) { in eadm_subchannel_irq() [all …]
|
D | device_ops.c | 47 cdev->private->options.fast = (flags & CCWDEV_EARLY_NOTIFICATION) != 0; in ccw_device_set_options_mask() 48 cdev->private->options.repall = (flags & CCWDEV_REPORT_ALL) != 0; in ccw_device_set_options_mask() 49 cdev->private->options.pgroup = (flags & CCWDEV_DO_PATHGROUP) != 0; in ccw_device_set_options_mask() 50 cdev->private->options.force = (flags & CCWDEV_ALLOW_FORCE) != 0; in ccw_device_set_options_mask() 51 cdev->private->options.mpath = (flags & CCWDEV_DO_MULTIPATH) != 0; in ccw_device_set_options_mask() 72 cdev->private->options.repall) || in ccw_device_set_options() 74 cdev->private->options.fast)) in ccw_device_set_options() 76 cdev->private->options.fast |= (flags & CCWDEV_EARLY_NOTIFICATION) != 0; in ccw_device_set_options() 77 cdev->private->options.repall |= (flags & CCWDEV_REPORT_ALL) != 0; in ccw_device_set_options() 78 cdev->private->options.pgroup |= (flags & CCWDEV_DO_PATHGROUP) != 0; in ccw_device_set_options() [all …]
|
D | device.c | 265 if (cdev->private->flags.initialized) { in ccw_device_unregister() 266 cdev->private->flags.initialized = 0; in ccw_device_unregister() 305 cdev->private->state != DEV_STATE_DISCONNECTED) { in ccw_device_set_offline() 307 wait_event(cdev->private->wait_q, (dev_fsm_final_state(cdev) || in ccw_device_set_offline() 308 cdev->private->state == DEV_STATE_DISCONNECTED)); in ccw_device_set_offline() 316 "0.%x.%04x\n", ret, cdev->private->dev_id.ssid, in ccw_device_set_offline() 317 cdev->private->dev_id.devno); in ccw_device_set_offline() 320 state = cdev->private->state; in ccw_device_set_offline() 324 cdev->private->state = state; in ccw_device_set_offline() 327 wait_event(cdev->private->wait_q, (dev_fsm_final_state(cdev) || in ccw_device_set_offline() [all …]
|
D | vfio_ccw_async.c | 15 static ssize_t vfio_ccw_async_region_read(struct vfio_ccw_private *private, in vfio_ccw_async_region_read() argument 27 mutex_lock(&private->io_mutex); in vfio_ccw_async_region_read() 28 region = private->region[i].data; in vfio_ccw_async_region_read() 33 mutex_unlock(&private->io_mutex); in vfio_ccw_async_region_read() 37 static ssize_t vfio_ccw_async_region_write(struct vfio_ccw_private *private, in vfio_ccw_async_region_write() argument 49 if (!mutex_trylock(&private->io_mutex)) in vfio_ccw_async_region_write() 52 region = private->region[i].data; in vfio_ccw_async_region_write() 58 vfio_ccw_fsm_event(private, VFIO_CCW_EVENT_ASYNC_REQ); in vfio_ccw_async_region_write() 63 mutex_unlock(&private->io_mutex); in vfio_ccw_async_region_write() 67 static void vfio_ccw_async_region_release(struct vfio_ccw_private *private, in vfio_ccw_async_region_release() argument [all …]
|
D | device_id.c | 101 struct ccw_dev_id *dev_id = &cdev->private->dev_id; in diag210_get_dev_info() 102 struct senseid *senseid = &cdev->private->dma_area->senseid; in diag210_get_dev_info() 136 cdev->private->flags.esid = 0; in snsid_init() 138 memset(&cdev->private->dma_area->senseid, 0, in snsid_init() 139 sizeof(cdev->private->dma_area->senseid)); in snsid_init() 140 cdev->private->dma_area->senseid.cu_type = 0xffff; in snsid_init() 148 struct cmd_scsw *scsw = &cdev->private->dma_area->irb.scsw.cmd; in snsid_check() 154 if (cdev->private->dma_area->senseid.cu_type == 0xffff) in snsid_check() 157 if (cdev->private->dma_area->senseid.reserved != 0xff) in snsid_check() 161 cdev->private->flags.esid = 1; in snsid_check() [all …]
|
D | cmf.c | 218 if (!cdev->private->cmb) in set_schib_wait() 226 if (cdev->private->state != DEV_STATE_ONLINE) in set_schib_wait() 235 cdev->private->state = DEV_STATE_CMFCHANGE; in set_schib_wait() 236 cdev->private->cmb_wait = &set_data; in set_schib_wait() 246 if (cdev->private->state == DEV_STATE_CMFCHANGE) in set_schib_wait() 247 cdev->private->state = DEV_STATE_ONLINE; in set_schib_wait() 250 cdev->private->cmb_wait = NULL; in set_schib_wait() 259 struct set_schib_struct *set_data = cdev->private->cmb_wait; in retry_set_schib() 286 cmb_data = cdev->private->cmb; in cmf_copy_block() 304 if (!cdev->private->cmb) in cmf_cmb_copy_wait() [all …]
|
D | device_status.c | 41 cdev->private->dev_id.devno, sch->schid.ssid, in ccw_device_msg_control_check() 68 cdev->private->flags.doverify = 1; in ccw_device_path_notoper() 82 cdev->private->dma_area->irb.scsw.cmd.ectl = 0; in ccw_device_accumulate_ecw() 85 cdev->private->dma_area->irb.scsw.cmd.ectl = irb->scsw.cmd.ectl; in ccw_device_accumulate_ecw() 87 if (!cdev->private->dma_area->irb.scsw.cmd.ectl) in ccw_device_accumulate_ecw() 90 memcpy(&cdev->private->dma_area->irb.ecw, irb->ecw, sizeof(irb->ecw)); in ccw_device_accumulate_ecw() 121 cdev_irb = &cdev->private->dma_area->irb; in ccw_device_accumulate_esw() 182 cdev->private->flags.doverify = 1; in ccw_device_accumulate_esw() 213 memcpy(&cdev->private->dma_area->irb, irb, sizeof(struct irb)); in ccw_device_accumulate_irb() 222 cdev_irb = &cdev->private->dma_area->irb; in ccw_device_accumulate_irb() [all …]
|
D | ccwreq.c | 43 struct ccw_request *req = &cdev->private->req; in ccwreq_next_path() 60 struct ccw_request *req = &cdev->private->req; in ccwreq_stop() 66 memset(&cdev->private->dma_area->irb, 0, sizeof(struct irb)); in ccwreq_stop() 77 struct ccw_request *req = &cdev->private->req; in ccwreq_do() 89 memset(&cdev->private->dma_area->irb, 0, sizeof(struct irb)); in ccwreq_do() 122 struct ccw_request *req = &cdev->private->req; in ccw_request_start() 154 struct ccw_request *req = &cdev->private->req; in ccw_request_cancel() 172 struct irb *irb = &cdev->private->dma_area->irb; in ccwreq_status() 188 CIO_HEX_EVENT(2, &cdev->private->dev_id, in ccwreq_status() 190 CIO_HEX_EVENT(2, &cdev->private->dma_area->irb.ecw, in ccwreq_status() [all …]
|
D | vfio_ccw_private.h | 35 ssize_t (*read)(struct vfio_ccw_private *private, char __user *buf, 37 ssize_t (*write)(struct vfio_ccw_private *private, 39 void (*release)(struct vfio_ccw_private *private, 52 int vfio_ccw_register_dev_region(struct vfio_ccw_private *private, 57 int vfio_ccw_register_async_dev_regions(struct vfio_ccw_private *private); 135 static inline void vfio_ccw_fsm_event(struct vfio_ccw_private *private, in vfio_ccw_fsm_event() argument 138 vfio_ccw_jumptable[private->state][event](private, event); in vfio_ccw_fsm_event()
|
/drivers/gpu/drm/mediatek/ |
D | mtk_drm_drv.c | 39 static void mtk_atomic_schedule(struct mtk_drm_private *private, in mtk_atomic_schedule() argument 42 private->commit.state = state; in mtk_atomic_schedule() 43 schedule_work(&private->commit.work); in mtk_atomic_schedule() 46 static void mtk_atomic_complete(struct mtk_drm_private *private, in mtk_atomic_complete() argument 49 struct drm_device *drm = private->drm; in mtk_atomic_complete() 80 struct mtk_drm_private *private = container_of(work, in mtk_atomic_work() local 83 mtk_atomic_complete(private, private->commit.state); in mtk_atomic_work() 90 struct mtk_drm_private *private = drm->dev_private; in mtk_atomic_commit() local 97 mutex_lock(&private->commit.lock); in mtk_atomic_commit() 98 flush_work(&private->commit.work); in mtk_atomic_commit() [all …]
|
/drivers/s390/block/ |
D | dasd_diag.c | 107 struct dasd_diag_private *private = device->private; in mdsk_init_io() local 108 struct dasd_diag_init_io *iib = &private->iib; in mdsk_init_io() 113 iib->dev_nr = private->dev_id.devno; in mdsk_init_io() 131 struct dasd_diag_private *private = device->private; in mdsk_term_io() local 132 struct dasd_diag_init_io *iib = &private->iib; in mdsk_term_io() 136 iib->dev_nr = private->dev_id.devno; in mdsk_term_io() 167 struct dasd_diag_private *private; in dasd_start_diag() local 178 private = device->private; in dasd_start_diag() 181 private->iob.dev_nr = private->dev_id.devno; in dasd_start_diag() 182 private->iob.key = 0; in dasd_start_diag() [all …]
|
D | dasd_eckd.c | 213 struct dasd_eckd_private *private = device->private; in set_timestamp() local 221 if ((rc && !private->rdc_data.facilities.XRC_supported) || in set_timestamp() 242 struct dasd_eckd_private *private = device->private; in define_extent() local 265 data->attributes.operation = private->attrib.operation; in define_extent() 274 data->attributes.operation = private->attrib.operation; in define_extent() 282 data->attributes.operation = private->attrib.operation; in define_extent() 300 data->attributes.operation = private->attrib.operation; in define_extent() 305 data->attributes.operation = private->attrib.operation; in define_extent() 317 if ((private->rdc_data.cu_type == 0x2105 || in define_extent() 318 private->rdc_data.cu_type == 0x2107 || in define_extent() [all …]
|
D | dasd_alias.c | 189 struct dasd_eckd_private *private = device->private; in dasd_alias_make_device_known_to_lcu() local 232 private->lcu = lcu; in dasd_alias_make_device_known_to_lcu() 246 struct dasd_eckd_private *private = device->private; in dasd_alias_disconnect_device_from_lcu() local 253 lcu = private->lcu; in dasd_alias_disconnect_device_from_lcu() 281 private->lcu = NULL; in dasd_alias_disconnect_device_from_lcu() 317 struct dasd_eckd_private *private = device->private; in _add_device_to_lcu() local 322 private->uid.type = lcu->uac->unit[private->uid.real_unit_addr].ua_type; in _add_device_to_lcu() 323 private->uid.base_unit_addr = in _add_device_to_lcu() 324 lcu->uac->unit[private->uid.real_unit_addr].base_ua; in _add_device_to_lcu() 325 uid = private->uid; in _add_device_to_lcu() [all …]
|
D | dasd_fba.c | 129 struct dasd_fba_private *private = device->private; in dasd_fba_check_characteristics() local 134 if (!private) { in dasd_fba_check_characteristics() 135 private = kzalloc(sizeof(*private), GFP_KERNEL | GFP_DMA); in dasd_fba_check_characteristics() 136 if (!private) { in dasd_fba_check_characteristics() 142 device->private = private; in dasd_fba_check_characteristics() 144 memset(private, 0, sizeof(*private)); in dasd_fba_check_characteristics() 150 device->private = NULL; in dasd_fba_check_characteristics() 151 kfree(private); in dasd_fba_check_characteristics() 159 &private->rdc_data, 32); in dasd_fba_check_characteristics() 165 device->private = NULL; in dasd_fba_check_characteristics() [all …]
|
/drivers/usb/serial/ |
D | upd78f0730.c | 165 struct upd78f0730_port_private *private; in upd78f0730_port_probe() local 167 private = kzalloc(sizeof(*private), GFP_KERNEL); in upd78f0730_port_probe() 168 if (!private) in upd78f0730_port_probe() 171 mutex_init(&private->lock); in upd78f0730_port_probe() 172 usb_set_serial_port_data(port, private); in upd78f0730_port_probe() 179 struct upd78f0730_port_private *private; in upd78f0730_port_remove() local 181 private = usb_get_serial_port_data(port); in upd78f0730_port_remove() 182 mutex_destroy(&private->lock); in upd78f0730_port_remove() 183 kfree(private); in upd78f0730_port_remove() 191 struct upd78f0730_port_private *private; in upd78f0730_tiocmget() local [all …]
|
/drivers/gpu/drm/rockchip/ |
D | rockchip_drm_drv.c | 48 struct rockchip_drm_private *private = drm_dev->dev_private; in rockchip_drm_dma_attach_device() local 54 ret = iommu_attach_device(private->domain, dev); in rockchip_drm_dma_attach_device() 66 struct rockchip_drm_private *private = drm_dev->dev_private; in rockchip_drm_dma_detach_device() local 67 struct iommu_domain *domain = private->domain; in rockchip_drm_dma_detach_device() 77 struct rockchip_drm_private *private = drm_dev->dev_private; in rockchip_drm_init_iommu() local 84 private->domain = iommu_domain_alloc(&platform_bus_type); in rockchip_drm_init_iommu() 85 if (!private->domain) in rockchip_drm_init_iommu() 88 geometry = &private->domain->geometry; in rockchip_drm_init_iommu() 94 drm_mm_init(&private->mm, start, end - start + 1); in rockchip_drm_init_iommu() 95 mutex_init(&private->mm_lock); in rockchip_drm_init_iommu() [all …]
|
D | rockchip_drm_fbdev.c | 25 struct rockchip_drm_private *private = to_drm_private(helper); in rockchip_fbdev_mmap() local 27 return rockchip_gem_mmap_buf(private->fbdev_bo, vma); in rockchip_fbdev_mmap() 42 struct rockchip_drm_private *private = to_drm_private(helper); in rockchip_drm_fbdev_create() local 67 private->fbdev_bo = &rk_obj->base; in rockchip_drm_fbdev_create() 77 private->fbdev_bo); in rockchip_drm_fbdev_create() 116 struct rockchip_drm_private *private = dev->dev_private; in rockchip_drm_fbdev_init() local 123 helper = &private->fbdev_helper; in rockchip_drm_fbdev_init() 159 struct rockchip_drm_private *private = dev->dev_private; in rockchip_drm_fbdev_fini() local 162 helper = &private->fbdev_helper; in rockchip_drm_fbdev_fini()
|
D | rockchip_drm_gem.c | 21 struct rockchip_drm_private *private = drm->dev_private; in rockchip_gem_iommu_map() local 25 mutex_lock(&private->mm_lock); in rockchip_gem_iommu_map() 26 ret = drm_mm_insert_node_generic(&private->mm, &rk_obj->mm, in rockchip_gem_iommu_map() 29 mutex_unlock(&private->mm_lock); in rockchip_gem_iommu_map() 38 ret = iommu_map_sg(private->domain, rk_obj->dma_addr, rk_obj->sgt->sgl, in rockchip_gem_iommu_map() 52 mutex_lock(&private->mm_lock); in rockchip_gem_iommu_map() 54 mutex_unlock(&private->mm_lock); in rockchip_gem_iommu_map() 62 struct rockchip_drm_private *private = drm->dev_private; in rockchip_gem_iommu_unmap() local 64 iommu_unmap(private->domain, rk_obj->dma_addr, rk_obj->size); in rockchip_gem_iommu_unmap() 66 mutex_lock(&private->mm_lock); in rockchip_gem_iommu_unmap() [all …]
|
/drivers/misc/ocxl/ |
D | afu_irq.c | 14 irqreturn_t (*handler)(void *private); 15 void (*free_private)(void *private); 16 void *private; member 30 irqreturn_t (*handler)(void *private), in ocxl_irq_set_handler() argument 31 void (*free_private)(void *private), in ocxl_irq_set_handler() argument 32 void *private) in ocxl_irq_set_handler() 45 irq->private = private; in ocxl_irq_set_handler() 64 return irq->handler(irq->private); in afu_irq_handler() 163 irq->free_private(irq->private); in afu_irq_free()
|
/drivers/md/ |
D | dm-linear.c | 66 ti->private = lc; in linear_ctr() 76 struct linear_c *lc = (struct linear_c *) ti->private; in linear_dtr() 84 struct linear_c *lc = ti->private; in linear_map_sector() 91 struct linear_c *lc = ti->private; in linear_map_bio() 109 struct linear_c *lc = (struct linear_c *) ti->private; in linear_status() 125 struct linear_c *lc = (struct linear_c *) ti->private; in linear_prepare_ioctl() 143 struct linear_c *lc = ti->private; in linear_report_zones() 155 struct linear_c *lc = ti->private; in linear_iterate_devices() 165 struct linear_c *lc = ti->private; in linear_dax_direct_access() 180 struct linear_c *lc = ti->private; in linear_dax_copy_from_iter() [all …]
|