/drivers/infiniband/sw/rdmavt/ |
D | cq.c | 122 struct kthread_worker *worker; in rvt_cq_enter() local 128 worker = cq->rdi->worker; in rvt_cq_enter() 129 if (likely(worker)) { in rvt_cq_enter() 132 kthread_queue_work(worker, &cq->comptask); in rvt_cq_enter() 515 if (rdi->worker) in rvt_driver_cq_init() 518 rdi->worker = kzalloc(sizeof(*rdi->worker), GFP_KERNEL); in rvt_driver_cq_init() 519 if (!rdi->worker) in rvt_driver_cq_init() 521 kthread_init_worker(rdi->worker); in rvt_driver_cq_init() 524 rdi->worker, in rvt_driver_cq_init() 528 kfree(rdi->worker); in rvt_driver_cq_init() [all …]
|
/drivers/platform/olpc/ |
D | olpc-ec.c | 36 struct work_struct worker; member 70 struct olpc_ec_priv *ec = container_of(w, struct olpc_ec_priv, worker); in olpc_ec_worker() 96 schedule_work(&ec->worker); in olpc_ec_worker() 114 schedule_work(&ec->worker); in queue_ec_descriptor() 271 INIT_WORK(&ec->worker, olpc_ec_worker); in olpc_ec_probe()
|
/drivers/gpu/drm/ |
D | drm_flip_work.c | 107 queue_work(wq, &work->worker); in drm_flip_work_commit() 113 struct drm_flip_work *work = container_of(w, struct drm_flip_work, worker); in flip_worker() 153 INIT_WORK(&work->worker, flip_worker); in drm_flip_work_init()
|
/drivers/macintosh/ams/ |
D | ams-core.c | 78 schedule_work(&ams_info.worker); in ams_handle_irq() 198 INIT_WORK(&ams_info.worker, ams_worker); in ams_init() 229 flush_work(&ams_info.worker); in ams_sensor_detach()
|
D | ams.h | 33 struct work_struct worker; member
|
/drivers/vhost/ |
D | vhost.c | 237 if (dev->worker) { in vhost_work_flush() 257 if (!dev->worker) in vhost_work_queue() 266 wake_up_process(dev->worker); in vhost_work_queue() 409 dev->worker = NULL; in vhost_dev_init() 475 struct task_struct *worker; in vhost_dev_set_owner() local 486 worker = kthread_create(vhost_worker, dev, "vhost-%d", current->pid); in vhost_dev_set_owner() 487 if (IS_ERR(worker)) { in vhost_dev_set_owner() 488 err = PTR_ERR(worker); in vhost_dev_set_owner() 492 dev->worker = worker; in vhost_dev_set_owner() 493 wake_up_process(worker); /* avoid contributing to loadavg */ in vhost_dev_set_owner() [all …]
|
D | vhost.h | 160 struct task_struct *worker; member
|
/drivers/hid/ |
D | uhid.c | 56 struct work_struct worker; member 63 struct uhid_device *uhid = container_of(work, struct uhid_device, worker); in uhid_device_add_worker() 534 schedule_work(&uhid->worker); in uhid_dev_create2() 578 cancel_work_sync(&uhid->worker); in uhid_dev_destroy() 642 INIT_WORK(&uhid->worker, uhid_device_add_worker); in uhid_char_open()
|
D | hid-wiimote.h | 110 struct work_struct worker; member
|
D | hid-wiimote-core.c | 47 worker); in wiimote_queue_worker() 104 schedule_work(&wdata->queue.worker); in wiimote_queue() 1734 INIT_WORK(&wdata->queue.worker, wiimote_queue_worker); in wiimote_create() 1768 cancel_work_sync(&wdata->queue.worker); in wiimote_destroy()
|
/drivers/block/ |
D | loop.h | 59 struct kthread_worker worker; member
|
D | loop.c | 846 kthread_flush_worker(&lo->worker); in loop_unprepare_queue() 852 kthread_init_worker(&lo->worker); in loop_prepare_queue() 854 &lo->worker, "loop%d", lo->lo_number); in loop_prepare_queue() 1682 kthread_queue_work(&lo->worker, &cmd->work); in loop_queue_rq()
|
/drivers/media/usb/hdpvr/ |
D | hdpvr-video.c | 251 worker); in hdpvr_transmit_buffers() 315 INIT_WORK(&dev->worker, hdpvr_transmit_buffers); in hdpvr_start_streaming() 316 schedule_work(&dev->worker); in hdpvr_start_streaming() 350 flush_work(&dev->worker); in hdpvr_stop_streaming() 1123 flush_work(&dev->worker); in hdpvr_device_release()
|
D | hdpvr.h | 110 struct work_struct worker; member
|
D | hdpvr-core.c | 428 flush_work(&dev->worker); in hdpvr_disconnect()
|
/drivers/s390/net/ |
D | qeth_l2_main.c | 1623 struct work_struct worker; member 1631 container_of(work, struct qeth_bridge_state_data, worker); in qeth_bridge_state_change_worker() 1686 INIT_WORK(&data->worker, qeth_bridge_state_change_worker); in qeth_bridge_state_change() 1690 queue_work(qeth_wq, &data->worker); in qeth_bridge_state_change() 1694 struct work_struct worker; member 1702 container_of(work, struct qeth_bridge_host_data, worker); in qeth_bridge_host_event_worker() 1758 INIT_WORK(&data->worker, qeth_bridge_host_event_worker); in qeth_bridge_host_event() 1762 queue_work(qeth_wq, &data->worker); in qeth_bridge_host_event()
|
/drivers/s390/block/ |
D | dasd_alias.c | 153 INIT_WORK(&lcu->suc_data.worker, summary_unit_check_handling_work); in _allocate_lcu() 262 cancel_work_sync(&lcu->suc_data.worker); in dasd_alias_disconnect_device_from_lcu() 879 worker); in summary_unit_check_handling_work() 941 if (!schedule_work(&lcu->suc_data.worker)) in dasd_alias_handle_summary_unit_check()
|
D | dasd_eckd.h | 495 struct work_struct worker; member
|
D | dasd_eckd.c | 107 struct work_struct worker; member 119 struct work_struct worker; member 1275 data = container_of(work, struct path_verification_work_data, worker); in do_path_verification_work() 1453 INIT_WORK(&data->worker, do_path_verification_work); in dasd_eckd_verify_path() 1457 schedule_work(&data->worker); in dasd_eckd_verify_path() 5680 data = container_of(work, struct check_attention_work_data, worker); in dasd_eckd_check_attention_work() 5707 INIT_WORK(&data->worker, dasd_eckd_check_attention_work); in dasd_eckd_check_attention() 5711 schedule_work(&data->worker); in dasd_eckd_check_attention()
|
/drivers/staging/speakup/ |
D | TODO | 21 in these cases. Pushing work to some worker thread would probably help,
|
/drivers/block/drbd/ |
D | drbd_main.c | 2286 struct work_struct worker; member 2294 struct retry_worker *retry = container_of(ws, struct retry_worker, worker); in do_retry() 2358 queue_work(retry.wq, &retry.worker); in drbd_restart_request() 2636 connection->worker.reset_cpu_mask = 1; in set_resource_options() 2722 drbd_thread_init(resource, &connection->worker, drbd_worker, "worker"); in conn_create() 2723 connection->worker.connection = connection; in conn_create() 2779 INIT_WORK(&device->submit.worker, do_submit); in init_submitter() 3021 INIT_WORK(&retry.worker, do_retry); in drbd_init() 3633 D_ASSERT(device, current == first_peer_device(device)->connection->worker.task); in drbd_queue_bitmap_io() 3675 D_ASSERT(device, current != first_peer_device(device)->connection->worker.task); in drbd_bitmap_io()
|
D | drbd_req.c | 1210 queue_work(device->submit.wq, &device->submit.worker); in drbd_queue_write() 1465 struct drbd_device *device = container_of(ws, struct drbd_device, submit.worker); in do_submit()
|
/drivers/md/ |
D | dm-era-target.c | 1152 struct work_struct worker; member 1206 queue_work(era->wq, &era->worker); in wake_worker() 1310 struct era *era = container_of(ws, struct era, worker); in do_work() 1505 INIT_WORK(&era->worker, do_work); in era_ctr()
|
D | dm-thin.c | 245 struct work_struct worker; member 399 queue_work(pool->wq, &pool->worker); in wake_worker() 2286 struct pool *pool = container_of(ws, struct pool, worker); in do_worker() 2334 struct work_struct worker; member 2340 return container_of(ws, struct pool_work, worker); in to_pool_work() 2351 INIT_WORK_ONSTACK(&pw->worker, fn); in pool_work_wait() 2353 queue_work(pool->wq, &pw->worker); in pool_work_wait() 2906 INIT_WORK(&pool->worker, do_worker); in pool_create()
|
D | raid5.c | 5787 struct r5worker *worker, in handle_active_stripes() argument 5837 struct r5worker *worker = container_of(work, struct r5worker, work); in raid5_do_work() local 5838 struct r5worker_group *group = worker->group; in raid5_do_work() 5852 released = release_stripe_list(conf, worker->temp_inactive_list); in raid5_do_work() 5854 batch_size = handle_active_stripes(conf, group_id, worker, in raid5_do_work() 5855 worker->temp_inactive_list); in raid5_do_work() 5856 worker->working = false; in raid5_do_work() 6314 struct r5worker *worker = group->workers + j; in alloc_thread_groups() local 6315 worker->group = group; in alloc_thread_groups() 6316 INIT_WORK(&worker->work, raid5_do_work); in alloc_thread_groups() [all …]
|