/drivers/isdn/mISDN/ |
D | hwchannel.c | 16 struct dchannel *dch = container_of(ws, struct dchannel, workq); in dchannel_bh() 39 struct bchannel *bch = container_of(ws, struct bchannel, workq); in bchannel_bh() 69 INIT_WORK(&ch->workq, dchannel_bh); in mISDN_initdchannel() 92 INIT_WORK(&ch->workq, bchannel_bh); in mISDN_initbchannel() 110 flush_work(&ch->workq); in mISDN_freedchannel() 150 cancel_work_sync(&ch->workq); in mISDN_freebchannel()
|
D | l1oip.h | 68 struct work_struct workq; member
|
D | dsp_core.c | 681 schedule_work(&dsp->workq); in dsp_function() 888 schedule_work(&dsp->workq); in dsp_function() 975 cancel_work_sync(&dsp->workq); in dsp_ctrl() 1007 struct dsp *dsp = container_of(work, struct dsp, workq); in dsp_send_bh() 1065 INIT_WORK(&ndsp->workq, (void *)dsp_send_bh); in dspcreate()
|
D | stack.c | 31 wake_up_interruptible(&st->workq); in _queue_message() 280 wait_event_interruptible(st->workq, (st->status & in mISDNStackd() 378 init_waitqueue_head(&newst->workq); in create_stack() 638 wake_up_interruptible(&st->workq); in delete_stack()
|
D | dsp.h | 185 struct work_struct workq; member
|
D | l1oip_core.c | 807 struct l1oip *hc = container_of(work, struct l1oip, workq); in l1oip_send_bh() 826 schedule_work(&hc->workq); in l1oip_keepalive() 1242 cancel_work_sync(&hc->workq); in release_card() 1488 INIT_WORK(&hc->workq, (void *)l1oip_send_bh); in l1oip_init()
|
D | dsp_cmx.c | 1585 schedule_work(&dsp->workq); 1618 schedule_work(&dsp->workq); 1940 schedule_work(&dsp->workq); 1956 schedule_work(&member->dsp->workq);
|
/drivers/hwmon/ |
D | xgene-hwmon.c | 105 struct work_struct workq; member 442 ctx = container_of(work, struct xgene_hwmon_dev, workq); in xgene_hwmon_evt_work() 522 schedule_work(&ctx->workq); in xgene_hwmon_rx_cb() 591 schedule_work(&ctx->workq); in xgene_hwmon_pcc_rx_cb() 637 INIT_WORK(&ctx->workq, xgene_hwmon_evt_work); in xgene_hwmon_probe() 750 schedule_work(&ctx->workq); in xgene_hwmon_probe() 771 cancel_work_sync(&ctx->workq); in xgene_hwmon_remove()
|
/drivers/net/ethernet/huawei/hinic/ |
D | hinic_hw_mgmt.c | 469 queue_work(pf_to_mgmt->workq, &mgmt_work->work); in mgmt_recv_msg_handler() 639 pf_to_mgmt->workq = create_singlethread_workqueue("hinic_mgmt"); in hinic_pf_to_mgmt_init() 640 if (!pf_to_mgmt->workq) { in hinic_pf_to_mgmt_init() 650 destroy_workqueue(pf_to_mgmt->workq); in hinic_pf_to_mgmt_init() 658 destroy_workqueue(pf_to_mgmt->workq); in hinic_pf_to_mgmt_init() 683 destroy_workqueue(pf_to_mgmt->workq); in hinic_pf_to_mgmt_free()
|
D | hinic_hw_eqs.c | 399 queue_work(aeqs->workq, &aeq_work->work); in aeq_interrupt() 867 aeqs->workq = create_singlethread_workqueue(HINIC_EQS_WQ_NAME); in hinic_aeqs_init() 868 if (!aeqs->workq) in hinic_aeqs_init() 889 destroy_workqueue(aeqs->workq); in hinic_aeqs_init() 904 destroy_workqueue(aeqs->workq); in hinic_aeqs_free()
|
D | hinic_dev.h | 91 struct workqueue_struct *workq; member
|
D | hinic_hw_mbox.c | 505 queue_work(func_to_func->workq, &mbox_work->work); in recv_mbox_handler() 619 queue_work(func_to_func->workq, &mbox_work->work); in check_vf_mbox_random_id() 1420 func_to_func->workq = create_singlethread_workqueue(HINIC_MBOX_WQ_NAME); in hinic_func_to_func_init() 1421 if (!func_to_func->workq) { in hinic_func_to_func_init() 1465 destroy_workqueue(func_to_func->workq); in hinic_func_to_func_init() 1484 destroy_workqueue(func_to_func->workq); in hinic_func_to_func_free()
|
D | hinic_hw_mgmt.h | 144 struct workqueue_struct *workq; member
|
D | hinic_main.c | 817 queue_work(nic_dev->workq, &rx_mode_work->work); in hinic_set_rx_mode() 1227 nic_dev->workq = create_singlethread_workqueue(HINIC_WQ_NAME); in nic_dev_init() 1228 if (!nic_dev->workq) { in nic_dev_init() 1338 destroy_workqueue(nic_dev->workq); in nic_dev_init() 1465 destroy_workqueue(nic_dev->workq); in hinic_remove()
|
D | hinic_hw_mbox.h | 96 struct workqueue_struct *workq; member
|
D | hinic_hw_eqs.h | 213 struct workqueue_struct *workq; member
|
/drivers/crypto/ccree/ |
D | cc_request_mgr.c | 39 struct workqueue_struct *workq; member 104 flush_workqueue(req_mgr_h->workq); in cc_req_mgr_fini() 105 destroy_workqueue(req_mgr_h->workq); in cc_req_mgr_fini() 134 req_mgr_h->workq = create_singlethread_workqueue("ccree"); in cc_req_mgr_init() 135 if (!req_mgr_h->workq) { in cc_req_mgr_init() 539 queue_delayed_work(request_mgr_handle->workq, in complete_request()
|
/drivers/net/ethernet/netronome/nfp/ |
D | ccm_mbox.c | 140 queue_work(nn->mbox_cmsg.workq, &nn->mbox_cmsg.runq_work); in nfp_ccm_mbox_mark_next_runner() 671 queue_work(nn->mbox_cmsg.workq, in nfp_ccm_mbox_post() 723 drain_workqueue(nn->mbox_cmsg.workq); in nfp_ccm_mbox_clean() 733 nn->mbox_cmsg.workq = alloc_workqueue("nfp-ccm-mbox", WQ_UNBOUND, 0); in nfp_ccm_mbox_alloc() 734 if (!nn->mbox_cmsg.workq) in nfp_ccm_mbox_alloc() 741 destroy_workqueue(nn->mbox_cmsg.workq); in nfp_ccm_mbox_free()
|
/drivers/gpu/drm/msm/hdmi/ |
D | hdmi.c | 66 if (hdmi->workq) { in msm_hdmi_destroy() 67 flush_workqueue(hdmi->workq); in msm_hdmi_destroy() 68 destroy_workqueue(hdmi->workq); in msm_hdmi_destroy() 266 hdmi->workq = alloc_ordered_workqueue("msm_hdmi", 0); in msm_hdmi_init() 267 if (!hdmi->workq) { in msm_hdmi_init()
|
D | hdmi_hdcp.c | 232 queue_work(hdmi->workq, &hdcp_ctrl->hdcp_reauth_work); in msm_hdmi_hdcp_irq() 452 queue_work(hdmi->workq, &hdcp_ctrl->hdcp_auth_work); in msm_hdmi_hdcp_reauth_work() 552 queue_work(hdmi->workq, &hdcp_ctrl->hdcp_reauth_work); in msm_hdmi_hdcp_auth_fail() 1326 queue_work(hdmi->workq, &hdcp_ctrl->hdcp_auth_work); in msm_hdmi_hdcp_on()
|
D | hdmi.h | 71 struct workqueue_struct *workq; member
|
D | hdmi_hpd.c | 206 queue_work(hdmi->workq, &hdmi_bridge->hpd_work); in msm_hdmi_hpd_irq()
|
/drivers/iio/adc/ |
D | at91-sama5d2_adc.c | 384 struct work_struct workq; member 1246 schedule_work(&st->touch_st.workq); in at91_adc_touch_data_handler() 1277 static void at91_adc_workq_handler(struct work_struct *workq) in at91_adc_workq_handler() argument 1279 struct at91_adc_touch *touch_st = container_of(workq, in at91_adc_workq_handler() 1280 struct at91_adc_touch, workq); in at91_adc_workq_handler() 1783 INIT_WORK(&st->touch_st.workq, at91_adc_workq_handler); in at91_adc_probe()
|
/drivers/scsi/ |
D | libiscsi.c | 91 if (ihost->workq) in iscsi_conn_queue_work() 92 queue_work(ihost->workq, &conn->xmitwork); in iscsi_conn_queue_work() 739 if (!ihost->workq) { in __iscsi_conn_send_pdu() 1729 if (!ihost->workq) { in iscsi_queuecommand() 1910 if (ihost->workq) in iscsi_suspend_tx() 1911 flush_workqueue(ihost->workq); in iscsi_suspend_tx() 2677 ihost->workq = alloc_workqueue("%s", in iscsi_host_alloc() 2680 if (!ihost->workq) in iscsi_host_alloc() 2724 if (ihost->workq) in iscsi_host_remove() 2725 destroy_workqueue(ihost->workq); in iscsi_host_remove()
|
/drivers/infiniband/hw/cxgb4/ |
D | cm.c | 138 static struct workqueue_struct *workq; variable 4340 queue_work(workq, &skb_work); in ep_timeout() 4358 queue_work(workq, &skb_work); in sched() 4461 workq = alloc_ordered_workqueue("iw_cxgb4", WQ_MEM_RECLAIM); in c4iw_cm_init() 4462 if (!workq) in c4iw_cm_init() 4471 flush_workqueue(workq); in c4iw_cm_term() 4472 destroy_workqueue(workq); in c4iw_cm_term()
|