Home
last modified time | relevance | path

Searched refs:workq (Results 1 – 25 of 35) sorted by relevance

12

/drivers/isdn/mISDN/
Dhwchannel.c16 struct dchannel *dch = container_of(ws, struct dchannel, workq); in dchannel_bh()
39 struct bchannel *bch = container_of(ws, struct bchannel, workq); in bchannel_bh()
69 INIT_WORK(&ch->workq, dchannel_bh); in mISDN_initdchannel()
92 INIT_WORK(&ch->workq, bchannel_bh); in mISDN_initbchannel()
110 flush_work(&ch->workq); in mISDN_freedchannel()
150 cancel_work_sync(&ch->workq); in mISDN_freebchannel()
Dl1oip.h68 struct work_struct workq; member
Ddsp_core.c681 schedule_work(&dsp->workq); in dsp_function()
888 schedule_work(&dsp->workq); in dsp_function()
975 cancel_work_sync(&dsp->workq); in dsp_ctrl()
1007 struct dsp *dsp = container_of(work, struct dsp, workq); in dsp_send_bh()
1065 INIT_WORK(&ndsp->workq, (void *)dsp_send_bh); in dspcreate()
Dstack.c31 wake_up_interruptible(&st->workq); in _queue_message()
280 wait_event_interruptible(st->workq, (st->status & in mISDNStackd()
378 init_waitqueue_head(&newst->workq); in create_stack()
638 wake_up_interruptible(&st->workq); in delete_stack()
Ddsp.h185 struct work_struct workq; member
Dl1oip_core.c807 struct l1oip *hc = container_of(work, struct l1oip, workq); in l1oip_send_bh()
826 schedule_work(&hc->workq); in l1oip_keepalive()
1242 cancel_work_sync(&hc->workq); in release_card()
1488 INIT_WORK(&hc->workq, (void *)l1oip_send_bh); in l1oip_init()
Ddsp_cmx.c1585 schedule_work(&dsp->workq);
1618 schedule_work(&dsp->workq);
1940 schedule_work(&dsp->workq);
1956 schedule_work(&member->dsp->workq);
/drivers/hwmon/
Dxgene-hwmon.c105 struct work_struct workq; member
442 ctx = container_of(work, struct xgene_hwmon_dev, workq); in xgene_hwmon_evt_work()
522 schedule_work(&ctx->workq); in xgene_hwmon_rx_cb()
591 schedule_work(&ctx->workq); in xgene_hwmon_pcc_rx_cb()
637 INIT_WORK(&ctx->workq, xgene_hwmon_evt_work); in xgene_hwmon_probe()
750 schedule_work(&ctx->workq); in xgene_hwmon_probe()
771 cancel_work_sync(&ctx->workq); in xgene_hwmon_remove()
/drivers/net/ethernet/huawei/hinic/
Dhinic_hw_mgmt.c469 queue_work(pf_to_mgmt->workq, &mgmt_work->work); in mgmt_recv_msg_handler()
639 pf_to_mgmt->workq = create_singlethread_workqueue("hinic_mgmt"); in hinic_pf_to_mgmt_init()
640 if (!pf_to_mgmt->workq) { in hinic_pf_to_mgmt_init()
650 destroy_workqueue(pf_to_mgmt->workq); in hinic_pf_to_mgmt_init()
658 destroy_workqueue(pf_to_mgmt->workq); in hinic_pf_to_mgmt_init()
683 destroy_workqueue(pf_to_mgmt->workq); in hinic_pf_to_mgmt_free()
Dhinic_hw_eqs.c399 queue_work(aeqs->workq, &aeq_work->work); in aeq_interrupt()
867 aeqs->workq = create_singlethread_workqueue(HINIC_EQS_WQ_NAME); in hinic_aeqs_init()
868 if (!aeqs->workq) in hinic_aeqs_init()
889 destroy_workqueue(aeqs->workq); in hinic_aeqs_init()
904 destroy_workqueue(aeqs->workq); in hinic_aeqs_free()
Dhinic_dev.h91 struct workqueue_struct *workq; member
Dhinic_hw_mbox.c505 queue_work(func_to_func->workq, &mbox_work->work); in recv_mbox_handler()
619 queue_work(func_to_func->workq, &mbox_work->work); in check_vf_mbox_random_id()
1420 func_to_func->workq = create_singlethread_workqueue(HINIC_MBOX_WQ_NAME); in hinic_func_to_func_init()
1421 if (!func_to_func->workq) { in hinic_func_to_func_init()
1465 destroy_workqueue(func_to_func->workq); in hinic_func_to_func_init()
1484 destroy_workqueue(func_to_func->workq); in hinic_func_to_func_free()
Dhinic_hw_mgmt.h144 struct workqueue_struct *workq; member
Dhinic_main.c817 queue_work(nic_dev->workq, &rx_mode_work->work); in hinic_set_rx_mode()
1227 nic_dev->workq = create_singlethread_workqueue(HINIC_WQ_NAME); in nic_dev_init()
1228 if (!nic_dev->workq) { in nic_dev_init()
1338 destroy_workqueue(nic_dev->workq); in nic_dev_init()
1465 destroy_workqueue(nic_dev->workq); in hinic_remove()
Dhinic_hw_mbox.h96 struct workqueue_struct *workq; member
Dhinic_hw_eqs.h213 struct workqueue_struct *workq; member
/drivers/crypto/ccree/
Dcc_request_mgr.c39 struct workqueue_struct *workq; member
104 flush_workqueue(req_mgr_h->workq); in cc_req_mgr_fini()
105 destroy_workqueue(req_mgr_h->workq); in cc_req_mgr_fini()
134 req_mgr_h->workq = create_singlethread_workqueue("ccree"); in cc_req_mgr_init()
135 if (!req_mgr_h->workq) { in cc_req_mgr_init()
539 queue_delayed_work(request_mgr_handle->workq, in complete_request()
/drivers/net/ethernet/netronome/nfp/
Dccm_mbox.c140 queue_work(nn->mbox_cmsg.workq, &nn->mbox_cmsg.runq_work); in nfp_ccm_mbox_mark_next_runner()
671 queue_work(nn->mbox_cmsg.workq, in nfp_ccm_mbox_post()
723 drain_workqueue(nn->mbox_cmsg.workq); in nfp_ccm_mbox_clean()
733 nn->mbox_cmsg.workq = alloc_workqueue("nfp-ccm-mbox", WQ_UNBOUND, 0); in nfp_ccm_mbox_alloc()
734 if (!nn->mbox_cmsg.workq) in nfp_ccm_mbox_alloc()
741 destroy_workqueue(nn->mbox_cmsg.workq); in nfp_ccm_mbox_free()
/drivers/gpu/drm/msm/hdmi/
Dhdmi.c66 if (hdmi->workq) { in msm_hdmi_destroy()
67 flush_workqueue(hdmi->workq); in msm_hdmi_destroy()
68 destroy_workqueue(hdmi->workq); in msm_hdmi_destroy()
266 hdmi->workq = alloc_ordered_workqueue("msm_hdmi", 0); in msm_hdmi_init()
267 if (!hdmi->workq) { in msm_hdmi_init()
Dhdmi_hdcp.c232 queue_work(hdmi->workq, &hdcp_ctrl->hdcp_reauth_work); in msm_hdmi_hdcp_irq()
452 queue_work(hdmi->workq, &hdcp_ctrl->hdcp_auth_work); in msm_hdmi_hdcp_reauth_work()
552 queue_work(hdmi->workq, &hdcp_ctrl->hdcp_reauth_work); in msm_hdmi_hdcp_auth_fail()
1326 queue_work(hdmi->workq, &hdcp_ctrl->hdcp_auth_work); in msm_hdmi_hdcp_on()
Dhdmi.h71 struct workqueue_struct *workq; member
Dhdmi_hpd.c206 queue_work(hdmi->workq, &hdmi_bridge->hpd_work); in msm_hdmi_hpd_irq()
/drivers/iio/adc/
Dat91-sama5d2_adc.c384 struct work_struct workq; member
1246 schedule_work(&st->touch_st.workq); in at91_adc_touch_data_handler()
1277 static void at91_adc_workq_handler(struct work_struct *workq) in at91_adc_workq_handler() argument
1279 struct at91_adc_touch *touch_st = container_of(workq, in at91_adc_workq_handler()
1280 struct at91_adc_touch, workq); in at91_adc_workq_handler()
1783 INIT_WORK(&st->touch_st.workq, at91_adc_workq_handler); in at91_adc_probe()
/drivers/scsi/
Dlibiscsi.c91 if (ihost->workq) in iscsi_conn_queue_work()
92 queue_work(ihost->workq, &conn->xmitwork); in iscsi_conn_queue_work()
739 if (!ihost->workq) { in __iscsi_conn_send_pdu()
1729 if (!ihost->workq) { in iscsi_queuecommand()
1910 if (ihost->workq) in iscsi_suspend_tx()
1911 flush_workqueue(ihost->workq); in iscsi_suspend_tx()
2677 ihost->workq = alloc_workqueue("%s", in iscsi_host_alloc()
2680 if (!ihost->workq) in iscsi_host_alloc()
2724 if (ihost->workq) in iscsi_host_remove()
2725 destroy_workqueue(ihost->workq); in iscsi_host_remove()
/drivers/infiniband/hw/cxgb4/
Dcm.c138 static struct workqueue_struct *workq; variable
4340 queue_work(workq, &skb_work); in ep_timeout()
4358 queue_work(workq, &skb_work); in sched()
4461 workq = alloc_ordered_workqueue("iw_cxgb4", WQ_MEM_RECLAIM); in c4iw_cm_init()
4462 if (!workq) in c4iw_cm_init()
4471 flush_workqueue(workq); in c4iw_cm_term()
4472 destroy_workqueue(workq); in c4iw_cm_term()

12