/net/ceph/crush/ |
D | mapper.c | 75 struct crush_work_bucket *work, in bucket_perm_choose() argument 82 if (work->perm_x != (__u32)x || work->perm_n == 0) { in bucket_perm_choose() 84 work->perm_x = x; in bucket_perm_choose() 90 work->perm[0] = s; in bucket_perm_choose() 91 work->perm_n = 0xffff; /* magic value, see below */ in bucket_perm_choose() 96 work->perm[i] = i; in bucket_perm_choose() 97 work->perm_n = 0; in bucket_perm_choose() 98 } else if (work->perm_n == 0xffff) { in bucket_perm_choose() 101 work->perm[i] = i; in bucket_perm_choose() 102 work->perm[work->perm[0]] = 0; in bucket_perm_choose() [all …]
|
/net/rds/ |
D | threads.c | 162 void rds_connect_worker(struct work_struct *work) in rds_connect_worker() argument 164 struct rds_conn_path *cp = container_of(work, in rds_connect_worker() 166 cp_conn_w.work); in rds_connect_worker() 191 void rds_send_worker(struct work_struct *work) in rds_send_worker() argument 193 struct rds_conn_path *cp = container_of(work, in rds_send_worker() 195 cp_send_w.work); in rds_send_worker() 218 void rds_recv_worker(struct work_struct *work) in rds_recv_worker() argument 220 struct rds_conn_path *cp = container_of(work, in rds_recv_worker() 222 cp_recv_w.work); in rds_recv_worker() 243 void rds_shutdown_worker(struct work_struct *work) in rds_shutdown_worker() argument [all …]
|
D | ib_rdma.c | 44 struct work_struct work; member 48 static void rds_ib_odp_mr_worker(struct work_struct *work); 477 static void rds_ib_mr_pool_flush_worker(struct work_struct *work) in rds_ib_mr_pool_flush_worker() argument 479 struct rds_ib_mr_pool *pool = container_of(work, struct rds_ib_mr_pool, flush_worker.work); in rds_ib_mr_pool_flush_worker() 497 INIT_DELAYED_WORK(&ibmr->work, rds_ib_odp_mr_worker); in rds_ib_free_mr() 498 queue_delayed_work(rds_ib_mr_wq, &ibmr->work, 0); in rds_ib_free_mr() 694 static void rds_ib_odp_mr_worker(struct work_struct *work) in rds_ib_odp_mr_worker() argument 698 ibmr = container_of(work, struct rds_ib_mr, work.work); in rds_ib_odp_mr_worker()
|
/net/netfilter/ |
D | xt_IDLETIMER.c | 41 struct work_struct work; member 87 schedule_work(&timer->work); in check_for_delayed_trigger() 200 static void idletimer_tg_work(struct work_struct *work) in idletimer_tg_work() argument 202 struct idletimer_tg *timer = container_of(work, struct idletimer_tg, in idletimer_tg_work() 203 work); in idletimer_tg_work() 220 schedule_work(&timer->work); in idletimer_tg_expired() 264 schedule_work(&timer->work); in idletimer_resume() 281 schedule_work(&timer->work); in idletimer_tg_alarmproc() 351 INIT_WORK(&info->timer->work, idletimer_tg_work); in idletimer_tg_create() 420 INIT_WORK(&info->timer->work, idletimer_tg_work); in idletimer_tg_create_v1() [all …]
|
D | nf_nat_masquerade.c | 13 struct work_struct work; member 77 static void iterate_cleanup_work(struct work_struct *work) in iterate_cleanup_work() argument 81 w = container_of(work, struct masq_dev_work, work); in iterate_cleanup_work() 119 INIT_WORK(&w->work, iterate_cleanup_work); in nf_nat_masq_schedule() 125 schedule_work(&w->work); in nf_nat_masq_schedule()
|
/net/strparser/ |
D | strparser.c | 379 queue_work(strp_wq, &strp->work); in strp_data_ready() 389 queue_work(strp_wq, &strp->work); in strp_data_ready() 407 queue_work(strp_wq, &strp->work); in do_strp_work() 415 do_strp_work(container_of(w, struct strparser, work)); in strp_work() 421 msg_timer_work.work); in strp_msg_timeout() 475 INIT_WORK(&strp->work, strp_work); in strp_init() 501 queue_work(strp_wq, &strp->work); in strp_unpause() 513 cancel_work_sync(&strp->work); in strp_done() 530 queue_work(strp_wq, &strp->work); in strp_check_rcv()
|
/net/core/ |
D | sock_diag.c | 111 struct work_struct work; member 121 static void sock_diag_broadcast_destroy_work(struct work_struct *work) in sock_diag_broadcast_destroy_work() argument 124 container_of(work, struct broadcast_sk, work); in sock_diag_broadcast_destroy_work() 161 INIT_WORK(&bsk->work, sock_diag_broadcast_destroy_work); in sock_diag_broadcast_destroy() 162 queue_work(broadcast_wq, &bsk->work); in sock_diag_broadcast_destroy()
|
D | drop_monitor.c | 116 int work, int budget); 117 void (*work_item_func)(struct work_struct *work); 118 void (*hw_work_item_func)(struct work_struct *work); 190 static void send_dm_alert(struct work_struct *work) in send_dm_alert() argument 195 data = container_of(work, struct per_cpu_dm_data, dm_alert_work); in send_dm_alert() 274 int work, int budget) in trace_napi_poll_hit() argument 414 static void net_dm_hw_summary_work(struct work_struct *work) in net_dm_hw_summary_work() argument 421 hw_data = container_of(work, struct per_cpu_dm_data, dm_alert_work); in net_dm_hw_summary_work() 544 int work, int budget) in net_dm_packet_trace_napi_poll_hit() argument 702 static void net_dm_packet_work(struct work_struct *work) in net_dm_packet_work() argument [all …]
|
/net/xdp/ |
D | xdp_umem.c | 69 static void xdp_umem_release_deferred(struct work_struct *work) in xdp_umem_release_deferred() argument 71 struct xdp_umem *umem = container_of(work, struct xdp_umem, work); in xdp_umem_release_deferred() 88 INIT_WORK(&umem->work, xdp_umem_release_deferred); in xdp_put_umem() 89 schedule_work(&umem->work); in xdp_put_umem()
|
/net/ceph/ |
D | osdmap.c | 976 struct crush_work *work; in alloc_workspace() local 983 work = ceph_kvmalloc(work_size, GFP_NOIO); in alloc_workspace() 984 if (!work) in alloc_workspace() 987 INIT_LIST_HEAD(&work->item); in alloc_workspace() 988 crush_init_workspace(c, work); in alloc_workspace() 989 return work; in alloc_workspace() 992 static void free_workspace(struct crush_work *work) in free_workspace() argument 994 WARN_ON(!list_empty(&work->item)); in free_workspace() 995 kvfree(work); in free_workspace() 1008 struct crush_work *work) in add_initial_workspace() argument [all …]
|
/net/tipc/ |
D | topsrv.c | 113 static void tipc_conn_recv_work(struct work_struct *work); 114 static void tipc_conn_send_work(struct work_struct *work); 300 static void tipc_conn_send_work(struct work_struct *work) in tipc_conn_send_work() argument 302 struct tipc_conn *con = container_of(work, struct tipc_conn, swork); in tipc_conn_send_work() 417 static void tipc_conn_recv_work(struct work_struct *work) in tipc_conn_recv_work() argument 419 struct tipc_conn *con = container_of(work, struct tipc_conn, rwork); in tipc_conn_recv_work() 452 static void tipc_topsrv_accept(struct work_struct *work) in tipc_topsrv_accept() argument 454 struct tipc_topsrv *srv = container_of(work, struct tipc_topsrv, awork); in tipc_topsrv_accept()
|
D | crypto.c | 213 struct delayed_work work; member 306 static void tipc_crypto_work_tx(struct work_struct *work); 307 static void tipc_crypto_work_rx(struct work_struct *work); 1223 if (cancel_delayed_work(&rx->work)) { in tipc_crypto_key_flush() 1422 if (queue_delayed_work(tx->wq, &rx->work, delay)) in tipc_crypto_key_synch() 1516 INIT_DELAYED_WORK(&c->work, tipc_crypto_work_rx); in tipc_crypto_start() 1518 INIT_DELAYED_WORK(&c->work, tipc_crypto_work_tx); in tipc_crypto_start() 1535 cancel_delayed_work_sync(&c->work); in tipc_crypto_stop() 2344 if (likely(skey && queue_delayed_work(tx->wq, &rx->work, 0))) in tipc_crypto_key_rcv() 2357 static void tipc_crypto_work_rx(struct work_struct *work) in tipc_crypto_work_rx() argument [all …]
|
/net/9p/ |
D | trans_xen.c | 70 struct work_struct work; member 194 static void p9_xen_response(struct work_struct *work) in p9_xen_response() argument 203 ring = container_of(work, struct xen_9pfs_dataring, work); in p9_xen_response() 273 schedule_work(&ring->work); in xen_9pfs_front_event_handler() 305 cancel_work_sync(&ring->work); in xen_9pfs_front_free() 351 INIT_WORK(&ring->work, p9_xen_response); in xen_9pfs_front_alloc_dataring()
|
/net/wireless/ |
D | core.h | 418 void cfg80211_autodisconnect_wk(struct work_struct *work); 421 void cfg80211_conn_work(struct work_struct *work); 445 void cfg80211_sched_scan_results_wk(struct work_struct *work); 482 void cfg80211_dfs_channels_update_work(struct work_struct *work); 497 void cfg80211_background_cac_done_wk(struct work_struct *work); 499 void cfg80211_background_cac_abort_wk(struct work_struct *work); 566 void cfg80211_pmsr_free_wk(struct work_struct *work);
|
/net/bluetooth/ |
D | 6lowpan.c | 597 static void do_notify_peers(struct work_struct *work) in do_notify_peers() argument 599 struct lowpan_btle_dev *dev = container_of(work, struct lowpan_btle_dev, in do_notify_peers() 600 notify_peers.work); in do_notify_peers() 751 static void delete_netdev(struct work_struct *work) in delete_netdev() argument 753 struct lowpan_btle_dev *entry = container_of(work, in delete_netdev() 1027 struct work_struct work; member 1031 static void do_enable_set(struct work_struct *work) in do_enable_set() argument 1033 struct set_enable *set_enable = container_of(work, in do_enable_set() 1034 struct set_enable, work); in do_enable_set() 1065 INIT_WORK(&set_enable->work, do_enable_set); in lowpan_enable_set() [all …]
|
D | hci_request.c | 1976 static void adv_timeout_expire(struct work_struct *work) in adv_timeout_expire() argument 1978 struct hci_dev *hdev = container_of(work, struct hci_dev, in adv_timeout_expire() 1979 adv_instance_expire.work); in adv_timeout_expire() 2038 static void interleave_scan_work(struct work_struct *work) in interleave_scan_work() argument 2040 struct hci_dev *hdev = container_of(work, struct hci_dev, in interleave_scan_work() 2041 interleave_scan.work); in interleave_scan_work() 2680 static void scan_update_work(struct work_struct *work) in scan_update_work() argument 2682 struct hci_dev *hdev = container_of(work, struct hci_dev, scan_update); in scan_update_work() 2718 static void connectable_update_work(struct work_struct *work) in connectable_update_work() argument 2720 struct hci_dev *hdev = container_of(work, struct hci_dev, in connectable_update_work() [all …]
|
D | hci_conn.c | 145 static void le_scan_cleanup(struct work_struct *work) in le_scan_cleanup() argument 147 struct hci_conn *conn = container_of(work, struct hci_conn, in le_scan_cleanup() 435 static void hci_conn_timeout(struct work_struct *work) in hci_conn_timeout() argument 437 struct hci_conn *conn = container_of(work, struct hci_conn, in hci_conn_timeout() 438 disc_work.work); in hci_conn_timeout() 466 static void hci_conn_idle(struct work_struct *work) in hci_conn_idle() argument 468 struct hci_conn *conn = container_of(work, struct hci_conn, in hci_conn_idle() 469 idle_work.work); in hci_conn_idle() 500 static void hci_conn_auto_accept(struct work_struct *work) in hci_conn_auto_accept() argument 502 struct hci_conn *conn = container_of(work, struct hci_conn, in hci_conn_auto_accept() [all …]
|
/net/batman-adv/ |
D | bat_v_elp.c | 163 void batadv_v_elp_throughput_metric_update(struct work_struct *work) in batadv_v_elp_throughput_metric_update() argument 168 neigh_bat_v = container_of(work, struct batadv_hardif_neigh_node_bat_v, in batadv_v_elp_throughput_metric_update() 251 static void batadv_v_elp_periodic_work(struct work_struct *work) in batadv_v_elp_periodic_work() argument 262 bat_v = container_of(work, struct batadv_hard_iface_bat_v, elp_wq.work); in batadv_v_elp_periodic_work()
|
/net/vmw_vsock/ |
D | virtio_transport.c | 88 virtio_transport_send_pkt_work(struct work_struct *work) in virtio_transport_send_pkt_work() argument 91 container_of(work, struct virtio_vsock, send_pkt_work); in virtio_transport_send_pkt_work() 288 static void virtio_transport_tx_work(struct work_struct *work) in virtio_transport_tx_work() argument 291 container_of(work, struct virtio_vsock, tx_work); in virtio_transport_tx_work() 399 static void virtio_transport_event_work(struct work_struct *work) in virtio_transport_event_work() argument 402 container_of(work, struct virtio_vsock, event_work); in virtio_transport_event_work() 521 static void virtio_transport_rx_work(struct work_struct *work) in virtio_transport_rx_work() argument 524 container_of(work, struct virtio_vsock, rx_work); in virtio_transport_rx_work()
|
/net/smc/ |
D | smc_ism.c | 303 struct work_struct work; member 364 static void smc_ism_event_work(struct work_struct *work) in smc_ism_event_work() argument 367 container_of(work, struct smc_ism_event_work, work); in smc_ism_event_work() 506 INIT_WORK(&wrk->work, smc_ism_event_work); in smcd_handle_event() 509 queue_work(smcd->event_wq, &wrk->work); in smcd_handle_event()
|
/net/xfrm/ |
D | espintcp.c | 405 static void espintcp_tx_work(struct work_struct *work) in espintcp_tx_work() argument 407 struct espintcp_ctx *ctx = container_of(work, in espintcp_tx_work() 408 struct espintcp_ctx, work); in espintcp_tx_work() 421 schedule_work(&ctx->work); in espintcp_write_space() 490 INIT_WORK(&ctx->work, espintcp_tx_work); in espintcp_init_sk() 527 cancel_work_sync(&ctx->work); in espintcp_close()
|
/net/mac80211/ |
D | ht.c | 322 cancel_work_sync(&sta->ampdu_mlme.work); in ieee80211_sta_tear_down_BA_sessions() 339 void ieee80211_ba_session_work(struct work_struct *work) in ieee80211_ba_session_work() argument 342 container_of(work, struct sta_info, ampdu_mlme.work); in ieee80211_ba_session_work() 542 void ieee80211_request_smps_mgd_work(struct work_struct *work) in ieee80211_request_smps_mgd_work() argument 545 container_of(work, struct ieee80211_sub_if_data, in ieee80211_request_smps_mgd_work()
|
/net/openvswitch/ |
D | dp_notify.c | 34 void ovs_dp_notify_wq(struct work_struct *work) in ovs_dp_notify_wq() argument 36 struct ovs_net *ovs_net = container_of(work, struct ovs_net, dp_notify_work); in ovs_dp_notify_wq()
|
/net/dsa/ |
D | tag_ocelot_8021q.c | 26 kthread_init_work(&xmit_work->work, felix_port->xmit_work_fn); in ocelot_defer_xmit() 33 kthread_queue_work(felix_port->xmit_worker, &xmit_work->work); in ocelot_defer_xmit()
|
/net/nfc/ |
D | digital_core.c | 110 static void digital_wq_cmd_complete(struct work_struct *work) in digital_wq_cmd_complete() argument 113 struct nfc_digital_dev *ddev = container_of(work, in digital_wq_cmd_complete() 152 static void digital_wq_cmd(struct work_struct *work) in digital_wq_cmd() argument 157 struct nfc_digital_dev *ddev = container_of(work, in digital_wq_cmd() 427 static void digital_wq_poll(struct work_struct *work) in digital_wq_poll() argument 431 struct nfc_digital_dev *ddev = container_of(work, in digital_wq_poll() 433 poll_work.work); in digital_wq_poll()
|