/kernel/linux/linux-5.10/drivers/infiniband/hw/cxgb4/ |
D | mem.c | 404 return xa_insert_irq(&mhp->rhp->mrs, mmid, mhp, GFP_KERNEL); in finish_mem_reg() 407 static int register_mem(struct c4iw_dev *rhp, struct c4iw_pd *php, in register_mem() argument 413 ret = write_tpt_entry(&rhp->rdev, 0, &stag, 1, mhp->attr.pdid, in register_mem() 426 dereg_mem(&rhp->rdev, mhp->attr.stag, mhp->attr.pbl_size, in register_mem() 435 mhp->attr.pbl_addr = c4iw_pblpool_alloc(&mhp->rhp->rdev, in alloc_pbl() 448 struct c4iw_dev *rhp; in c4iw_get_dma_mr() local 456 rhp = php->rhp; in c4iw_get_dma_mr() 474 mhp->rhp = rhp; in c4iw_get_dma_mr() 484 ret = write_tpt_entry(&rhp->rdev, 0, &stag, 1, php->pdid, in c4iw_get_dma_mr() 496 dereg_mem(&rhp->rdev, mhp->attr.stag, mhp->attr.pbl_size, in c4iw_get_dma_mr() [all …]
|
D | provider.c | 64 struct c4iw_dev *rhp; in c4iw_dealloc_ucontext() local 68 rhp = to_c4iw_dev(ucontext->ibucontext.device); in c4iw_dealloc_ucontext() 72 c4iw_release_dev_ucontext(&rhp->rdev, &ucontext->uctx); in c4iw_dealloc_ucontext() 80 struct c4iw_dev *rhp = to_c4iw_dev(ibdev); in c4iw_alloc_ucontext() local 86 c4iw_init_dev_ucontext(&rhp->rdev, &context->uctx); in c4iw_alloc_ucontext() 92 rhp->rdev.flags |= T4_STATUS_PAGE_DISABLED; in c4iw_alloc_ucontext() 113 mm->addr = virt_to_phys(rhp->rdev.status_page); in c4iw_alloc_ucontext() 195 struct c4iw_dev *rhp; in c4iw_deallocate_pd() local 199 rhp = php->rhp; in c4iw_deallocate_pd() 201 c4iw_put_resource(&rhp->rdev.resource.pdid_table, php->pdid); in c4iw_deallocate_pd() [all …]
|
D | qp.c | 720 cxgb4_read_sge_timestamp(qhp->rhp->rdev.lldi.ports[0]); in post_write_cmpl() 744 cxgb4_read_sge_timestamp(qhp->rhp->rdev.lldi.ports[0]); in post_write_cmpl() 807 PBL_OFF(&mhp->rhp->rdev, mhp->attr.pbl_addr)>>3)); in build_tpte_memreg() 916 xa_lock_irqsave(&qhp->rhp->qps, flags); in ring_kernel_sq_db() 918 if (qhp->rhp->db_state == NORMAL) in ring_kernel_sq_db() 921 add_to_fc_list(&qhp->rhp->db_fc_list, &qhp->db_fc_entry); in ring_kernel_sq_db() 925 xa_unlock_irqrestore(&qhp->rhp->qps, flags); in ring_kernel_sq_db() 933 xa_lock_irqsave(&qhp->rhp->qps, flags); in ring_kernel_rq_db() 935 if (qhp->rhp->db_state == NORMAL) in ring_kernel_rq_db() 938 add_to_fc_list(&qhp->rhp->db_fc_list, &qhp->db_fc_entry); in ring_kernel_rq_db() [all …]
|
D | cq.c | 351 qhp = get_qhp(chp->rhp, CQE_QPID(hw_cqe)); in c4iw_flush_hw_cq() 803 c4iw_invalidate_mr(qhp->rhp, wc->ex.invalidate_rkey); in __c4iw_poll_cq_one() 844 c4iw_invalidate_mr(qhp->rhp, in __c4iw_poll_cq_one() 933 qhp = get_qhp(chp->rhp, CQE_QPID(rd_cqe)); in c4iw_poll_cq_one() 978 xa_erase_irq(&chp->rhp->cqs, chp->cq.cqid); in c4iw_destroy_cq() 984 destroy_cq(&chp->rhp->rdev, &chp->cq, in c4iw_destroy_cq() 997 struct c4iw_dev *rhp = to_c4iw_dev(ibcq->device); in c4iw_create_cq() local 1014 if (vector >= rhp->rdev.lldi.nciq) in c4iw_create_cq() 1050 hwentries = min(entries * 2, rhp->rdev.hw_queue.t4_max_iq_size); in c4iw_create_cq() 1072 ret = create_cq(&rhp->rdev, &chp->cq, in c4iw_create_cq() [all …]
|
D | iw_cxgb4.h | 349 static inline struct c4iw_cq *get_chp(struct c4iw_dev *rhp, u32 cqid) in get_chp() argument 351 return xa_load(&rhp->cqs, cqid); in get_chp() 354 static inline struct c4iw_qp *get_qhp(struct c4iw_dev *rhp, u32 qpid) in get_qhp() argument 356 return xa_load(&rhp->qps, qpid); in get_qhp() 369 struct c4iw_dev *rhp; member 398 struct c4iw_dev *rhp; member 416 struct c4iw_dev *rhp; member 430 struct c4iw_dev *rhp; member 487 struct c4iw_dev *rhp; member 510 struct c4iw_dev *rhp; member [all …]
|
D | device.c | 1376 xa_lock_irq(&qp->rhp->qps); in recover_lost_dbs() 1378 ret = cxgb4_sync_txq_pidx(qp->rhp->rdev.lldi.ports[0], in recover_lost_dbs() 1386 xa_unlock_irq(&qp->rhp->qps); in recover_lost_dbs() 1391 ret = cxgb4_sync_txq_pidx(qp->rhp->rdev.lldi.ports[0], in recover_lost_dbs() 1400 xa_unlock_irq(&qp->rhp->qps); in recover_lost_dbs() 1405 xa_unlock_irq(&qp->rhp->qps); in recover_lost_dbs() 1408 while (cxgb4_dbfifo_count(qp->rhp->rdev.lldi.ports[0], 1) > 0) { in recover_lost_dbs()
|
/kernel/linux/linux-5.10/include/trace/events/ |
D | rcu.h | 485 TP_PROTO(const char *rcuname, struct rcu_head *rhp, long qlen), 487 TP_ARGS(rcuname, rhp, qlen), 491 __field(void *, rhp) 498 __entry->rhp = rhp; 499 __entry->func = rhp->func; 504 __entry->rcuname, __entry->rhp, __entry->func, 518 TP_PROTO(const char *rcuname, struct rcu_head *rhp, unsigned long offset, 521 TP_ARGS(rcuname, rhp, offset, qlen), 525 __field(void *, rhp) 532 __entry->rhp = rhp; [all …]
|
/kernel/linux/linux-5.10/kernel/rcu/ |
D | rcu_segcblist.c | 28 void rcu_cblist_enqueue(struct rcu_cblist *rclp, struct rcu_head *rhp) in rcu_cblist_enqueue() argument 30 *rclp->tail = rhp; in rcu_cblist_enqueue() 31 rclp->tail = &rhp->next; in rcu_cblist_enqueue() 45 struct rcu_head *rhp) in rcu_cblist_flush_enqueue() argument 53 if (!rhp) { in rcu_cblist_flush_enqueue() 56 rhp->next = NULL; in rcu_cblist_flush_enqueue() 57 srclp->head = rhp; in rcu_cblist_flush_enqueue() 58 srclp->tail = &rhp->next; in rcu_cblist_flush_enqueue() 69 struct rcu_head *rhp; in rcu_cblist_dequeue() local 71 rhp = rclp->head; in rcu_cblist_dequeue() [all …]
|
D | srcutiny.c | 116 struct rcu_head *rhp; in srcu_drive_gp() local 139 rhp = lh; in srcu_drive_gp() 142 rhp->func(rhp); in srcu_drive_gp() 178 void call_srcu(struct srcu_struct *ssp, struct rcu_head *rhp, in call_srcu() argument 183 rhp->func = func; in call_srcu() 184 rhp->next = NULL; in call_srcu() 186 *ssp->srcu_cb_tail = rhp; in call_srcu() 187 ssp->srcu_cb_tail = &rhp->next; in call_srcu()
|
D | update.c | 467 void do_trace_rcu_torture_read(const char *rcutorturename, struct rcu_head *rhp, in do_trace_rcu_torture_read() argument 471 trace_rcu_torture_read(rcutorturename, rhp, secs, c_old, c); in do_trace_rcu_torture_read() 475 #define do_trace_rcu_torture_read(rcutorturename, rhp, secs, c_old, c) \ argument 534 struct early_boot_kfree_rcu *rhp; in early_boot_test_call_rcu() local 539 rhp = kmalloc(sizeof(*rhp), GFP_KERNEL); in early_boot_test_call_rcu() 540 if (!WARN_ON_ONCE(!rhp)) in early_boot_test_call_rcu() 541 kfree_rcu(rhp, rh); in early_boot_test_call_rcu()
|
D | rcu_segcblist.h | 19 void rcu_cblist_enqueue(struct rcu_cblist *rclp, struct rcu_head *rhp); 22 struct rcu_head *rhp); 88 struct rcu_head *rhp); 90 struct rcu_head *rhp);
|
D | tasks.h | 152 static void call_rcu_tasks_generic(struct rcu_head *rhp, rcu_callback_t func, in call_rcu_tasks_generic() argument 158 rhp->next = NULL; in call_rcu_tasks_generic() 159 rhp->func = func; in call_rcu_tasks_generic() 162 WRITE_ONCE(*rtp->cbs_tail, rhp); in call_rcu_tasks_generic() 163 rtp->cbs_tail = &rhp->next; in call_rcu_tasks_generic() 495 void call_rcu_tasks(struct rcu_head *rhp, rcu_callback_t func); 516 void call_rcu_tasks(struct rcu_head *rhp, rcu_callback_t func) in call_rcu_tasks() argument 518 call_rcu_tasks_generic(rhp, func, &rcu_tasks); in call_rcu_tasks() 630 void call_rcu_tasks_rude(struct rcu_head *rhp, rcu_callback_t func); 652 void call_rcu_tasks_rude(struct rcu_head *rhp, rcu_callback_t func) in call_rcu_tasks_rude() argument [all …]
|
D | srcutree.c | 809 static void srcu_leak_callback(struct rcu_head *rhp) in srcu_leak_callback() argument 817 struct rcu_head *rhp, bool do_norm) in srcu_gp_start_if_needed() argument 830 if (rhp) in srcu_gp_start_if_needed() 831 rcu_segcblist_enqueue(&sdp->srcu_cblist, rhp); in srcu_gp_start_if_needed() 881 static void __call_srcu(struct srcu_struct *ssp, struct rcu_head *rhp, in __call_srcu() argument 884 if (debug_rcu_head_queue(rhp)) { in __call_srcu() 886 WRITE_ONCE(rhp->func, srcu_leak_callback); in __call_srcu() 890 rhp->func = func; in __call_srcu() 891 (void)srcu_gp_start_if_needed(ssp, rhp, do_norm); in __call_srcu() 911 void call_srcu(struct srcu_struct *ssp, struct rcu_head *rhp, in call_srcu() argument [all …]
|
D | rcuscale.c | 348 static void rcu_scale_async_cb(struct rcu_head *rhp) in rcu_scale_async_cb() argument 351 kfree(rhp); in rcu_scale_async_cb() 363 struct rcu_head *rhp = NULL; in rcu_scale_writer() local 403 if (!rhp) in rcu_scale_writer() 404 rhp = kmalloc(sizeof(*rhp), GFP_KERNEL); in rcu_scale_writer() 405 if (rhp && atomic_read(this_cpu_ptr(&n_async_inflight)) < gp_async_max) { in rcu_scale_writer() 407 cur_ops->async(rhp, rcu_scale_async_cb); in rcu_scale_writer() 408 rhp = NULL; in rcu_scale_writer() 413 kfree(rhp); /* Because we are stopping. */ in rcu_scale_writer()
|
D | sync.c | 43 static void rcu_sync_func(struct rcu_head *rhp); 73 static void rcu_sync_func(struct rcu_head *rhp) in rcu_sync_func() argument 75 struct rcu_sync *rsp = container_of(rhp, struct rcu_sync, cb_head); in rcu_sync_func()
|
D | rcu.h | 458 struct rcu_head *rhp, 472 struct rcu_head *rhp, 477 #define do_trace_rcu_torture_read(rcutorturename, rhp, secs, c_old, c) \ argument
|
D | rcutorture.c | 1210 static void rcu_torture_timer_cb(struct rcu_head *rhp) in rcu_torture_timer_cb() argument 1212 kfree(rhp); in rcu_torture_timer_cb() 1459 struct rcu_head *rhp = kmalloc(sizeof(*rhp), GFP_NOWAIT); in rcu_torture_timer() local 1461 if (rhp) in rcu_torture_timer() 1462 cur_ops->call(rhp, rcu_torture_timer_cb); in rcu_torture_timer() 1787 static void rcu_torture_fwd_prog_cb(struct rcu_head *rhp) in rcu_torture_fwd_prog_cb() argument 1789 struct fwd_cb_state *fcsp = container_of(rhp, struct fwd_cb_state, rh); in rcu_torture_fwd_prog_cb() 1856 static void rcu_torture_fwd_cb_cr(struct rcu_head *rhp) in rcu_torture_fwd_cb_cr() argument 1860 struct rcu_fwd_cb *rfcp = container_of(rhp, struct rcu_fwd_cb, rh); in rcu_torture_fwd_cb_cr() 2214 struct rcu_head *rhp = rcu_void; in rcu_torture_barrier1cb() local [all …]
|
D | tree_plugin.h | 1689 static bool rcu_nocb_do_flush_bypass(struct rcu_data *rdp, struct rcu_head *rhp, in rcu_nocb_do_flush_bypass() argument 1697 if (rhp && !rcu_cblist_n_cbs(&rdp->nocb_bypass)) { in rcu_nocb_do_flush_bypass() 1702 if (rhp) in rcu_nocb_do_flush_bypass() 1704 rcu_cblist_flush_enqueue(&rcl, &rdp->nocb_bypass, rhp); in rcu_nocb_do_flush_bypass() 1719 static bool rcu_nocb_flush_bypass(struct rcu_data *rdp, struct rcu_head *rhp, in rcu_nocb_flush_bypass() argument 1726 return rcu_nocb_do_flush_bypass(rdp, rhp, j); in rcu_nocb_flush_bypass() 1760 static bool rcu_nocb_try_bypass(struct rcu_data *rdp, struct rcu_head *rhp, in rcu_nocb_try_bypass() argument 1816 if (!rcu_nocb_flush_bypass(rdp, rhp, j)) { in rcu_nocb_try_bypass() 1839 rcu_cblist_enqueue(&rdp->nocb_bypass, rhp); in rcu_nocb_try_bypass() 2504 static bool rcu_nocb_flush_bypass(struct rcu_data *rdp, struct rcu_head *rhp, in rcu_nocb_flush_bypass() argument [all …]
|
D | tree.h | 428 static bool rcu_nocb_flush_bypass(struct rcu_data *rdp, struct rcu_head *rhp, 430 static bool rcu_nocb_try_bypass(struct rcu_data *rdp, struct rcu_head *rhp,
|
D | tree.c | 2438 struct rcu_head *rhp; in rcu_do_batch() local 2481 rhp = rcu_cblist_dequeue(&rcl); in rcu_do_batch() 2482 for (; rhp; rhp = rcu_cblist_dequeue(&rcl)) { in rcu_do_batch() 2485 debug_rcu_head_unqueue(rhp); in rcu_do_batch() 2488 trace_rcu_invoke_callback(rcu_state.name, rhp); in rcu_do_batch() 2490 f = rhp->func; in rcu_do_batch() 2491 WRITE_ONCE(rhp->func, (rcu_callback_t)0L); in rcu_do_batch() 2492 f(rhp); in rcu_do_batch() 2903 static void rcu_leak_callback(struct rcu_head *rhp) in rcu_leak_callback() argument 3806 static void rcu_barrier_callback(struct rcu_head *rhp) in rcu_barrier_callback() argument
|
/kernel/linux/linux-5.10/include/linux/ |
D | rcupdate_trace.h | 86 void call_rcu_tasks_trace(struct rcu_head *rhp, rcu_callback_t func); 94 static inline void call_rcu_tasks_trace(struct rcu_head *rhp, rcu_callback_t func) { BUG(); } in call_rcu_tasks_trace() argument
|
D | rcupdate.h | 959 static inline void rcu_head_init(struct rcu_head *rhp) in rcu_head_init() argument 961 rhp->func = (rcu_callback_t)~0L; in rcu_head_init() 978 rcu_head_after_call_rcu(struct rcu_head *rhp, rcu_callback_t f) in rcu_head_after_call_rcu() argument 980 rcu_callback_t func = READ_ONCE(rhp->func); in rcu_head_after_call_rcu()
|
/kernel/linux/linux-5.10/drivers/media/usb/pvrusb2/ |
D | pvrusb2-v4l2.c | 42 struct pvr2_ioread *rhp; member 919 if (fhp->rhp) { in pvr2_v4l2_release() 922 sp = pvr2_ioread_get_stream(fhp->rhp); in pvr2_v4l2_release() 924 pvr2_ioread_destroy(fhp->rhp); in pvr2_v4l2_release() 925 fhp->rhp = NULL; in pvr2_v4l2_release() 1050 if (fh->rhp) return 0; in pvr2_v4l2_iosetup() 1066 fh->rhp = pvr2_channel_create_mpeg_stream(fh->pdi->stream); in pvr2_v4l2_iosetup() 1067 if (!fh->rhp) { in pvr2_v4l2_iosetup() 1077 return pvr2_ioread_set_enabled(fh->rhp,!0); in pvr2_v4l2_iosetup() 1120 if (!fh->rhp) { in pvr2_v4l2_read() [all …]
|
/kernel/linux/linux-5.10/tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/ |
D | misc.h | 33 #define trace_rcu_torture_read(rcutorturename, rhp, secs, c_old, c) \ argument
|
/kernel/linux/linux-5.10/kernel/ |
D | pid.c | 120 static void delayed_put_pid(struct rcu_head *rhp) in delayed_put_pid() argument 122 struct pid *pid = container_of(rhp, struct pid, rcu); in delayed_put_pid()
|