Home
last modified time | relevance | path

Searched refs:ctxt (Results 1 – 25 of 46) sorted by relevance

12

/drivers/staging/lustre/lustre/obdclass/
Dllog_obd.c46 struct llog_ctxt *ctxt; in llog_new_ctxt() local
48 ctxt = kzalloc(sizeof(*ctxt), GFP_NOFS); in llog_new_ctxt()
49 if (!ctxt) in llog_new_ctxt()
52 ctxt->loc_obd = obd; in llog_new_ctxt()
53 atomic_set(&ctxt->loc_refcount, 1); in llog_new_ctxt()
55 return ctxt; in llog_new_ctxt()
58 static void llog_ctxt_destroy(struct llog_ctxt *ctxt) in llog_ctxt_destroy() argument
60 if (ctxt->loc_exp) { in llog_ctxt_destroy()
61 class_export_put(ctxt->loc_exp); in llog_ctxt_destroy()
62 ctxt->loc_exp = NULL; in llog_ctxt_destroy()
[all …]
Dllog.c409 int llog_open(const struct lu_env *env, struct llog_ctxt *ctxt, in llog_open() argument
416 LASSERT(ctxt); in llog_open()
417 LASSERT(ctxt->loc_logops); in llog_open()
419 if (ctxt->loc_logops->lop_open == NULL) { in llog_open()
427 (*lgh)->lgh_ctxt = ctxt; in llog_open()
428 (*lgh)->lgh_logops = ctxt->loc_logops; in llog_open()
433 rc = ctxt->loc_logops->lop_open(env, *lgh, logid, name, open_param); in llog_open()
/drivers/staging/lustre/lustre/ptlrpc/
Dllog_net.c55 int llog_initiator_connect(struct llog_ctxt *ctxt) in llog_initiator_connect() argument
59 LASSERT(ctxt); in llog_initiator_connect()
60 new_imp = ctxt->loc_obd->u.cli.cl_import; in llog_initiator_connect()
61 LASSERTF(ctxt->loc_imp == NULL || ctxt->loc_imp == new_imp, in llog_initiator_connect()
62 "%p - %p\n", ctxt->loc_imp, new_imp); in llog_initiator_connect()
63 mutex_lock(&ctxt->loc_mutex); in llog_initiator_connect()
64 if (ctxt->loc_imp != new_imp) { in llog_initiator_connect()
65 if (ctxt->loc_imp) in llog_initiator_connect()
66 class_import_put(ctxt->loc_imp); in llog_initiator_connect()
67 ctxt->loc_imp = class_import_get(new_imp); in llog_initiator_connect()
[all …]
Dllog_client.c52 #define LLOG_CLIENT_ENTRY(ctxt, imp) do { \ argument
53 mutex_lock(&ctxt->loc_mutex); \
54 if (ctxt->loc_imp) { \
55 imp = class_import_get(ctxt->loc_imp); \
60 ctxt->loc_idx); \
62 mutex_unlock(&ctxt->loc_mutex); \
65 mutex_unlock(&ctxt->loc_mutex); \
68 #define LLOG_CLIENT_EXIT(ctxt, imp) do { \ argument
69 mutex_lock(&ctxt->loc_mutex); \
70 if (ctxt->loc_imp != imp) \
[all …]
/drivers/staging/lustre/lustre/include/
Dlustre_log.h93 int llog_open(const struct lu_env *env, struct llog_ctxt *ctxt,
98 struct llog_ctxt *ctxt, struct llog_ctxt *bak_ctxt,
144 int __llog_ctxt_put(const struct lu_env *env, struct llog_ctxt *ctxt);
148 int llog_initiator_connect(struct llog_ctxt *ctxt);
161 int (*lop_sync)(struct llog_ctxt *ctxt, struct obd_export *exp,
163 int (*lop_cleanup)(const struct lu_env *env, struct llog_ctxt *ctxt);
164 int (*lop_cancel)(const struct lu_env *env, struct llog_ctxt *ctxt,
166 int (*lop_connect)(struct llog_ctxt *ctxt, struct llog_logid *logid,
268 static inline struct llog_ctxt *llog_ctxt_get(struct llog_ctxt *ctxt) in llog_ctxt_get() argument
270 atomic_inc(&ctxt->loc_refcount); in llog_ctxt_get()
[all …]
/drivers/net/wireless/iwlwifi/mvm/
Dphy-ctxt.c127 static void iwl_mvm_phy_ctxt_cmd_hdr(struct iwl_mvm_phy_ctxt *ctxt, in iwl_mvm_phy_ctxt_cmd_hdr() argument
133 cmd->id_and_color = cpu_to_le32(FW_CMD_ID_AND_COLOR(ctxt->id, in iwl_mvm_phy_ctxt_cmd_hdr()
134 ctxt->color)); in iwl_mvm_phy_ctxt_cmd_hdr()
193 struct iwl_mvm_phy_ctxt *ctxt, in iwl_mvm_phy_ctxt_apply() argument
202 iwl_mvm_phy_ctxt_cmd_hdr(ctxt, &cmd, action, apply_time); in iwl_mvm_phy_ctxt_apply()
219 int iwl_mvm_phy_ctxt_add(struct iwl_mvm *mvm, struct iwl_mvm_phy_ctxt *ctxt, in iwl_mvm_phy_ctxt_add() argument
224 ctxt->ref); in iwl_mvm_phy_ctxt_add()
227 ctxt->channel = chandef->chan; in iwl_mvm_phy_ctxt_add()
229 return iwl_mvm_phy_ctxt_apply(mvm, ctxt, chandef, in iwl_mvm_phy_ctxt_add()
238 void iwl_mvm_phy_ctxt_ref(struct iwl_mvm *mvm, struct iwl_mvm_phy_ctxt *ctxt) in iwl_mvm_phy_ctxt_ref() argument
[all …]
DMakefile2 iwlmvm-y += fw.o mac80211.o nvm.o ops.o phy-ctxt.o mac-ctxt.o
/drivers/staging/rdma/hfi1/
Dtrace.h81 u32 ctxt,
87 TP_ARGS(dd, ctxt, eflags, etype, hlen, tlen, updegr, etail),
91 __field(u32, ctxt)
101 __entry->ctxt = ctxt;
111 __entry->ctxt,
122 TP_PROTO(struct hfi1_devdata *dd, u32 ctxt),
123 TP_ARGS(dd, ctxt),
126 __field(u32, ctxt)
132 __entry->ctxt = ctxt;
133 if (dd->rcd[ctxt]->do_interrupt ==
[all …]
Dfile_ops.c168 #define HFI1_MMAP_TOKEN(type, ctxt, subctxt, addr) \ argument
171 HFI1_MMAP_TOKEN_SET(CTXT, ctxt) | \
413 uctxt->ctxt); in hfi1_file_write()
457 ctxt_fp(kiocb->ki_filp)->ctxt, subctxt_fp(kiocb->ki_filp), in hfi1_write_iter()
493 u16 ctxt; in hfi1_file_mmap() local
502 ctxt = HFI1_MMAP_TOKEN_GET(CTXT, token); in hfi1_file_mmap()
505 if (ctxt != uctxt->ctxt || subctxt != subctxt_fp(fp)) { in hfi1_file_mmap()
601 + (uctxt->ctxt * RXE_PER_CONTEXT_SIZE); in hfi1_file_mmap()
617 ((uctxt->ctxt - dd->first_user_ctxt) * in hfi1_file_mmap()
689 uctxt->ctxt, subctxt_fp(fp), in hfi1_file_mmap()
[all …]
Dinit.c167 dd->rcd[rcd->ctxt] = NULL; in hfi1_create_ctxts()
177 dd->rcd[rcd->ctxt] = NULL; in hfi1_create_ctxts()
196 struct hfi1_ctxtdata *hfi1_create_ctxtdata(struct hfi1_pportdata *ppd, u32 ctxt) in hfi1_create_ctxtdata() argument
211 dd_dev_info(dd, "%s: setting up context %u\n", __func__, ctxt); in hfi1_create_ctxtdata()
217 rcd->ctxt = ctxt; in hfi1_create_ctxtdata()
218 dd->rcd[ctxt] = rcd; in hfi1_create_ctxtdata()
231 if (ctxt < dd->first_user_ctxt) { in hfi1_create_ctxtdata()
232 if (ctxt < kctxt_ngroups) { in hfi1_create_ctxtdata()
233 base = ctxt * (dd->rcv_entries.ngroups + 1); in hfi1_create_ctxtdata()
237 (ctxt * dd->rcv_entries.ngroups); in hfi1_create_ctxtdata()
[all …]
Dchip.h560 static inline u64 read_kctxt_csr(const struct hfi1_devdata *dd, int ctxt, in read_kctxt_csr() argument
564 return read_csr(dd, offset0 + (0x100 * ctxt)); in read_kctxt_csr()
567 static inline void write_kctxt_csr(struct hfi1_devdata *dd, int ctxt, in write_kctxt_csr() argument
571 write_csr(dd, offset0 + (0x100 * ctxt), value); in write_kctxt_csr()
583 int ctxt, in get_kctxt_csr_addr() argument
586 return get_csr_addr(dd, offset0 + (0x100 * ctxt)); in get_kctxt_csr_addr()
595 static inline u64 read_uctxt_csr(const struct hfi1_devdata *dd, int ctxt, in read_uctxt_csr() argument
599 return read_csr(dd, offset0 + (0x1000 * ctxt)); in read_uctxt_csr()
602 static inline void write_uctxt_csr(struct hfi1_devdata *dd, int ctxt, in write_uctxt_csr() argument
606 write_csr(dd, offset0 + (0x1000 * ctxt), value); in write_uctxt_csr()
[all …]
Duser_sdma.c276 (req)->pq->ctxt, (req)->pq->subctxt, (req)->info.comp_idx, \
279 hfi1_cdbg(SDMA, "[%u:%u:%u] " fmt, (pq)->dd->unit, (pq)->ctxt, \
386 pq->ctxt = uctxt->ctxt; in hfi1_user_sdma_alloc_queues()
395 snprintf(buf, 64, "txreq-kmem-cache-%u-%u-%u", dd->unit, uctxt->ctxt, in hfi1_user_sdma_alloc_queues()
404 uctxt->ctxt); in hfi1_user_sdma_alloc_queues()
448 uctxt->ctxt, fd->subctxt); in hfi1_user_sdma_free_queues()
502 dd->unit, uctxt->ctxt, subctxt_fp(fp), in hfi1_user_sdma_process_request()
510 dd->unit, uctxt->ctxt, subctxt_fp(fp), ret); in hfi1_user_sdma_process_request()
514 trace_hfi1_sdma_user_reqinfo(dd, uctxt->ctxt, subctxt_fp(fp), in hfi1_user_sdma_process_request()
518 dd->unit, uctxt->ctxt, subctxt_fp(fp), in hfi1_user_sdma_process_request()
[all …]
Dchip.c4640 trace_hfi1_receive_interrupt(dd, rcd->ctxt); in receive_context_interrupt()
7595 write_kctxt_csr(dd, rcd->ctxt, RCV_AVAIL_TIME_OUT, in adjust_rcv_timeout()
7604 u32 ctxt = rcd->ctxt; in update_usrhead() local
7615 write_uctxt_csr(dd, ctxt, RCV_EGR_INDEX_HEAD, reg); in update_usrhead()
7621 write_uctxt_csr(dd, ctxt, RCV_HDR_HEAD, reg); in update_usrhead()
7629 head = (read_uctxt_csr(rcd->dd, rcd->ctxt, RCV_HDR_HEAD) in hdrqempty()
7635 tail = read_uctxt_csr(rcd->dd, rcd->ctxt, RCV_HDR_TAIL); in hdrqempty()
7676 void hfi1_rcvctrl(struct hfi1_devdata *dd, unsigned int op, int ctxt) in hfi1_rcvctrl() argument
7682 rcd = dd->rcd[ctxt]; in hfi1_rcvctrl()
7686 hfi1_cdbg(RCVCTRL, "ctxt %d op 0x%x", ctxt, op); in hfi1_rcvctrl()
[all …]
/drivers/scsi/be2iscsi/
Dbe_cmds.c957 void *ctxt = &req->context; in beiscsi_cmd_cq_create() local
971 ctxt, coalesce_wm); in beiscsi_cmd_cq_create()
972 AMAP_SET_BITS(struct amap_cq_context, nodelay, ctxt, no_delay); in beiscsi_cmd_cq_create()
973 AMAP_SET_BITS(struct amap_cq_context, count, ctxt, in beiscsi_cmd_cq_create()
975 AMAP_SET_BITS(struct amap_cq_context, valid, ctxt, 1); in beiscsi_cmd_cq_create()
976 AMAP_SET_BITS(struct amap_cq_context, solevent, ctxt, sol_evts); in beiscsi_cmd_cq_create()
977 AMAP_SET_BITS(struct amap_cq_context, eventable, ctxt, 1); in beiscsi_cmd_cq_create()
978 AMAP_SET_BITS(struct amap_cq_context, eqid, ctxt, eq->id); in beiscsi_cmd_cq_create()
979 AMAP_SET_BITS(struct amap_cq_context, armed, ctxt, 1); in beiscsi_cmd_cq_create()
980 AMAP_SET_BITS(struct amap_cq_context, func, ctxt, in beiscsi_cmd_cq_create()
[all …]
/drivers/infiniband/hw/qib/
Dqib_file_ops.c193 kinfo->spi_uregbase = (u64) dd->uregbase + dd->ureg_align * rcd->ctxt; in qib_get_base_info()
242 kinfo->spi_ctxt = rcd->ctxt; in qib_get_base_info()
308 ctxttid = rcd->ctxt * dd->rcvtidcnt; in qib_tid_update()
503 ctxttid = rcd->ctxt * dd->rcvtidcnt; in qib_tid_free()
693 dd->f_rcvctrl(rcd->ppd, rcvctrl_op, rcd->ctxt); in qib_manage_rcvq()
770 what, rcd->ctxt, pfn, len, ret); in qib_mmap_mem()
1050 ureg = dd->uregbase + dd->ureg_align * rcd->ctxt; in qib_mmapf()
1143 dd->f_rcvctrl(rcd->ppd, QIB_RCVCTRL_INTRAVAIL_ENB, rcd->ctxt); in qib_poll_next()
1312 static int setup_ctxt(struct qib_pportdata *ppd, int ctxt, in setup_ctxt() argument
1328 rcd = qib_create_ctxtdata(ppd, ctxt, numa_id); in setup_ctxt()
[all …]
Dqib_tx.c135 unsigned ctxt; in find_ctxt() local
139 for (ctxt = dd->first_user_ctxt; ctxt < dd->cfgctxts; ctxt++) { in find_ctxt()
140 rcd = dd->rcd[ctxt]; in find_ctxt()
462 unsigned ctxt; in qib_cancel_sends() local
474 for (ctxt = dd->first_user_ctxt; ctxt < dd->cfgctxts; ctxt++) { in qib_cancel_sends()
476 rcd = dd->rcd[ctxt]; in qib_cancel_sends()
Dqib_init.c168 struct qib_ctxtdata *qib_create_ctxtdata(struct qib_pportdata *ppd, u32 ctxt, in qib_create_ctxtdata() argument
181 rcd->ctxt = ctxt; in qib_create_ctxtdata()
182 dd->rcd[ctxt] = rcd; in qib_create_ctxtdata()
184 if (ctxt < dd->first_user_ctxt) { /* N/A for PSM contexts */ in qib_create_ctxtdata()
1347 int ctxt; in cleanup_device_data() local
1385 for (ctxt = 0; ctxt < dd->cfgctxts; ctxt++) { in cleanup_device_data()
1386 int ctxt_tidbase = ctxt * dd->rcvtidcnt; in cleanup_device_data()
1416 for (ctxt = 0; tmp && ctxt < dd->ctxtcnt; ctxt++) { in cleanup_device_data()
1417 struct qib_ctxtdata *rcd = tmp[ctxt]; in cleanup_device_data()
1419 tmp[ctxt] = NULL; /* debugging paranoia */ in cleanup_device_data()
[all …]
Dqib_iba6120.c307 enum qib_ureg regno, int ctxt) in qib_read_ureg32() argument
315 dd->ureg_align * ctxt)); in qib_read_ureg32()
320 dd->ureg_align * ctxt)); in qib_read_ureg32()
333 enum qib_ureg regno, u64 value, int ctxt) in qib_write_ureg() argument
340 dd->ureg_align * ctxt); in qib_write_ureg()
345 dd->ureg_align * ctxt); in qib_write_ureg()
383 const u16 regno, unsigned ctxt, in qib_write_kreg_ctxt() argument
386 qib_write_kreg(dd, regno + ctxt, value); in qib_write_kreg_ctxt()
1985 u32 ctxt; in qib_6120_clear_tids() local
1991 ctxt = rcd->ctxt; in qib_6120_clear_tids()
[all …]
Dqib_iba7220.c230 enum qib_ureg regno, int ctxt) in qib_read_ureg32() argument
238 dd->ureg_align * ctxt)); in qib_read_ureg32()
243 dd->ureg_align * ctxt)); in qib_read_ureg32()
256 enum qib_ureg regno, u64 value, int ctxt) in qib_write_ureg() argument
263 dd->ureg_align * ctxt); in qib_write_ureg()
268 dd->ureg_align * ctxt); in qib_write_ureg()
282 const u16 regno, unsigned ctxt, in qib_write_kreg_ctxt() argument
285 qib_write_kreg(dd, regno + ctxt, value); in qib_write_kreg_ctxt()
2226 u32 ctxt; in qib_7220_clear_tids() local
2232 ctxt = rcd->ctxt; in qib_7220_clear_tids()
[all …]
Dqib_iba7322.c776 enum qib_ureg regno, int ctxt) in qib_read_ureg32() argument
781 (dd->ureg_align * ctxt) + (dd->userbase ? in qib_read_ureg32()
797 enum qib_ureg regno, int ctxt) in qib_read_ureg() argument
803 (dd->ureg_align * ctxt) + (dd->userbase ? in qib_read_ureg()
818 enum qib_ureg regno, u64 value, int ctxt) in qib_write_ureg() argument
825 dd->ureg_align * ctxt); in qib_write_ureg()
830 dd->ureg_align * ctxt); in qib_write_ureg()
887 const u16 regno, unsigned ctxt, in qib_write_kreg_ctxt() argument
890 qib_write_kreg(dd, regno + ctxt, value); in qib_write_kreg_ctxt()
2712 if (cspec->rhdr_cpu[rcd->ctxt] != cpu) { in qib_update_rhdrq_dca()
[all …]
/drivers/net/ethernet/emulex/benet/
Dbe_cmds.c1102 void *ctxt; in be_cmd_cq_create() local
1110 ctxt = &req->context; in be_cmd_cq_create()
1119 AMAP_SET_BITS(struct amap_cq_context_be, coalescwm, ctxt, in be_cmd_cq_create()
1122 ctxt, no_delay); in be_cmd_cq_create()
1123 AMAP_SET_BITS(struct amap_cq_context_be, count, ctxt, in be_cmd_cq_create()
1125 AMAP_SET_BITS(struct amap_cq_context_be, valid, ctxt, 1); in be_cmd_cq_create()
1126 AMAP_SET_BITS(struct amap_cq_context_be, eventable, ctxt, 1); in be_cmd_cq_create()
1127 AMAP_SET_BITS(struct amap_cq_context_be, eqid, ctxt, eq->id); in be_cmd_cq_create()
1137 ctxt, coalesce_wm); in be_cmd_cq_create()
1138 AMAP_SET_BITS(struct amap_cq_context_v2, nodelay, ctxt, in be_cmd_cq_create()
[all …]
/drivers/net/ethernet/intel/i40e/
Di40e_main.c1605 struct i40e_vsi_context *ctxt, in i40e_vsi_setup_queue_map() argument
1610 struct i40e_vsi_context *ctxt, in i40e_vsi_setup_queue_map()
1709 ctxt->info.tc_mapping[i] = cpu_to_le16(qmap); in i40e_vsi_setup_queue_map()
1725 ctxt->info.up_enable_bits = enabled_tc; in i40e_vsi_setup_queue_map()
1728 ctxt->info.mapping_flags |= in i40e_vsi_setup_queue_map()
1731 ctxt->info.queue_mapping[i] = in i40e_vsi_setup_queue_map()
1734 ctxt->info.mapping_flags |= in i40e_vsi_setup_queue_map()
1736 ctxt->info.queue_mapping[0] = cpu_to_le16(vsi->base_queue); in i40e_vsi_setup_queue_map()
1738 ctxt->info.valid_sections |= cpu_to_le16(sections); in i40e_vsi_setup_queue_map()
2263 struct i40e_vsi_context ctxt; in i40e_vlan_stripping_enable() local
[all …]
Di40e_fcoe.c366 int i40e_fcoe_vsi_init(struct i40e_vsi *vsi, struct i40e_vsi_context *ctxt) in i40e_fcoe_vsi_init() argument
368 struct i40e_aqc_vsi_properties_data *info = &ctxt->info; in i40e_fcoe_vsi_init()
380 ctxt->pf_num = hw->pf_id; in i40e_fcoe_vsi_init()
381 ctxt->vf_num = 0; in i40e_fcoe_vsi_init()
382 ctxt->uplink_seid = vsi->uplink_seid; in i40e_fcoe_vsi_init()
383 ctxt->connection_type = I40E_AQ_VSI_CONN_TYPE_NORMAL; in i40e_fcoe_vsi_init()
384 ctxt->flags = I40E_AQ_VSI_TYPE_PF; in i40e_fcoe_vsi_init()
403 i40e_vsi_setup_queue_map(vsi, ctxt, enabled_tc, true); in i40e_fcoe_vsi_init()
/drivers/staging/lustre/lustre/mgc/
Dmgc_request.c634 struct llog_ctxt *ctxt; in mgc_llog_init() local
644 ctxt = llog_get_context(obd, LLOG_CONFIG_REPL_CTXT); in mgc_llog_init()
645 LASSERT(ctxt); in mgc_llog_init()
647 llog_initiator_connect(ctxt); in mgc_llog_init()
648 llog_ctxt_put(ctxt); in mgc_llog_init()
655 struct llog_ctxt *ctxt; in mgc_llog_fini() local
657 ctxt = llog_get_context(obd, LLOG_CONFIG_REPL_CTXT); in mgc_llog_fini()
658 if (ctxt) in mgc_llog_fini()
659 llog_cleanup(env, ctxt); in mgc_llog_fini()
1450 struct llog_ctxt *ctxt; in mgc_process_cfg_log() local
[all …]
/drivers/staging/lustre/lustre/mdc/
Dmdc_request.c1504 struct llog_ctxt *ctxt = NULL; in mdc_changelog_send_thread() local
1519 ctxt = llog_get_context(cs->cs_obd, LLOG_CHANGELOG_REPL_CTXT); in mdc_changelog_send_thread()
1520 if (ctxt == NULL) { in mdc_changelog_send_thread()
1524 rc = llog_open(NULL, ctxt, &llh, NULL, CHANGELOG_CATALOG, in mdc_changelog_send_thread()
1550 if (ctxt) in mdc_changelog_send_thread()
1551 llog_ctxt_put(ctxt); in mdc_changelog_send_thread()
2276 struct llog_ctxt *ctxt; in mdc_llog_init() local
2284 ctxt = llog_group_get_ctxt(olg, LLOG_CHANGELOG_REPL_CTXT); in mdc_llog_init()
2285 llog_initiator_connect(ctxt); in mdc_llog_init()
2286 llog_ctxt_put(ctxt); in mdc_llog_init()
[all …]

12