/drivers/staging/lustre/lustre/obdclass/ |
D | llog_obd.c | 46 struct llog_ctxt *ctxt; in llog_new_ctxt() local 48 ctxt = kzalloc(sizeof(*ctxt), GFP_NOFS); in llog_new_ctxt() 49 if (!ctxt) in llog_new_ctxt() 52 ctxt->loc_obd = obd; in llog_new_ctxt() 53 atomic_set(&ctxt->loc_refcount, 1); in llog_new_ctxt() 55 return ctxt; in llog_new_ctxt() 58 static void llog_ctxt_destroy(struct llog_ctxt *ctxt) in llog_ctxt_destroy() argument 60 if (ctxt->loc_exp) { in llog_ctxt_destroy() 61 class_export_put(ctxt->loc_exp); in llog_ctxt_destroy() 62 ctxt->loc_exp = NULL; in llog_ctxt_destroy() [all …]
|
D | llog.c | 409 int llog_open(const struct lu_env *env, struct llog_ctxt *ctxt, in llog_open() argument 416 LASSERT(ctxt); in llog_open() 417 LASSERT(ctxt->loc_logops); in llog_open() 419 if (ctxt->loc_logops->lop_open == NULL) { in llog_open() 427 (*lgh)->lgh_ctxt = ctxt; in llog_open() 428 (*lgh)->lgh_logops = ctxt->loc_logops; in llog_open() 433 rc = ctxt->loc_logops->lop_open(env, *lgh, logid, name, open_param); in llog_open()
|
/drivers/staging/lustre/lustre/ptlrpc/ |
D | llog_net.c | 55 int llog_initiator_connect(struct llog_ctxt *ctxt) in llog_initiator_connect() argument 59 LASSERT(ctxt); in llog_initiator_connect() 60 new_imp = ctxt->loc_obd->u.cli.cl_import; in llog_initiator_connect() 61 LASSERTF(ctxt->loc_imp == NULL || ctxt->loc_imp == new_imp, in llog_initiator_connect() 62 "%p - %p\n", ctxt->loc_imp, new_imp); in llog_initiator_connect() 63 mutex_lock(&ctxt->loc_mutex); in llog_initiator_connect() 64 if (ctxt->loc_imp != new_imp) { in llog_initiator_connect() 65 if (ctxt->loc_imp) in llog_initiator_connect() 66 class_import_put(ctxt->loc_imp); in llog_initiator_connect() 67 ctxt->loc_imp = class_import_get(new_imp); in llog_initiator_connect() [all …]
|
D | llog_client.c | 52 #define LLOG_CLIENT_ENTRY(ctxt, imp) do { \ argument 53 mutex_lock(&ctxt->loc_mutex); \ 54 if (ctxt->loc_imp) { \ 55 imp = class_import_get(ctxt->loc_imp); \ 60 ctxt->loc_idx); \ 62 mutex_unlock(&ctxt->loc_mutex); \ 65 mutex_unlock(&ctxt->loc_mutex); \ 68 #define LLOG_CLIENT_EXIT(ctxt, imp) do { \ argument 69 mutex_lock(&ctxt->loc_mutex); \ 70 if (ctxt->loc_imp != imp) \ [all …]
|
/drivers/staging/lustre/lustre/include/ |
D | lustre_log.h | 93 int llog_open(const struct lu_env *env, struct llog_ctxt *ctxt, 98 struct llog_ctxt *ctxt, struct llog_ctxt *bak_ctxt, 144 int __llog_ctxt_put(const struct lu_env *env, struct llog_ctxt *ctxt); 148 int llog_initiator_connect(struct llog_ctxt *ctxt); 161 int (*lop_sync)(struct llog_ctxt *ctxt, struct obd_export *exp, 163 int (*lop_cleanup)(const struct lu_env *env, struct llog_ctxt *ctxt); 164 int (*lop_cancel)(const struct lu_env *env, struct llog_ctxt *ctxt, 166 int (*lop_connect)(struct llog_ctxt *ctxt, struct llog_logid *logid, 268 static inline struct llog_ctxt *llog_ctxt_get(struct llog_ctxt *ctxt) in llog_ctxt_get() argument 270 atomic_inc(&ctxt->loc_refcount); in llog_ctxt_get() [all …]
|
/drivers/net/wireless/iwlwifi/mvm/ |
D | phy-ctxt.c | 127 static void iwl_mvm_phy_ctxt_cmd_hdr(struct iwl_mvm_phy_ctxt *ctxt, in iwl_mvm_phy_ctxt_cmd_hdr() argument 133 cmd->id_and_color = cpu_to_le32(FW_CMD_ID_AND_COLOR(ctxt->id, in iwl_mvm_phy_ctxt_cmd_hdr() 134 ctxt->color)); in iwl_mvm_phy_ctxt_cmd_hdr() 193 struct iwl_mvm_phy_ctxt *ctxt, in iwl_mvm_phy_ctxt_apply() argument 202 iwl_mvm_phy_ctxt_cmd_hdr(ctxt, &cmd, action, apply_time); in iwl_mvm_phy_ctxt_apply() 219 int iwl_mvm_phy_ctxt_add(struct iwl_mvm *mvm, struct iwl_mvm_phy_ctxt *ctxt, in iwl_mvm_phy_ctxt_add() argument 224 ctxt->ref); in iwl_mvm_phy_ctxt_add() 227 ctxt->channel = chandef->chan; in iwl_mvm_phy_ctxt_add() 229 return iwl_mvm_phy_ctxt_apply(mvm, ctxt, chandef, in iwl_mvm_phy_ctxt_add() 238 void iwl_mvm_phy_ctxt_ref(struct iwl_mvm *mvm, struct iwl_mvm_phy_ctxt *ctxt) in iwl_mvm_phy_ctxt_ref() argument [all …]
|
D | Makefile | 2 iwlmvm-y += fw.o mac80211.o nvm.o ops.o phy-ctxt.o mac-ctxt.o
|
/drivers/staging/rdma/hfi1/ |
D | trace.h | 81 u32 ctxt, 87 TP_ARGS(dd, ctxt, eflags, etype, hlen, tlen, updegr, etail), 91 __field(u32, ctxt) 101 __entry->ctxt = ctxt; 111 __entry->ctxt, 122 TP_PROTO(struct hfi1_devdata *dd, u32 ctxt), 123 TP_ARGS(dd, ctxt), 126 __field(u32, ctxt) 132 __entry->ctxt = ctxt; 133 if (dd->rcd[ctxt]->do_interrupt == [all …]
|
D | file_ops.c | 168 #define HFI1_MMAP_TOKEN(type, ctxt, subctxt, addr) \ argument 171 HFI1_MMAP_TOKEN_SET(CTXT, ctxt) | \ 413 uctxt->ctxt); in hfi1_file_write() 457 ctxt_fp(kiocb->ki_filp)->ctxt, subctxt_fp(kiocb->ki_filp), in hfi1_write_iter() 493 u16 ctxt; in hfi1_file_mmap() local 502 ctxt = HFI1_MMAP_TOKEN_GET(CTXT, token); in hfi1_file_mmap() 505 if (ctxt != uctxt->ctxt || subctxt != subctxt_fp(fp)) { in hfi1_file_mmap() 601 + (uctxt->ctxt * RXE_PER_CONTEXT_SIZE); in hfi1_file_mmap() 617 ((uctxt->ctxt - dd->first_user_ctxt) * in hfi1_file_mmap() 689 uctxt->ctxt, subctxt_fp(fp), in hfi1_file_mmap() [all …]
|
D | init.c | 167 dd->rcd[rcd->ctxt] = NULL; in hfi1_create_ctxts() 177 dd->rcd[rcd->ctxt] = NULL; in hfi1_create_ctxts() 196 struct hfi1_ctxtdata *hfi1_create_ctxtdata(struct hfi1_pportdata *ppd, u32 ctxt) in hfi1_create_ctxtdata() argument 211 dd_dev_info(dd, "%s: setting up context %u\n", __func__, ctxt); in hfi1_create_ctxtdata() 217 rcd->ctxt = ctxt; in hfi1_create_ctxtdata() 218 dd->rcd[ctxt] = rcd; in hfi1_create_ctxtdata() 231 if (ctxt < dd->first_user_ctxt) { in hfi1_create_ctxtdata() 232 if (ctxt < kctxt_ngroups) { in hfi1_create_ctxtdata() 233 base = ctxt * (dd->rcv_entries.ngroups + 1); in hfi1_create_ctxtdata() 237 (ctxt * dd->rcv_entries.ngroups); in hfi1_create_ctxtdata() [all …]
|
D | chip.h | 560 static inline u64 read_kctxt_csr(const struct hfi1_devdata *dd, int ctxt, in read_kctxt_csr() argument 564 return read_csr(dd, offset0 + (0x100 * ctxt)); in read_kctxt_csr() 567 static inline void write_kctxt_csr(struct hfi1_devdata *dd, int ctxt, in write_kctxt_csr() argument 571 write_csr(dd, offset0 + (0x100 * ctxt), value); in write_kctxt_csr() 583 int ctxt, in get_kctxt_csr_addr() argument 586 return get_csr_addr(dd, offset0 + (0x100 * ctxt)); in get_kctxt_csr_addr() 595 static inline u64 read_uctxt_csr(const struct hfi1_devdata *dd, int ctxt, in read_uctxt_csr() argument 599 return read_csr(dd, offset0 + (0x1000 * ctxt)); in read_uctxt_csr() 602 static inline void write_uctxt_csr(struct hfi1_devdata *dd, int ctxt, in write_uctxt_csr() argument 606 write_csr(dd, offset0 + (0x1000 * ctxt), value); in write_uctxt_csr() [all …]
|
D | user_sdma.c | 276 (req)->pq->ctxt, (req)->pq->subctxt, (req)->info.comp_idx, \ 279 hfi1_cdbg(SDMA, "[%u:%u:%u] " fmt, (pq)->dd->unit, (pq)->ctxt, \ 386 pq->ctxt = uctxt->ctxt; in hfi1_user_sdma_alloc_queues() 395 snprintf(buf, 64, "txreq-kmem-cache-%u-%u-%u", dd->unit, uctxt->ctxt, in hfi1_user_sdma_alloc_queues() 404 uctxt->ctxt); in hfi1_user_sdma_alloc_queues() 448 uctxt->ctxt, fd->subctxt); in hfi1_user_sdma_free_queues() 502 dd->unit, uctxt->ctxt, subctxt_fp(fp), in hfi1_user_sdma_process_request() 510 dd->unit, uctxt->ctxt, subctxt_fp(fp), ret); in hfi1_user_sdma_process_request() 514 trace_hfi1_sdma_user_reqinfo(dd, uctxt->ctxt, subctxt_fp(fp), in hfi1_user_sdma_process_request() 518 dd->unit, uctxt->ctxt, subctxt_fp(fp), in hfi1_user_sdma_process_request() [all …]
|
D | chip.c | 4640 trace_hfi1_receive_interrupt(dd, rcd->ctxt); in receive_context_interrupt() 7595 write_kctxt_csr(dd, rcd->ctxt, RCV_AVAIL_TIME_OUT, in adjust_rcv_timeout() 7604 u32 ctxt = rcd->ctxt; in update_usrhead() local 7615 write_uctxt_csr(dd, ctxt, RCV_EGR_INDEX_HEAD, reg); in update_usrhead() 7621 write_uctxt_csr(dd, ctxt, RCV_HDR_HEAD, reg); in update_usrhead() 7629 head = (read_uctxt_csr(rcd->dd, rcd->ctxt, RCV_HDR_HEAD) in hdrqempty() 7635 tail = read_uctxt_csr(rcd->dd, rcd->ctxt, RCV_HDR_TAIL); in hdrqempty() 7676 void hfi1_rcvctrl(struct hfi1_devdata *dd, unsigned int op, int ctxt) in hfi1_rcvctrl() argument 7682 rcd = dd->rcd[ctxt]; in hfi1_rcvctrl() 7686 hfi1_cdbg(RCVCTRL, "ctxt %d op 0x%x", ctxt, op); in hfi1_rcvctrl() [all …]
|
/drivers/scsi/be2iscsi/ |
D | be_cmds.c | 957 void *ctxt = &req->context; in beiscsi_cmd_cq_create() local 971 ctxt, coalesce_wm); in beiscsi_cmd_cq_create() 972 AMAP_SET_BITS(struct amap_cq_context, nodelay, ctxt, no_delay); in beiscsi_cmd_cq_create() 973 AMAP_SET_BITS(struct amap_cq_context, count, ctxt, in beiscsi_cmd_cq_create() 975 AMAP_SET_BITS(struct amap_cq_context, valid, ctxt, 1); in beiscsi_cmd_cq_create() 976 AMAP_SET_BITS(struct amap_cq_context, solevent, ctxt, sol_evts); in beiscsi_cmd_cq_create() 977 AMAP_SET_BITS(struct amap_cq_context, eventable, ctxt, 1); in beiscsi_cmd_cq_create() 978 AMAP_SET_BITS(struct amap_cq_context, eqid, ctxt, eq->id); in beiscsi_cmd_cq_create() 979 AMAP_SET_BITS(struct amap_cq_context, armed, ctxt, 1); in beiscsi_cmd_cq_create() 980 AMAP_SET_BITS(struct amap_cq_context, func, ctxt, in beiscsi_cmd_cq_create() [all …]
|
/drivers/infiniband/hw/qib/ |
D | qib_file_ops.c | 193 kinfo->spi_uregbase = (u64) dd->uregbase + dd->ureg_align * rcd->ctxt; in qib_get_base_info() 242 kinfo->spi_ctxt = rcd->ctxt; in qib_get_base_info() 308 ctxttid = rcd->ctxt * dd->rcvtidcnt; in qib_tid_update() 503 ctxttid = rcd->ctxt * dd->rcvtidcnt; in qib_tid_free() 693 dd->f_rcvctrl(rcd->ppd, rcvctrl_op, rcd->ctxt); in qib_manage_rcvq() 770 what, rcd->ctxt, pfn, len, ret); in qib_mmap_mem() 1050 ureg = dd->uregbase + dd->ureg_align * rcd->ctxt; in qib_mmapf() 1143 dd->f_rcvctrl(rcd->ppd, QIB_RCVCTRL_INTRAVAIL_ENB, rcd->ctxt); in qib_poll_next() 1312 static int setup_ctxt(struct qib_pportdata *ppd, int ctxt, in setup_ctxt() argument 1328 rcd = qib_create_ctxtdata(ppd, ctxt, numa_id); in setup_ctxt() [all …]
|
D | qib_tx.c | 135 unsigned ctxt; in find_ctxt() local 139 for (ctxt = dd->first_user_ctxt; ctxt < dd->cfgctxts; ctxt++) { in find_ctxt() 140 rcd = dd->rcd[ctxt]; in find_ctxt() 462 unsigned ctxt; in qib_cancel_sends() local 474 for (ctxt = dd->first_user_ctxt; ctxt < dd->cfgctxts; ctxt++) { in qib_cancel_sends() 476 rcd = dd->rcd[ctxt]; in qib_cancel_sends()
|
D | qib_init.c | 168 struct qib_ctxtdata *qib_create_ctxtdata(struct qib_pportdata *ppd, u32 ctxt, in qib_create_ctxtdata() argument 181 rcd->ctxt = ctxt; in qib_create_ctxtdata() 182 dd->rcd[ctxt] = rcd; in qib_create_ctxtdata() 184 if (ctxt < dd->first_user_ctxt) { /* N/A for PSM contexts */ in qib_create_ctxtdata() 1347 int ctxt; in cleanup_device_data() local 1385 for (ctxt = 0; ctxt < dd->cfgctxts; ctxt++) { in cleanup_device_data() 1386 int ctxt_tidbase = ctxt * dd->rcvtidcnt; in cleanup_device_data() 1416 for (ctxt = 0; tmp && ctxt < dd->ctxtcnt; ctxt++) { in cleanup_device_data() 1417 struct qib_ctxtdata *rcd = tmp[ctxt]; in cleanup_device_data() 1419 tmp[ctxt] = NULL; /* debugging paranoia */ in cleanup_device_data() [all …]
|
D | qib_iba6120.c | 307 enum qib_ureg regno, int ctxt) in qib_read_ureg32() argument 315 dd->ureg_align * ctxt)); in qib_read_ureg32() 320 dd->ureg_align * ctxt)); in qib_read_ureg32() 333 enum qib_ureg regno, u64 value, int ctxt) in qib_write_ureg() argument 340 dd->ureg_align * ctxt); in qib_write_ureg() 345 dd->ureg_align * ctxt); in qib_write_ureg() 383 const u16 regno, unsigned ctxt, in qib_write_kreg_ctxt() argument 386 qib_write_kreg(dd, regno + ctxt, value); in qib_write_kreg_ctxt() 1985 u32 ctxt; in qib_6120_clear_tids() local 1991 ctxt = rcd->ctxt; in qib_6120_clear_tids() [all …]
|
D | qib_iba7220.c | 230 enum qib_ureg regno, int ctxt) in qib_read_ureg32() argument 238 dd->ureg_align * ctxt)); in qib_read_ureg32() 243 dd->ureg_align * ctxt)); in qib_read_ureg32() 256 enum qib_ureg regno, u64 value, int ctxt) in qib_write_ureg() argument 263 dd->ureg_align * ctxt); in qib_write_ureg() 268 dd->ureg_align * ctxt); in qib_write_ureg() 282 const u16 regno, unsigned ctxt, in qib_write_kreg_ctxt() argument 285 qib_write_kreg(dd, regno + ctxt, value); in qib_write_kreg_ctxt() 2226 u32 ctxt; in qib_7220_clear_tids() local 2232 ctxt = rcd->ctxt; in qib_7220_clear_tids() [all …]
|
D | qib_iba7322.c | 776 enum qib_ureg regno, int ctxt) in qib_read_ureg32() argument 781 (dd->ureg_align * ctxt) + (dd->userbase ? in qib_read_ureg32() 797 enum qib_ureg regno, int ctxt) in qib_read_ureg() argument 803 (dd->ureg_align * ctxt) + (dd->userbase ? in qib_read_ureg() 818 enum qib_ureg regno, u64 value, int ctxt) in qib_write_ureg() argument 825 dd->ureg_align * ctxt); in qib_write_ureg() 830 dd->ureg_align * ctxt); in qib_write_ureg() 887 const u16 regno, unsigned ctxt, in qib_write_kreg_ctxt() argument 890 qib_write_kreg(dd, regno + ctxt, value); in qib_write_kreg_ctxt() 2712 if (cspec->rhdr_cpu[rcd->ctxt] != cpu) { in qib_update_rhdrq_dca() [all …]
|
/drivers/net/ethernet/emulex/benet/ |
D | be_cmds.c | 1102 void *ctxt; in be_cmd_cq_create() local 1110 ctxt = &req->context; in be_cmd_cq_create() 1119 AMAP_SET_BITS(struct amap_cq_context_be, coalescwm, ctxt, in be_cmd_cq_create() 1122 ctxt, no_delay); in be_cmd_cq_create() 1123 AMAP_SET_BITS(struct amap_cq_context_be, count, ctxt, in be_cmd_cq_create() 1125 AMAP_SET_BITS(struct amap_cq_context_be, valid, ctxt, 1); in be_cmd_cq_create() 1126 AMAP_SET_BITS(struct amap_cq_context_be, eventable, ctxt, 1); in be_cmd_cq_create() 1127 AMAP_SET_BITS(struct amap_cq_context_be, eqid, ctxt, eq->id); in be_cmd_cq_create() 1137 ctxt, coalesce_wm); in be_cmd_cq_create() 1138 AMAP_SET_BITS(struct amap_cq_context_v2, nodelay, ctxt, in be_cmd_cq_create() [all …]
|
/drivers/net/ethernet/intel/i40e/ |
D | i40e_main.c | 1605 struct i40e_vsi_context *ctxt, in i40e_vsi_setup_queue_map() argument 1610 struct i40e_vsi_context *ctxt, in i40e_vsi_setup_queue_map() 1709 ctxt->info.tc_mapping[i] = cpu_to_le16(qmap); in i40e_vsi_setup_queue_map() 1725 ctxt->info.up_enable_bits = enabled_tc; in i40e_vsi_setup_queue_map() 1728 ctxt->info.mapping_flags |= in i40e_vsi_setup_queue_map() 1731 ctxt->info.queue_mapping[i] = in i40e_vsi_setup_queue_map() 1734 ctxt->info.mapping_flags |= in i40e_vsi_setup_queue_map() 1736 ctxt->info.queue_mapping[0] = cpu_to_le16(vsi->base_queue); in i40e_vsi_setup_queue_map() 1738 ctxt->info.valid_sections |= cpu_to_le16(sections); in i40e_vsi_setup_queue_map() 2263 struct i40e_vsi_context ctxt; in i40e_vlan_stripping_enable() local [all …]
|
D | i40e_fcoe.c | 366 int i40e_fcoe_vsi_init(struct i40e_vsi *vsi, struct i40e_vsi_context *ctxt) in i40e_fcoe_vsi_init() argument 368 struct i40e_aqc_vsi_properties_data *info = &ctxt->info; in i40e_fcoe_vsi_init() 380 ctxt->pf_num = hw->pf_id; in i40e_fcoe_vsi_init() 381 ctxt->vf_num = 0; in i40e_fcoe_vsi_init() 382 ctxt->uplink_seid = vsi->uplink_seid; in i40e_fcoe_vsi_init() 383 ctxt->connection_type = I40E_AQ_VSI_CONN_TYPE_NORMAL; in i40e_fcoe_vsi_init() 384 ctxt->flags = I40E_AQ_VSI_TYPE_PF; in i40e_fcoe_vsi_init() 403 i40e_vsi_setup_queue_map(vsi, ctxt, enabled_tc, true); in i40e_fcoe_vsi_init()
|
/drivers/staging/lustre/lustre/mgc/ |
D | mgc_request.c | 634 struct llog_ctxt *ctxt; in mgc_llog_init() local 644 ctxt = llog_get_context(obd, LLOG_CONFIG_REPL_CTXT); in mgc_llog_init() 645 LASSERT(ctxt); in mgc_llog_init() 647 llog_initiator_connect(ctxt); in mgc_llog_init() 648 llog_ctxt_put(ctxt); in mgc_llog_init() 655 struct llog_ctxt *ctxt; in mgc_llog_fini() local 657 ctxt = llog_get_context(obd, LLOG_CONFIG_REPL_CTXT); in mgc_llog_fini() 658 if (ctxt) in mgc_llog_fini() 659 llog_cleanup(env, ctxt); in mgc_llog_fini() 1450 struct llog_ctxt *ctxt; in mgc_process_cfg_log() local [all …]
|
/drivers/staging/lustre/lustre/mdc/ |
D | mdc_request.c | 1504 struct llog_ctxt *ctxt = NULL; in mdc_changelog_send_thread() local 1519 ctxt = llog_get_context(cs->cs_obd, LLOG_CHANGELOG_REPL_CTXT); in mdc_changelog_send_thread() 1520 if (ctxt == NULL) { in mdc_changelog_send_thread() 1524 rc = llog_open(NULL, ctxt, &llh, NULL, CHANGELOG_CATALOG, in mdc_changelog_send_thread() 1550 if (ctxt) in mdc_changelog_send_thread() 1551 llog_ctxt_put(ctxt); in mdc_changelog_send_thread() 2276 struct llog_ctxt *ctxt; in mdc_llog_init() local 2284 ctxt = llog_group_get_ctxt(olg, LLOG_CHANGELOG_REPL_CTXT); in mdc_llog_init() 2285 llog_initiator_connect(ctxt); in mdc_llog_init() 2286 llog_ctxt_put(ctxt); in mdc_llog_init() [all …]
|