Home
last modified time | relevance | path

Searched refs:pnode (Results 1 – 19 of 19) sorted by relevance

/drivers/scsi/lpfc/
Dlpfc_scsi.c147 struct lpfc_nodelist *pnode; in lpfc_update_stats() local
162 pnode = rdata->pnode; in lpfc_update_stats()
165 if (!pnode || in lpfc_update_stats()
166 !pnode->lat_data || in lpfc_update_stats()
187 pnode->lat_data[i].cmd_count++; in lpfc_update_stats()
487 if (psb->rdata && psb->rdata->pnode && in lpfc_sli4_vport_delete_fcp_xri_aborted()
488 psb->rdata->pnode->vport == vport) in lpfc_sli4_vport_delete_fcp_xri_aborted()
541 if (psb->rdata && psb->rdata->pnode) in lpfc_sli4_io_xri_aborted()
542 ndlp = psb->rdata->pnode; in lpfc_sli4_io_xri_aborted()
1023 if (rdata && rdata->pnode) { in lpfc_bg_err_inject()
[all …]
Dlpfc_scsi.h43 struct lpfc_nodelist *pnode; /* Pointer to the node structure. */ member
Dlpfc_nvme.c1198 struct lpfc_nodelist *pnode, in lpfc_nvme_prep_io_cmd() argument
1207 if (!NLP_CHK_NODE_ACT(pnode)) in lpfc_nvme_prep_io_cmd()
1226 (pnode->nlp_flag & NLP_FIRSTBURST)) { in lpfc_nvme_prep_io_cmd()
1228 if (req_len < pnode->nvme_fb_size) in lpfc_nvme_prep_io_cmd()
1233 pnode->nvme_fb_size; in lpfc_nvme_prep_io_cmd()
1259 if (pnode->nlp_nvme_info & NLP_NVME_NSLER) in lpfc_nvme_prep_io_cmd()
1272 phba->sli4_hba.rpi_ids[pnode->nlp_rpi]); in lpfc_nvme_prep_io_cmd()
Dlpfc_hbadisc.c84 ndlp = rdata->pnode; in lpfc_terminate_rport_io()
124 ndlp = rdata->pnode; in lpfc_dev_loss_tmo_callbk()
144 put_node = rdata->pnode != NULL; in lpfc_dev_loss_tmo_callbk()
146 rdata->pnode = NULL; in lpfc_dev_loss_tmo_callbk()
261 put_node = rdata->pnode != NULL; in lpfc_dev_loss_tmo_handler()
262 rdata->pnode = NULL; in lpfc_dev_loss_tmo_handler()
282 put_node = rdata->pnode != NULL; in lpfc_dev_loss_tmo_handler()
283 rdata->pnode = NULL; in lpfc_dev_loss_tmo_handler()
4128 if (rdata->pnode == ndlp) in lpfc_register_remote_port()
4130 rdata->pnode = NULL; in lpfc_register_remote_port()
[all …]
Dlpfc_els.c1768 if (rdata->pnode == ndlp) { in lpfc_plogi_confirm_nport()
1772 rdata->pnode = lpfc_nlp_get(new_ndlp); in lpfc_plogi_confirm_nport()
1849 put_node = rdata->pnode != NULL; in lpfc_plogi_confirm_nport()
1851 rdata->pnode = NULL; in lpfc_plogi_confirm_nport()
Dlpfc_bsg.c393 struct lpfc_nodelist *ndlp = rdata->pnode; in lpfc_bsg_send_mgmt_cmd()
667 struct lpfc_nodelist *ndlp = rdata->pnode; in lpfc_bsg_rport_els()
Dlpfc_sli.c1130 ndlp = lpfc_cmd->rdata->pnode; in __lpfc_sli_get_els_sglq()
11341 if ((lpfc_cmd->rdata) && (lpfc_cmd->rdata->pnode) && in lpfc_sli_validate_fcp_iocb()
11342 (lpfc_cmd->rdata->pnode->nlp_sid == tgt_id) && in lpfc_sli_validate_fcp_iocb()
11347 if ((lpfc_cmd->rdata) && (lpfc_cmd->rdata->pnode) && in lpfc_sli_validate_fcp_iocb()
11348 (lpfc_cmd->rdata->pnode->nlp_sid == tgt_id)) in lpfc_sli_validate_fcp_iocb()
11646 ndlp = lpfc_cmd->rdata->pnode; in lpfc_sli_abort_taskmgmt()
Dlpfc_attr.c6861 struct lpfc_nodelist *ndlp = rdata->pnode; in lpfc_set_rport_loss_tmo()
/drivers/macintosh/
Dmacio_asic.c454 struct device_node *np, *pnode; in macio_pci_add_devices() local
466 pnode = of_node_get(chip->of_node); in macio_pci_add_devices()
467 if (pnode == NULL) in macio_pci_add_devices()
471 rdev = macio_add_one_device(chip, parent, pnode, NULL, root_res); in macio_pci_add_devices()
477 for (np = NULL; (np = of_get_next_child(pnode, np)) != NULL;) { in macio_pci_add_devices()
493 pnode = mbdev->ofdev.dev.of_node; in macio_pci_add_devices()
494 for (np = NULL; (np = of_get_next_child(pnode, np)) != NULL;) { in macio_pci_add_devices()
506 pnode = sdev->ofdev.dev.of_node; in macio_pci_add_devices()
507 for (np = NULL; (np = of_get_next_child(pnode, np)) != NULL;) { in macio_pci_add_devices()
/drivers/base/power/
Dqos.c162 &req->data.pnode, action, value); in apply_constraint()
166 &req->data.pnode, action, value); in apply_constraint()
273 plist_for_each_entry_safe(req, tmp, &c->list, data.pnode) { in dev_pm_qos_constraints_destroy()
283 plist_for_each_entry_safe(req, tmp, &c->list, data.pnode) { in dev_pm_qos_constraints_destroy()
289 plist_for_each_entry_safe(req, tmp, &c->list, data.freq.pnode) { in dev_pm_qos_constraints_destroy()
296 plist_for_each_entry_safe(req, tmp, &c->list, data.freq.pnode) { in dev_pm_qos_constraints_destroy()
424 curr_value = req->data.pnode.prio; in __dev_pm_qos_update_request()
428 curr_value = req->data.freq.pnode.prio; in __dev_pm_qos_update_request()
894 dev->power.qos->latency_tolerance_req->data.pnode.prio; in dev_pm_qos_get_user_latency_tolerance()
/drivers/clk/st/
Dclk-flexgen.c312 struct device_node *pnode; in st_of_flexgen_setup() local
324 pnode = of_get_parent(np); in st_of_flexgen_setup()
325 if (!pnode) in st_of_flexgen_setup()
328 reg = of_iomap(pnode, 0); in st_of_flexgen_setup()
329 of_node_put(pnode); in st_of_flexgen_setup()
Dclkgen-pll.c624 struct device_node *pnode; in clkgen_get_register_base() local
627 pnode = of_get_parent(np); in clkgen_get_register_base()
628 if (!pnode) in clkgen_get_register_base()
631 reg = of_iomap(pnode, 0); in clkgen_get_register_base()
633 of_node_put(pnode); in clkgen_get_register_base()
/drivers/misc/sgi-gru/
Dgruhandles.h143 static inline unsigned long gru_chiplet_paddr(unsigned long paddr, int pnode, in gru_chiplet_paddr() argument
146 return paddr + GRU_SIZE * (2 * pnode + chiplet); in gru_chiplet_paddr()
149 static inline void *gru_chiplet_vaddr(void *vaddr, int pnode, int chiplet) in gru_chiplet_vaddr() argument
151 return vaddr + GRU_SIZE * (2 * pnode + chiplet); in gru_chiplet_vaddr()
Dgrufile.c253 int pnode, nid, bid, chip; in gru_init_tables() local
264 pnode = uv_blade_to_pnode(bid); in gru_init_tables()
280 paddr = gru_chiplet_paddr(gru_base_paddr, pnode, chip); in gru_init_tables()
281 vaddr = gru_chiplet_vaddr(gru_base_vaddr, pnode, chip); in gru_init_tables()
/drivers/net/ipvlan/
Dipvlan_main.c16 list_for_each_entry(ipvlan, &port->ipvlans, pnode) { in ipvlan_set_port_mode()
45 list_for_each_entry_continue_reverse(ipvlan, &port->ipvlans, pnode) { in ipvlan_set_port_mode()
616 list_add_tail_rcu(&ipvlan->pnode, &port->ipvlans); in ipvlan_link_new()
644 list_del_rcu(&ipvlan->pnode); in ipvlan_link_delete()
709 list_for_each_entry(ipvlan, &port->ipvlans, pnode) in ipvlan_device_event()
730 list_for_each_entry_safe(ipvlan, next, &port->ipvlans, pnode) in ipvlan_device_event()
737 list_for_each_entry(ipvlan, &port->ipvlans, pnode) { in ipvlan_device_event()
746 list_for_each_entry(ipvlan, &port->ipvlans, pnode) in ipvlan_device_event()
752 list_for_each_entry(ipvlan, &port->ipvlans, pnode) { in ipvlan_device_event()
762 list_for_each_entry(ipvlan, &port->ipvlans, pnode) { in ipvlan_device_event()
Dipvlan.h64 struct list_head pnode; member
Dipvlan_core.c126 list_for_each_entry_rcu(ipvlan, &port->ipvlans, pnode) { in ipvlan_addr_busy()
258 list_for_each_entry_rcu(ipvlan, &port->ipvlans, pnode) { in ipvlan_process_multicast()
/drivers/gpu/drm/nouveau/nvkm/core/
Dmm.c112 u32 align, struct nvkm_mm_node **pnode) in nvkm_mm_head() argument
153 *pnode = this; in nvkm_mm_head()
187 u32 align, struct nvkm_mm_node **pnode) in nvkm_mm_tail() argument
232 *pnode = this; in nvkm_mm_tail()
/drivers/staging/mt7621-pci/
Dpci-mt7621.c331 struct device_node *pnode = dev->of_node; in mt7621_pcie_parse_port() local
340 err = of_address_to_resource(pnode, slot + 1, &regs); in mt7621_pcie_parse_port()