/drivers/scsi/lpfc/ |
D | lpfc_scsi.c | 147 struct lpfc_nodelist *pnode; in lpfc_update_stats() local 162 pnode = rdata->pnode; in lpfc_update_stats() 165 if (!pnode || in lpfc_update_stats() 166 !pnode->lat_data || in lpfc_update_stats() 187 pnode->lat_data[i].cmd_count++; in lpfc_update_stats() 487 if (psb->rdata && psb->rdata->pnode && in lpfc_sli4_vport_delete_fcp_xri_aborted() 488 psb->rdata->pnode->vport == vport) in lpfc_sli4_vport_delete_fcp_xri_aborted() 541 if (psb->rdata && psb->rdata->pnode) in lpfc_sli4_io_xri_aborted() 542 ndlp = psb->rdata->pnode; in lpfc_sli4_io_xri_aborted() 1023 if (rdata && rdata->pnode) { in lpfc_bg_err_inject() [all …]
|
D | lpfc_scsi.h | 43 struct lpfc_nodelist *pnode; /* Pointer to the node structure. */ member
|
D | lpfc_nvme.c | 1198 struct lpfc_nodelist *pnode, in lpfc_nvme_prep_io_cmd() argument 1207 if (!NLP_CHK_NODE_ACT(pnode)) in lpfc_nvme_prep_io_cmd() 1226 (pnode->nlp_flag & NLP_FIRSTBURST)) { in lpfc_nvme_prep_io_cmd() 1228 if (req_len < pnode->nvme_fb_size) in lpfc_nvme_prep_io_cmd() 1233 pnode->nvme_fb_size; in lpfc_nvme_prep_io_cmd() 1259 if (pnode->nlp_nvme_info & NLP_NVME_NSLER) in lpfc_nvme_prep_io_cmd() 1272 phba->sli4_hba.rpi_ids[pnode->nlp_rpi]); in lpfc_nvme_prep_io_cmd()
|
D | lpfc_hbadisc.c | 84 ndlp = rdata->pnode; in lpfc_terminate_rport_io() 124 ndlp = rdata->pnode; in lpfc_dev_loss_tmo_callbk() 144 put_node = rdata->pnode != NULL; in lpfc_dev_loss_tmo_callbk() 146 rdata->pnode = NULL; in lpfc_dev_loss_tmo_callbk() 261 put_node = rdata->pnode != NULL; in lpfc_dev_loss_tmo_handler() 262 rdata->pnode = NULL; in lpfc_dev_loss_tmo_handler() 282 put_node = rdata->pnode != NULL; in lpfc_dev_loss_tmo_handler() 283 rdata->pnode = NULL; in lpfc_dev_loss_tmo_handler() 4128 if (rdata->pnode == ndlp) in lpfc_register_remote_port() 4130 rdata->pnode = NULL; in lpfc_register_remote_port() [all …]
|
D | lpfc_els.c | 1768 if (rdata->pnode == ndlp) { in lpfc_plogi_confirm_nport() 1772 rdata->pnode = lpfc_nlp_get(new_ndlp); in lpfc_plogi_confirm_nport() 1849 put_node = rdata->pnode != NULL; in lpfc_plogi_confirm_nport() 1851 rdata->pnode = NULL; in lpfc_plogi_confirm_nport()
|
D | lpfc_bsg.c | 393 struct lpfc_nodelist *ndlp = rdata->pnode; in lpfc_bsg_send_mgmt_cmd() 667 struct lpfc_nodelist *ndlp = rdata->pnode; in lpfc_bsg_rport_els()
|
D | lpfc_sli.c | 1130 ndlp = lpfc_cmd->rdata->pnode; in __lpfc_sli_get_els_sglq() 11341 if ((lpfc_cmd->rdata) && (lpfc_cmd->rdata->pnode) && in lpfc_sli_validate_fcp_iocb() 11342 (lpfc_cmd->rdata->pnode->nlp_sid == tgt_id) && in lpfc_sli_validate_fcp_iocb() 11347 if ((lpfc_cmd->rdata) && (lpfc_cmd->rdata->pnode) && in lpfc_sli_validate_fcp_iocb() 11348 (lpfc_cmd->rdata->pnode->nlp_sid == tgt_id)) in lpfc_sli_validate_fcp_iocb() 11646 ndlp = lpfc_cmd->rdata->pnode; in lpfc_sli_abort_taskmgmt()
|
D | lpfc_attr.c | 6861 struct lpfc_nodelist *ndlp = rdata->pnode; in lpfc_set_rport_loss_tmo()
|
/drivers/macintosh/ |
D | macio_asic.c | 454 struct device_node *np, *pnode; in macio_pci_add_devices() local 466 pnode = of_node_get(chip->of_node); in macio_pci_add_devices() 467 if (pnode == NULL) in macio_pci_add_devices() 471 rdev = macio_add_one_device(chip, parent, pnode, NULL, root_res); in macio_pci_add_devices() 477 for (np = NULL; (np = of_get_next_child(pnode, np)) != NULL;) { in macio_pci_add_devices() 493 pnode = mbdev->ofdev.dev.of_node; in macio_pci_add_devices() 494 for (np = NULL; (np = of_get_next_child(pnode, np)) != NULL;) { in macio_pci_add_devices() 506 pnode = sdev->ofdev.dev.of_node; in macio_pci_add_devices() 507 for (np = NULL; (np = of_get_next_child(pnode, np)) != NULL;) { in macio_pci_add_devices()
|
/drivers/base/power/ |
D | qos.c | 162 &req->data.pnode, action, value); in apply_constraint() 166 &req->data.pnode, action, value); in apply_constraint() 273 plist_for_each_entry_safe(req, tmp, &c->list, data.pnode) { in dev_pm_qos_constraints_destroy() 283 plist_for_each_entry_safe(req, tmp, &c->list, data.pnode) { in dev_pm_qos_constraints_destroy() 289 plist_for_each_entry_safe(req, tmp, &c->list, data.freq.pnode) { in dev_pm_qos_constraints_destroy() 296 plist_for_each_entry_safe(req, tmp, &c->list, data.freq.pnode) { in dev_pm_qos_constraints_destroy() 424 curr_value = req->data.pnode.prio; in __dev_pm_qos_update_request() 428 curr_value = req->data.freq.pnode.prio; in __dev_pm_qos_update_request() 894 dev->power.qos->latency_tolerance_req->data.pnode.prio; in dev_pm_qos_get_user_latency_tolerance()
|
/drivers/clk/st/ |
D | clk-flexgen.c | 312 struct device_node *pnode; in st_of_flexgen_setup() local 324 pnode = of_get_parent(np); in st_of_flexgen_setup() 325 if (!pnode) in st_of_flexgen_setup() 328 reg = of_iomap(pnode, 0); in st_of_flexgen_setup() 329 of_node_put(pnode); in st_of_flexgen_setup()
|
D | clkgen-pll.c | 624 struct device_node *pnode; in clkgen_get_register_base() local 627 pnode = of_get_parent(np); in clkgen_get_register_base() 628 if (!pnode) in clkgen_get_register_base() 631 reg = of_iomap(pnode, 0); in clkgen_get_register_base() 633 of_node_put(pnode); in clkgen_get_register_base()
|
/drivers/misc/sgi-gru/ |
D | gruhandles.h | 143 static inline unsigned long gru_chiplet_paddr(unsigned long paddr, int pnode, in gru_chiplet_paddr() argument 146 return paddr + GRU_SIZE * (2 * pnode + chiplet); in gru_chiplet_paddr() 149 static inline void *gru_chiplet_vaddr(void *vaddr, int pnode, int chiplet) in gru_chiplet_vaddr() argument 151 return vaddr + GRU_SIZE * (2 * pnode + chiplet); in gru_chiplet_vaddr()
|
D | grufile.c | 253 int pnode, nid, bid, chip; in gru_init_tables() local 264 pnode = uv_blade_to_pnode(bid); in gru_init_tables() 280 paddr = gru_chiplet_paddr(gru_base_paddr, pnode, chip); in gru_init_tables() 281 vaddr = gru_chiplet_vaddr(gru_base_vaddr, pnode, chip); in gru_init_tables()
|
/drivers/net/ipvlan/ |
D | ipvlan_main.c | 16 list_for_each_entry(ipvlan, &port->ipvlans, pnode) { in ipvlan_set_port_mode() 45 list_for_each_entry_continue_reverse(ipvlan, &port->ipvlans, pnode) { in ipvlan_set_port_mode() 616 list_add_tail_rcu(&ipvlan->pnode, &port->ipvlans); in ipvlan_link_new() 644 list_del_rcu(&ipvlan->pnode); in ipvlan_link_delete() 709 list_for_each_entry(ipvlan, &port->ipvlans, pnode) in ipvlan_device_event() 730 list_for_each_entry_safe(ipvlan, next, &port->ipvlans, pnode) in ipvlan_device_event() 737 list_for_each_entry(ipvlan, &port->ipvlans, pnode) { in ipvlan_device_event() 746 list_for_each_entry(ipvlan, &port->ipvlans, pnode) in ipvlan_device_event() 752 list_for_each_entry(ipvlan, &port->ipvlans, pnode) { in ipvlan_device_event() 762 list_for_each_entry(ipvlan, &port->ipvlans, pnode) { in ipvlan_device_event()
|
D | ipvlan.h | 64 struct list_head pnode; member
|
D | ipvlan_core.c | 126 list_for_each_entry_rcu(ipvlan, &port->ipvlans, pnode) { in ipvlan_addr_busy() 258 list_for_each_entry_rcu(ipvlan, &port->ipvlans, pnode) { in ipvlan_process_multicast()
|
/drivers/gpu/drm/nouveau/nvkm/core/ |
D | mm.c | 112 u32 align, struct nvkm_mm_node **pnode) in nvkm_mm_head() argument 153 *pnode = this; in nvkm_mm_head() 187 u32 align, struct nvkm_mm_node **pnode) in nvkm_mm_tail() argument 232 *pnode = this; in nvkm_mm_tail()
|
/drivers/staging/mt7621-pci/ |
D | pci-mt7621.c | 331 struct device_node *pnode = dev->of_node; in mt7621_pcie_parse_port() local 340 err = of_address_to_resource(pnode, slot + 1, ®s); in mt7621_pcie_parse_port()
|