/drivers/scsi/elx/libefc/ |
D | efc_device.c | 16 efc_d_send_prli_rsp(struct efc_node *node, u16 ox_id) in efc_d_send_prli_rsp() argument 19 struct efc *efc = node->efc; in efc_d_send_prli_rsp() 21 node->ls_acc_oxid = ox_id; in efc_d_send_prli_rsp() 22 node->send_ls_acc = EFC_NODE_SEND_LS_ACC_PRLI; in efc_d_send_prli_rsp() 29 if (node->init) { in efc_d_send_prli_rsp() 31 node->display_name, node->wwpn, node->wwnn); in efc_d_send_prli_rsp() 32 if (node->nport->enable_tgt) in efc_d_send_prli_rsp() 33 rc = efc->tt.scsi_new_node(efc, node); in efc_d_send_prli_rsp() 37 efc_node_post_event(node, EFC_EVT_NODE_SESS_REG_FAIL, NULL); in efc_d_send_prli_rsp() 40 efc_node_post_event(node, EFC_EVT_NODE_SESS_REG_OK, NULL); in efc_d_send_prli_rsp() [all …]
|
D | efc_node.c | 14 struct efc_node *node = rnode->node; in efc_remote_node_cb() local 18 efc_node_post_event(node, event, NULL); in efc_remote_node_cb() 34 struct efc_node *node = container_of(arg, struct efc_node, ref); in _efc_node_free() local 35 struct efc *efc = node->efc; in _efc_node_free() 38 dma = &node->sparm_dma_buf; in _efc_node_free() 41 mempool_free(node, efc->node_pool); in _efc_node_free() 48 struct efc_node *node = NULL; in efc_node_alloc() local 58 node = mempool_alloc(efc->node_pool, GFP_ATOMIC); in efc_node_alloc() 59 if (!node) { in efc_node_alloc() 63 memset(node, 0, sizeof(*node)); in efc_node_alloc() [all …]
|
D | efc_fabric.c | 24 efc_fabric_initiate_shutdown(struct efc_node *node) in efc_fabric_initiate_shutdown() argument 26 struct efc *efc = node->efc; in efc_fabric_initiate_shutdown() 28 node->els_io_enabled = false; in efc_fabric_initiate_shutdown() 30 if (node->attached) { in efc_fabric_initiate_shutdown() 37 rc = efc_cmd_node_detach(efc, &node->rnode); in efc_fabric_initiate_shutdown() 39 node_printf(node, "Failed freeing HW node, rc=%d\n", in efc_fabric_initiate_shutdown() 47 efc_node_initiate_cleanup(node); in efc_fabric_initiate_shutdown() 54 struct efc_node *node = NULL; in __efc_fabric_common() local 56 node = ctx->app; in __efc_fabric_common() 62 node->shutdown_reason = EFC_NODE_SHUTDOWN_DEFAULT; in __efc_fabric_common() [all …]
|
D | efc_node.h | 19 io->node->display_name, io->instance_index, io->init_task_tag, \ 26 struct efc_node *node = ctx->app; in efc_node_evt_set() local 29 strncpy(node->current_state_name, handler, in efc_node_evt_set() 30 sizeof(node->current_state_name)); in efc_node_evt_set() 32 strncpy(node->prev_state_name, node->current_state_name, in efc_node_evt_set() 33 sizeof(node->prev_state_name)); in efc_node_evt_set() 34 strncpy(node->current_state_name, "invalid", in efc_node_evt_set() 35 sizeof(node->current_state_name)); in efc_node_evt_set() 37 node->prev_evt = node->current_evt; in efc_node_evt_set() 38 node->current_evt = evt; in efc_node_evt_set() [all …]
|
D | efc_els.c | 22 node->display_name, __func__); \ 26 efc_log_err((struct efc *)els->node->efc,\ 28 els->node->display_name,\ 35 efc_els_io_alloc(struct efc_node *node, u32 reqlen) in efc_els_io_alloc() argument 37 return efc_els_io_alloc_size(node, reqlen, EFC_ELS_RSP_LEN); in efc_els_io_alloc() 41 efc_els_io_alloc_size(struct efc_node *node, u32 reqlen, u32 rsplen) in efc_els_io_alloc_size() argument 47 efc = node->efc; in efc_els_io_alloc_size() 49 if (!node->els_io_enabled) { in efc_els_io_alloc_size() 65 els->node = node; in efc_els_io_alloc_size() 92 spin_lock_irqsave(&node->els_ios_lock, flags); in efc_els_io_alloc_size() [all …]
|
D | efc_els.h | 17 struct efc_node *node; member 33 efc_els_io_alloc(struct efc_node *node, u32 reqlen); 35 efc_els_io_alloc_size(struct efc_node *node, u32 reqlen, u32 rsplen); 39 typedef void (*els_cb_t)(struct efc_node *node, 42 efc_send_plogi(struct efc_node *node); 44 efc_send_flogi(struct efc_node *node); 46 efc_send_fdisc(struct efc_node *node); 48 efc_send_prli(struct efc_node *node); 50 efc_send_prlo(struct efc_node *node); 52 efc_send_logo(struct efc_node *node); [all …]
|
/drivers/net/wireguard/ |
D | allowedips.c | 23 static void copy_and_assign_cidr(struct allowedips_node *node, const u8 *src, in copy_and_assign_cidr() argument 26 node->cidr = cidr; in copy_and_assign_cidr() 27 node->bit_at_a = cidr / 8U; in copy_and_assign_cidr() 29 node->bit_at_a ^= (bits / 8U - 1U) % 8U; in copy_and_assign_cidr() 31 node->bit_at_b = 7U - (cidr % 8U); in copy_and_assign_cidr() 32 node->bitlen = bits; in copy_and_assign_cidr() 33 memcpy(node->bits, src, bits / 8U); in copy_and_assign_cidr() 36 static inline u8 choose(struct allowedips_node *node, const u8 *key) in choose() argument 38 return (key[node->bit_at_a] >> node->bit_at_b) & 1; in choose() 58 struct allowedips_node *node, *stack[MAX_ALLOWEDIPS_DEPTH] = { in root_free_rcu() local [all …]
|
/drivers/net/ethernet/marvell/octeontx2/nic/ |
D | qos.c | 44 static void otx2_qos_get_regaddr(struct otx2_qos_node *node, in otx2_qos_get_regaddr() argument 48 if (node->level == NIX_TXSCH_LVL_SMQ) { in otx2_qos_get_regaddr() 49 cfg->reg[index++] = NIX_AF_MDQX_PARENT(node->schq); in otx2_qos_get_regaddr() 50 cfg->reg[index++] = NIX_AF_MDQX_SCHEDULE(node->schq); in otx2_qos_get_regaddr() 51 cfg->reg[index++] = NIX_AF_MDQX_PIR(node->schq); in otx2_qos_get_regaddr() 52 cfg->reg[index] = NIX_AF_MDQX_CIR(node->schq); in otx2_qos_get_regaddr() 53 } else if (node->level == NIX_TXSCH_LVL_TL4) { in otx2_qos_get_regaddr() 54 cfg->reg[index++] = NIX_AF_TL4X_PARENT(node->schq); in otx2_qos_get_regaddr() 55 cfg->reg[index++] = NIX_AF_TL4X_SCHEDULE(node->schq); in otx2_qos_get_regaddr() 56 cfg->reg[index++] = NIX_AF_TL4X_PIR(node->schq); in otx2_qos_get_regaddr() [all …]
|
/drivers/gpu/drm/ |
D | drm_mm.c | 107 static noinline void save_stack(struct drm_mm_node *node) in save_stack() argument 115 node->stack = stack_depot_save(entries, n, GFP_NOWAIT); in save_stack() 120 struct drm_mm_node *node; in show_leaks() local 127 list_for_each_entry(node, drm_mm_nodes(mm), node_list) { in show_leaks() 128 if (!node->stack) { in show_leaks() 130 node->start, node->size); in show_leaks() 134 stack_depot_snprint(node->stack, buf, BUFSZ, 0); in show_leaks() 136 node->start, node->size, buf); in show_leaks() 145 static void save_stack(struct drm_mm_node *node) { } in save_stack() argument 149 #define START(node) ((node)->start) argument [all …]
|
/drivers/gpu/drm/i915/gt/ |
D | intel_gt_buffer_pool.c | 30 static void node_free(struct intel_gt_buffer_pool_node *node) in node_free() argument 32 i915_gem_object_put(node->obj); in node_free() 33 i915_active_fini(&node->active); in node_free() 34 kfree_rcu(node, rcu); in node_free() 39 struct intel_gt_buffer_pool_node *node, *stale = NULL; in pool_free_older_than() local 57 node = list_entry(pos, typeof(*node), link); in pool_free_older_than() 59 age = READ_ONCE(node->age); in pool_free_older_than() 64 if (!xchg(&node->age, 0)) in pool_free_older_than() 67 node->free = stale; in pool_free_older_than() 68 stale = node; in pool_free_older_than() [all …]
|
/drivers/net/ethernet/mellanox/mlx5/core/en/ |
D | htb.c | 36 struct mlx5e_qos_node *node = NULL; in mlx5e_htb_enumerate_leaves() local 39 hash_for_each(htb->qos_tc2node, bkt, node, hnode) { in mlx5e_htb_enumerate_leaves() 40 if (node->qid == MLX5E_QOS_QID_INNER) in mlx5e_htb_enumerate_leaves() 42 err = callback(data, node->qid, node->hw_id); in mlx5e_htb_enumerate_leaves() 73 struct mlx5e_qos_node *node; in mlx5e_htb_node_create_leaf() local 75 node = kzalloc(sizeof(*node), GFP_KERNEL); in mlx5e_htb_node_create_leaf() 76 if (!node) in mlx5e_htb_node_create_leaf() 79 node->parent = parent; in mlx5e_htb_node_create_leaf() 81 node->qid = qid; in mlx5e_htb_node_create_leaf() 84 node->classid = classid; in mlx5e_htb_node_create_leaf() [all …]
|
/drivers/base/test/ |
D | property-entry-test.c | 20 struct fwnode_handle *node; in pe_test_uints() local 27 node = fwnode_create_software_node(entries, NULL); in pe_test_uints() 28 KUNIT_ASSERT_NOT_ERR_OR_NULL(test, node); in pe_test_uints() 30 error = fwnode_property_count_u8(node, "prop-u8"); in pe_test_uints() 33 error = fwnode_property_read_u8(node, "prop-u8", &val_u8); in pe_test_uints() 37 error = fwnode_property_read_u8_array(node, "prop-u8", array_u8, 1); in pe_test_uints() 41 error = fwnode_property_read_u8_array(node, "prop-u8", array_u8, 2); in pe_test_uints() 44 error = fwnode_property_read_u8(node, "no-prop-u8", &val_u8); in pe_test_uints() 47 error = fwnode_property_read_u8_array(node, "no-prop-u8", array_u8, 1); in pe_test_uints() 50 error = fwnode_property_read_u16(node, "prop-u16", &val_u16); in pe_test_uints() [all …]
|
/drivers/interconnect/ |
D | core.c | 169 struct icc_node *node = dst; in path_init() local 182 node->provider->users++; in path_init() 183 hlist_add_head(&path->reqs[i].req_node, &node->req_list); in path_init() 184 path->reqs[i].node = node; in path_init() 188 node = node->reverse; in path_init() 200 struct icc_node *n, *node = NULL; in path_find() local 215 list_for_each_entry_safe(node, n, &traverse_list, search_list) { in path_find() 216 if (node == dst) { in path_find() 222 for (i = 0; i < node->num_links; i++) { in path_find() 223 struct icc_node *tmp = node->links[i]; in path_find() [all …]
|
/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/ |
D | gk20a.c | 176 struct gk20a_instobj *node = gk20a_instobj(memory); in gk20a_instobj_acquire_dma() local 177 struct gk20a_instmem *imem = node->imem; in gk20a_instobj_acquire_dma() 182 return node->vaddr; in gk20a_instobj_acquire_dma() 188 struct gk20a_instobj_iommu *node = gk20a_instobj_iommu(memory); in gk20a_instobj_acquire_iommu() local 189 struct gk20a_instmem *imem = node->base.imem; in gk20a_instobj_acquire_iommu() 197 if (node->base.vaddr) { in gk20a_instobj_acquire_iommu() 198 if (!node->use_cpt) { in gk20a_instobj_acquire_iommu() 200 list_del(&node->vaddr_node); in gk20a_instobj_acquire_iommu() 209 node->base.vaddr = vmap(node->pages, size >> PAGE_SHIFT, VM_MAP, in gk20a_instobj_acquire_iommu() 211 if (!node->base.vaddr) { in gk20a_instobj_acquire_iommu() [all …]
|
/drivers/acpi/acpica/ |
D | nsobject.c | 37 acpi_ns_attach_object(struct acpi_namespace_node *node, in acpi_ns_attach_object() argument 49 if (!node) { in acpi_ns_attach_object() 66 if (ACPI_GET_DESCRIPTOR_TYPE(node) != ACPI_DESC_TYPE_NAMED) { in acpi_ns_attach_object() 71 node, acpi_ut_get_descriptor_name(node))); in acpi_ns_attach_object() 77 if (node->object == object) { in acpi_ns_attach_object() 80 object, node)); in acpi_ns_attach_object() 119 obj_desc, node, acpi_ut_get_node_name(node))); in acpi_ns_attach_object() 123 if (node->object) { in acpi_ns_attach_object() 124 acpi_ns_detach_object(node); in acpi_ns_attach_object() 145 last_obj_desc->common.next_object = node->object; in acpi_ns_attach_object() [all …]
|
D | dsargs.c | 23 acpi_ds_execute_arguments(struct acpi_namespace_node *node, 43 acpi_ds_execute_arguments(struct acpi_namespace_node *node, in acpi_ds_execute_arguments() argument 62 op->common.node = scope_node; in acpi_ds_execute_arguments() 82 walk_state->deferred_node = node; in acpi_ds_execute_arguments() 93 op->common.node = node; in acpi_ds_execute_arguments() 103 op->common.node = scope_node; in acpi_ds_execute_arguments() 124 walk_state->deferred_node = node; in acpi_ds_execute_arguments() 149 struct acpi_namespace_node *node; in acpi_ds_get_buffer_field_arguments() local 161 node = obj_desc->buffer_field.node; in acpi_ds_get_buffer_field_arguments() 164 (ACPI_TYPE_BUFFER_FIELD, node, NULL)); in acpi_ds_get_buffer_field_arguments() [all …]
|
/drivers/firmware/efi/ |
D | dev-path-parser.c | 15 static long __init parse_acpi_path(const struct efi_dev_path *node, in parse_acpi_path() argument 24 if (node->header.length != 12) in parse_acpi_path() 28 'A' + ((node->acpi.hid >> 10) & 0x1f) - 1, in parse_acpi_path() 29 'A' + ((node->acpi.hid >> 5) & 0x1f) - 1, in parse_acpi_path() 30 'A' + ((node->acpi.hid >> 0) & 0x1f) - 1, in parse_acpi_path() 31 node->acpi.hid >> 16); in parse_acpi_path() 35 if (ret == 0 && node->acpi.uid == uid) in parse_acpi_path() 37 if (ret == -ENODATA && node->acpi.uid == 0) in parse_acpi_path() 60 static long __init parse_pci_path(const struct efi_dev_path *node, in parse_pci_path() argument 65 if (node->header.length != 6) in parse_pci_path() [all …]
|
/drivers/gpu/drm/tests/ |
D | drm_mm_test.c | 102 struct drm_mm_node *node, *check, *found; in assert_continuous() local 111 drm_mm_for_each_node(node, mm) { in assert_continuous() 112 if (node->start != addr) { in assert_continuous() 114 n, addr, node->start); in assert_continuous() 118 if (node->size != size) { in assert_continuous() 120 n, size, node->size); in assert_continuous() 124 if (drm_mm_hole_follows(node)) { in assert_continuous() 131 if (node != check) { in assert_continuous() 134 node->start, check->start); in assert_continuous() 151 static u64 misalignment(struct drm_mm_node *node, u64 alignment) in misalignment() argument [all …]
|
/drivers/acpi/arm64/ |
D | iort.c | 83 struct acpi_iort_node *node) in iort_get_fwnode() argument 90 if (curr->iort_node == node) { in iort_get_fwnode() 105 static inline void iort_delete_fwnode(struct acpi_iort_node *node) in iort_delete_fwnode() argument 111 if (curr->iort_node == node) { in iort_delete_fwnode() 146 (struct acpi_iort_node *node, void *context); 261 static acpi_status iort_match_node_callback(struct acpi_iort_node *node, in iort_match_node_callback() argument 267 if (node->type == ACPI_IORT_NODE_NAMED_COMPONENT) { in iort_match_node_callback() 297 ncomp = (struct acpi_iort_named_component *)node->node_data; in iort_match_node_callback() 301 } else if (node->type == ACPI_IORT_NODE_PCI_ROOT_COMPLEX) { in iort_match_node_callback() 306 pci_rc = (struct acpi_iort_root_complex *)node->node_data; in iort_match_node_callback() [all …]
|
/drivers/clk/ti/ |
D | dpll.c | 154 struct device_node *node) in _register_dpll() argument 163 clk = of_clk_get(node, 0); in _register_dpll() 166 node); in _register_dpll() 167 if (!ti_clk_retry_init(node, hw, _register_dpll)) in _register_dpll() 175 clk = of_clk_get(node, 1); in _register_dpll() 179 node); in _register_dpll() 180 if (!ti_clk_retry_init(node, hw, _register_dpll)) in _register_dpll() 189 name = ti_dt_clk_name(node); in _register_dpll() 190 clk = of_ti_clk_register_omap_hw(node, &clk_hw->hw, name); in _register_dpll() 193 of_clk_add_provider(node, of_clk_src_simple_get, clk); in _register_dpll() [all …]
|
/drivers/media/platform/st/sti/bdisp/ |
D | bdisp-hw.c | 433 if (ctx && ctx->node[0]) in bdisp_hw_free_nodes() 436 ctx->node[0], ctx->node_paddr[0], in bdisp_hw_free_nodes() 467 ctx->node[i] = base; in bdisp_hw_alloc_nodes() 469 dev_dbg(dev, "node[%d]=0x%p (paddr=%pad)\n", i, ctx->node[i], in bdisp_hw_alloc_nodes() 759 struct bdisp_node *node, in bdisp_hw_build_node() argument 774 memset(node, 0, sizeof(*node)); in bdisp_hw_build_node() 789 node->nip = 0; in bdisp_hw_build_node() 790 node->cic = BLT_CIC_ALL_GRP; in bdisp_hw_build_node() 791 node->ack = BLT_ACK_BYPASS_S2S3; in bdisp_hw_build_node() 796 node->ins = BLT_INS_S1_OFF | BLT_INS_S2_MEM | BLT_INS_S3_OFF; in bdisp_hw_build_node() [all …]
|
/drivers/gpu/drm/i915/ |
D | i915_scheduler.c | 19 node_to_request(const struct i915_sched_node *node) in node_to_request() argument 21 return container_of(node, const struct i915_request, sched); in node_to_request() 24 static inline bool node_started(const struct i915_sched_node *node) in node_started() argument 26 return i915_request_started(node_to_request(node)); in node_started() 29 static inline bool node_signaled(const struct i915_sched_node *node) in node_signaled() argument 31 return i915_request_completed(node_to_request(node)); in node_signaled() 36 return rb_entry(rb, struct i915_priolist, node); in to_priolist() 113 rb_link_node(&p->node, rb, parent); in i915_sched_lookup_priolist() 114 rb_insert_color_cached(&p->node, &sched_engine->queue, first); in i915_sched_lookup_priolist() 129 lock_sched_engine(struct i915_sched_node *node, in lock_sched_engine() argument [all …]
|
/drivers/gpu/drm/amd/amdkfd/ |
D | kfd_interrupt.c | 53 int kfd_interrupt_init(struct kfd_node *node) in kfd_interrupt_init() argument 57 r = kfifo_alloc(&node->ih_fifo, in kfd_interrupt_init() 58 KFD_IH_NUM_ENTRIES * node->kfd->device_info.ih_ring_entry_size, in kfd_interrupt_init() 61 dev_err(node->adev->dev, "Failed to allocate IH fifo\n"); in kfd_interrupt_init() 65 node->ih_wq = alloc_workqueue("KFD IH", WQ_HIGHPRI, 1); in kfd_interrupt_init() 66 if (unlikely(!node->ih_wq)) { in kfd_interrupt_init() 67 kfifo_free(&node->ih_fifo); in kfd_interrupt_init() 68 dev_err(node->adev->dev, "Failed to allocate KFD IH workqueue\n"); in kfd_interrupt_init() 71 spin_lock_init(&node->interrupt_lock); in kfd_interrupt_init() 73 INIT_WORK(&node->interrupt_work, interrupt_wq); in kfd_interrupt_init() [all …]
|
D | kfd_device.c | 483 static int kfd_gws_init(struct kfd_node *node) in kfd_gws_init() argument 486 struct kfd_dev *kfd = node->kfd; in kfd_gws_init() 487 uint32_t mes_rev = node->adev->mes.sched_version & AMDGPU_MES_VERSION_MASK; in kfd_gws_init() 489 if (node->dqm->sched_policy == KFD_SCHED_POLICY_NO_HWS) in kfd_gws_init() 492 if (hws_gws_support || (KFD_IS_SOC15(node) && in kfd_gws_init() 493 ((KFD_GC_VERSION(node) == IP_VERSION(9, 0, 1) in kfd_gws_init() 495 (KFD_GC_VERSION(node) <= IP_VERSION(9, 4, 0) in kfd_gws_init() 497 (KFD_GC_VERSION(node) == IP_VERSION(9, 4, 1) in kfd_gws_init() 499 (KFD_GC_VERSION(node) == IP_VERSION(9, 4, 2) in kfd_gws_init() 501 (KFD_GC_VERSION(node) == IP_VERSION(9, 4, 3)) || in kfd_gws_init() [all …]
|
/drivers/net/wireless/quantenna/qtnfmac/ |
D | util.c | 19 struct qtnf_sta_node *node; in qtnf_sta_list_lookup() local 24 list_for_each_entry(node, &list->head, list) { in qtnf_sta_list_lookup() 25 if (ether_addr_equal(node->mac_addr, mac)) in qtnf_sta_list_lookup() 26 return node; in qtnf_sta_list_lookup() 35 struct qtnf_sta_node *node; in qtnf_sta_list_lookup_index() local 40 list_for_each_entry(node, &list->head, list) { in qtnf_sta_list_lookup_index() 42 return node; in qtnf_sta_list_lookup_index() 52 struct qtnf_sta_node *node; in qtnf_sta_list_add() local 57 node = qtnf_sta_list_lookup(list, mac); in qtnf_sta_list_add() 59 if (node) in qtnf_sta_list_add() [all …]
|