Home
last modified time | relevance | path

Searched refs:node (Results 1 – 25 of 2070) sorted by relevance

12345678910>>...83

/drivers/scsi/elx/libefc/
Defc_device.c16 efc_d_send_prli_rsp(struct efc_node *node, u16 ox_id) in efc_d_send_prli_rsp() argument
19 struct efc *efc = node->efc; in efc_d_send_prli_rsp()
21 node->ls_acc_oxid = ox_id; in efc_d_send_prli_rsp()
22 node->send_ls_acc = EFC_NODE_SEND_LS_ACC_PRLI; in efc_d_send_prli_rsp()
29 if (node->init) { in efc_d_send_prli_rsp()
31 node->display_name, node->wwpn, node->wwnn); in efc_d_send_prli_rsp()
32 if (node->nport->enable_tgt) in efc_d_send_prli_rsp()
33 rc = efc->tt.scsi_new_node(efc, node); in efc_d_send_prli_rsp()
37 efc_node_post_event(node, EFC_EVT_NODE_SESS_REG_FAIL, NULL); in efc_d_send_prli_rsp()
40 efc_node_post_event(node, EFC_EVT_NODE_SESS_REG_OK, NULL); in efc_d_send_prli_rsp()
[all …]
Defc_node.c14 struct efc_node *node = rnode->node; in efc_remote_node_cb() local
18 efc_node_post_event(node, event, NULL); in efc_remote_node_cb()
34 struct efc_node *node = container_of(arg, struct efc_node, ref); in _efc_node_free() local
35 struct efc *efc = node->efc; in _efc_node_free()
38 dma = &node->sparm_dma_buf; in _efc_node_free()
41 mempool_free(node, efc->node_pool); in _efc_node_free()
48 struct efc_node *node = NULL; in efc_node_alloc() local
58 node = mempool_alloc(efc->node_pool, GFP_ATOMIC); in efc_node_alloc()
59 if (!node) { in efc_node_alloc()
63 memset(node, 0, sizeof(*node)); in efc_node_alloc()
[all …]
Defc_fabric.c24 efc_fabric_initiate_shutdown(struct efc_node *node) in efc_fabric_initiate_shutdown() argument
26 struct efc *efc = node->efc; in efc_fabric_initiate_shutdown()
28 node->els_io_enabled = false; in efc_fabric_initiate_shutdown()
30 if (node->attached) { in efc_fabric_initiate_shutdown()
37 rc = efc_cmd_node_detach(efc, &node->rnode); in efc_fabric_initiate_shutdown()
39 node_printf(node, "Failed freeing HW node, rc=%d\n", in efc_fabric_initiate_shutdown()
47 efc_node_initiate_cleanup(node); in efc_fabric_initiate_shutdown()
54 struct efc_node *node = NULL; in __efc_fabric_common() local
56 node = ctx->app; in __efc_fabric_common()
62 node->shutdown_reason = EFC_NODE_SHUTDOWN_DEFAULT; in __efc_fabric_common()
[all …]
Defc_node.h19 io->node->display_name, io->instance_index, io->init_task_tag, \
26 struct efc_node *node = ctx->app; in efc_node_evt_set() local
29 strncpy(node->current_state_name, handler, in efc_node_evt_set()
30 sizeof(node->current_state_name)); in efc_node_evt_set()
32 strncpy(node->prev_state_name, node->current_state_name, in efc_node_evt_set()
33 sizeof(node->prev_state_name)); in efc_node_evt_set()
34 strncpy(node->current_state_name, "invalid", in efc_node_evt_set()
35 sizeof(node->current_state_name)); in efc_node_evt_set()
37 node->prev_evt = node->current_evt; in efc_node_evt_set()
38 node->current_evt = evt; in efc_node_evt_set()
[all …]
Defc_els.c22 node->display_name, __func__); \
26 efc_log_err((struct efc *)els->node->efc,\
28 els->node->display_name,\
35 efc_els_io_alloc(struct efc_node *node, u32 reqlen) in efc_els_io_alloc() argument
37 return efc_els_io_alloc_size(node, reqlen, EFC_ELS_RSP_LEN); in efc_els_io_alloc()
41 efc_els_io_alloc_size(struct efc_node *node, u32 reqlen, u32 rsplen) in efc_els_io_alloc_size() argument
47 efc = node->efc; in efc_els_io_alloc_size()
49 if (!node->els_io_enabled) { in efc_els_io_alloc_size()
65 els->node = node; in efc_els_io_alloc_size()
92 spin_lock_irqsave(&node->els_ios_lock, flags); in efc_els_io_alloc_size()
[all …]
Defc_els.h17 struct efc_node *node; member
33 efc_els_io_alloc(struct efc_node *node, u32 reqlen);
35 efc_els_io_alloc_size(struct efc_node *node, u32 reqlen, u32 rsplen);
39 typedef void (*els_cb_t)(struct efc_node *node,
42 efc_send_plogi(struct efc_node *node);
44 efc_send_flogi(struct efc_node *node);
46 efc_send_fdisc(struct efc_node *node);
48 efc_send_prli(struct efc_node *node);
50 efc_send_prlo(struct efc_node *node);
52 efc_send_logo(struct efc_node *node);
[all …]
/drivers/net/wireguard/
Dallowedips.c23 static void copy_and_assign_cidr(struct allowedips_node *node, const u8 *src, in copy_and_assign_cidr() argument
26 node->cidr = cidr; in copy_and_assign_cidr()
27 node->bit_at_a = cidr / 8U; in copy_and_assign_cidr()
29 node->bit_at_a ^= (bits / 8U - 1U) % 8U; in copy_and_assign_cidr()
31 node->bit_at_b = 7U - (cidr % 8U); in copy_and_assign_cidr()
32 node->bitlen = bits; in copy_and_assign_cidr()
33 memcpy(node->bits, src, bits / 8U); in copy_and_assign_cidr()
36 static inline u8 choose(struct allowedips_node *node, const u8 *key) in choose() argument
38 return (key[node->bit_at_a] >> node->bit_at_b) & 1; in choose()
58 struct allowedips_node *node, *stack[MAX_ALLOWEDIPS_DEPTH] = { in root_free_rcu() local
[all …]
/drivers/net/ethernet/marvell/octeontx2/nic/
Dqos.c44 static void otx2_qos_get_regaddr(struct otx2_qos_node *node, in otx2_qos_get_regaddr() argument
48 if (node->level == NIX_TXSCH_LVL_SMQ) { in otx2_qos_get_regaddr()
49 cfg->reg[index++] = NIX_AF_MDQX_PARENT(node->schq); in otx2_qos_get_regaddr()
50 cfg->reg[index++] = NIX_AF_MDQX_SCHEDULE(node->schq); in otx2_qos_get_regaddr()
51 cfg->reg[index++] = NIX_AF_MDQX_PIR(node->schq); in otx2_qos_get_regaddr()
52 cfg->reg[index] = NIX_AF_MDQX_CIR(node->schq); in otx2_qos_get_regaddr()
53 } else if (node->level == NIX_TXSCH_LVL_TL4) { in otx2_qos_get_regaddr()
54 cfg->reg[index++] = NIX_AF_TL4X_PARENT(node->schq); in otx2_qos_get_regaddr()
55 cfg->reg[index++] = NIX_AF_TL4X_SCHEDULE(node->schq); in otx2_qos_get_regaddr()
56 cfg->reg[index++] = NIX_AF_TL4X_PIR(node->schq); in otx2_qos_get_regaddr()
[all …]
/drivers/gpu/drm/
Ddrm_mm.c107 static noinline void save_stack(struct drm_mm_node *node) in save_stack() argument
115 node->stack = stack_depot_save(entries, n, GFP_NOWAIT); in save_stack()
120 struct drm_mm_node *node; in show_leaks() local
127 list_for_each_entry(node, drm_mm_nodes(mm), node_list) { in show_leaks()
128 if (!node->stack) { in show_leaks()
130 node->start, node->size); in show_leaks()
134 stack_depot_snprint(node->stack, buf, BUFSZ, 0); in show_leaks()
136 node->start, node->size, buf); in show_leaks()
145 static void save_stack(struct drm_mm_node *node) { } in save_stack() argument
149 #define START(node) ((node)->start) argument
[all …]
/drivers/gpu/drm/i915/gt/
Dintel_gt_buffer_pool.c30 static void node_free(struct intel_gt_buffer_pool_node *node) in node_free() argument
32 i915_gem_object_put(node->obj); in node_free()
33 i915_active_fini(&node->active); in node_free()
34 kfree_rcu(node, rcu); in node_free()
39 struct intel_gt_buffer_pool_node *node, *stale = NULL; in pool_free_older_than() local
57 node = list_entry(pos, typeof(*node), link); in pool_free_older_than()
59 age = READ_ONCE(node->age); in pool_free_older_than()
64 if (!xchg(&node->age, 0)) in pool_free_older_than()
67 node->free = stale; in pool_free_older_than()
68 stale = node; in pool_free_older_than()
[all …]
/drivers/net/ethernet/mellanox/mlx5/core/en/
Dhtb.c36 struct mlx5e_qos_node *node = NULL; in mlx5e_htb_enumerate_leaves() local
39 hash_for_each(htb->qos_tc2node, bkt, node, hnode) { in mlx5e_htb_enumerate_leaves()
40 if (node->qid == MLX5E_QOS_QID_INNER) in mlx5e_htb_enumerate_leaves()
42 err = callback(data, node->qid, node->hw_id); in mlx5e_htb_enumerate_leaves()
73 struct mlx5e_qos_node *node; in mlx5e_htb_node_create_leaf() local
75 node = kzalloc(sizeof(*node), GFP_KERNEL); in mlx5e_htb_node_create_leaf()
76 if (!node) in mlx5e_htb_node_create_leaf()
79 node->parent = parent; in mlx5e_htb_node_create_leaf()
81 node->qid = qid; in mlx5e_htb_node_create_leaf()
84 node->classid = classid; in mlx5e_htb_node_create_leaf()
[all …]
/drivers/base/test/
Dproperty-entry-test.c20 struct fwnode_handle *node; in pe_test_uints() local
27 node = fwnode_create_software_node(entries, NULL); in pe_test_uints()
28 KUNIT_ASSERT_NOT_ERR_OR_NULL(test, node); in pe_test_uints()
30 error = fwnode_property_count_u8(node, "prop-u8"); in pe_test_uints()
33 error = fwnode_property_read_u8(node, "prop-u8", &val_u8); in pe_test_uints()
37 error = fwnode_property_read_u8_array(node, "prop-u8", array_u8, 1); in pe_test_uints()
41 error = fwnode_property_read_u8_array(node, "prop-u8", array_u8, 2); in pe_test_uints()
44 error = fwnode_property_read_u8(node, "no-prop-u8", &val_u8); in pe_test_uints()
47 error = fwnode_property_read_u8_array(node, "no-prop-u8", array_u8, 1); in pe_test_uints()
50 error = fwnode_property_read_u16(node, "prop-u16", &val_u16); in pe_test_uints()
[all …]
/drivers/interconnect/
Dcore.c169 struct icc_node *node = dst; in path_init() local
182 node->provider->users++; in path_init()
183 hlist_add_head(&path->reqs[i].req_node, &node->req_list); in path_init()
184 path->reqs[i].node = node; in path_init()
188 node = node->reverse; in path_init()
200 struct icc_node *n, *node = NULL; in path_find() local
215 list_for_each_entry_safe(node, n, &traverse_list, search_list) { in path_find()
216 if (node == dst) { in path_find()
222 for (i = 0; i < node->num_links; i++) { in path_find()
223 struct icc_node *tmp = node->links[i]; in path_find()
[all …]
/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/
Dgk20a.c176 struct gk20a_instobj *node = gk20a_instobj(memory); in gk20a_instobj_acquire_dma() local
177 struct gk20a_instmem *imem = node->imem; in gk20a_instobj_acquire_dma()
182 return node->vaddr; in gk20a_instobj_acquire_dma()
188 struct gk20a_instobj_iommu *node = gk20a_instobj_iommu(memory); in gk20a_instobj_acquire_iommu() local
189 struct gk20a_instmem *imem = node->base.imem; in gk20a_instobj_acquire_iommu()
197 if (node->base.vaddr) { in gk20a_instobj_acquire_iommu()
198 if (!node->use_cpt) { in gk20a_instobj_acquire_iommu()
200 list_del(&node->vaddr_node); in gk20a_instobj_acquire_iommu()
209 node->base.vaddr = vmap(node->pages, size >> PAGE_SHIFT, VM_MAP, in gk20a_instobj_acquire_iommu()
211 if (!node->base.vaddr) { in gk20a_instobj_acquire_iommu()
[all …]
/drivers/acpi/acpica/
Dnsobject.c37 acpi_ns_attach_object(struct acpi_namespace_node *node, in acpi_ns_attach_object() argument
49 if (!node) { in acpi_ns_attach_object()
66 if (ACPI_GET_DESCRIPTOR_TYPE(node) != ACPI_DESC_TYPE_NAMED) { in acpi_ns_attach_object()
71 node, acpi_ut_get_descriptor_name(node))); in acpi_ns_attach_object()
77 if (node->object == object) { in acpi_ns_attach_object()
80 object, node)); in acpi_ns_attach_object()
119 obj_desc, node, acpi_ut_get_node_name(node))); in acpi_ns_attach_object()
123 if (node->object) { in acpi_ns_attach_object()
124 acpi_ns_detach_object(node); in acpi_ns_attach_object()
145 last_obj_desc->common.next_object = node->object; in acpi_ns_attach_object()
[all …]
Ddsargs.c23 acpi_ds_execute_arguments(struct acpi_namespace_node *node,
43 acpi_ds_execute_arguments(struct acpi_namespace_node *node, in acpi_ds_execute_arguments() argument
62 op->common.node = scope_node; in acpi_ds_execute_arguments()
82 walk_state->deferred_node = node; in acpi_ds_execute_arguments()
93 op->common.node = node; in acpi_ds_execute_arguments()
103 op->common.node = scope_node; in acpi_ds_execute_arguments()
124 walk_state->deferred_node = node; in acpi_ds_execute_arguments()
149 struct acpi_namespace_node *node; in acpi_ds_get_buffer_field_arguments() local
161 node = obj_desc->buffer_field.node; in acpi_ds_get_buffer_field_arguments()
164 (ACPI_TYPE_BUFFER_FIELD, node, NULL)); in acpi_ds_get_buffer_field_arguments()
[all …]
/drivers/firmware/efi/
Ddev-path-parser.c15 static long __init parse_acpi_path(const struct efi_dev_path *node, in parse_acpi_path() argument
24 if (node->header.length != 12) in parse_acpi_path()
28 'A' + ((node->acpi.hid >> 10) & 0x1f) - 1, in parse_acpi_path()
29 'A' + ((node->acpi.hid >> 5) & 0x1f) - 1, in parse_acpi_path()
30 'A' + ((node->acpi.hid >> 0) & 0x1f) - 1, in parse_acpi_path()
31 node->acpi.hid >> 16); in parse_acpi_path()
35 if (ret == 0 && node->acpi.uid == uid) in parse_acpi_path()
37 if (ret == -ENODATA && node->acpi.uid == 0) in parse_acpi_path()
60 static long __init parse_pci_path(const struct efi_dev_path *node, in parse_pci_path() argument
65 if (node->header.length != 6) in parse_pci_path()
[all …]
/drivers/gpu/drm/tests/
Ddrm_mm_test.c102 struct drm_mm_node *node, *check, *found; in assert_continuous() local
111 drm_mm_for_each_node(node, mm) { in assert_continuous()
112 if (node->start != addr) { in assert_continuous()
114 n, addr, node->start); in assert_continuous()
118 if (node->size != size) { in assert_continuous()
120 n, size, node->size); in assert_continuous()
124 if (drm_mm_hole_follows(node)) { in assert_continuous()
131 if (node != check) { in assert_continuous()
134 node->start, check->start); in assert_continuous()
151 static u64 misalignment(struct drm_mm_node *node, u64 alignment) in misalignment() argument
[all …]
/drivers/acpi/arm64/
Diort.c83 struct acpi_iort_node *node) in iort_get_fwnode() argument
90 if (curr->iort_node == node) { in iort_get_fwnode()
105 static inline void iort_delete_fwnode(struct acpi_iort_node *node) in iort_delete_fwnode() argument
111 if (curr->iort_node == node) { in iort_delete_fwnode()
146 (struct acpi_iort_node *node, void *context);
261 static acpi_status iort_match_node_callback(struct acpi_iort_node *node, in iort_match_node_callback() argument
267 if (node->type == ACPI_IORT_NODE_NAMED_COMPONENT) { in iort_match_node_callback()
297 ncomp = (struct acpi_iort_named_component *)node->node_data; in iort_match_node_callback()
301 } else if (node->type == ACPI_IORT_NODE_PCI_ROOT_COMPLEX) { in iort_match_node_callback()
306 pci_rc = (struct acpi_iort_root_complex *)node->node_data; in iort_match_node_callback()
[all …]
/drivers/clk/ti/
Ddpll.c154 struct device_node *node) in _register_dpll() argument
163 clk = of_clk_get(node, 0); in _register_dpll()
166 node); in _register_dpll()
167 if (!ti_clk_retry_init(node, hw, _register_dpll)) in _register_dpll()
175 clk = of_clk_get(node, 1); in _register_dpll()
179 node); in _register_dpll()
180 if (!ti_clk_retry_init(node, hw, _register_dpll)) in _register_dpll()
189 name = ti_dt_clk_name(node); in _register_dpll()
190 clk = of_ti_clk_register_omap_hw(node, &clk_hw->hw, name); in _register_dpll()
193 of_clk_add_provider(node, of_clk_src_simple_get, clk); in _register_dpll()
[all …]
/drivers/media/platform/st/sti/bdisp/
Dbdisp-hw.c433 if (ctx && ctx->node[0]) in bdisp_hw_free_nodes()
436 ctx->node[0], ctx->node_paddr[0], in bdisp_hw_free_nodes()
467 ctx->node[i] = base; in bdisp_hw_alloc_nodes()
469 dev_dbg(dev, "node[%d]=0x%p (paddr=%pad)\n", i, ctx->node[i], in bdisp_hw_alloc_nodes()
759 struct bdisp_node *node, in bdisp_hw_build_node() argument
774 memset(node, 0, sizeof(*node)); in bdisp_hw_build_node()
789 node->nip = 0; in bdisp_hw_build_node()
790 node->cic = BLT_CIC_ALL_GRP; in bdisp_hw_build_node()
791 node->ack = BLT_ACK_BYPASS_S2S3; in bdisp_hw_build_node()
796 node->ins = BLT_INS_S1_OFF | BLT_INS_S2_MEM | BLT_INS_S3_OFF; in bdisp_hw_build_node()
[all …]
/drivers/gpu/drm/i915/
Di915_scheduler.c19 node_to_request(const struct i915_sched_node *node) in node_to_request() argument
21 return container_of(node, const struct i915_request, sched); in node_to_request()
24 static inline bool node_started(const struct i915_sched_node *node) in node_started() argument
26 return i915_request_started(node_to_request(node)); in node_started()
29 static inline bool node_signaled(const struct i915_sched_node *node) in node_signaled() argument
31 return i915_request_completed(node_to_request(node)); in node_signaled()
36 return rb_entry(rb, struct i915_priolist, node); in to_priolist()
113 rb_link_node(&p->node, rb, parent); in i915_sched_lookup_priolist()
114 rb_insert_color_cached(&p->node, &sched_engine->queue, first); in i915_sched_lookup_priolist()
129 lock_sched_engine(struct i915_sched_node *node, in lock_sched_engine() argument
[all …]
/drivers/gpu/drm/amd/amdkfd/
Dkfd_interrupt.c53 int kfd_interrupt_init(struct kfd_node *node) in kfd_interrupt_init() argument
57 r = kfifo_alloc(&node->ih_fifo, in kfd_interrupt_init()
58 KFD_IH_NUM_ENTRIES * node->kfd->device_info.ih_ring_entry_size, in kfd_interrupt_init()
61 dev_err(node->adev->dev, "Failed to allocate IH fifo\n"); in kfd_interrupt_init()
65 node->ih_wq = alloc_workqueue("KFD IH", WQ_HIGHPRI, 1); in kfd_interrupt_init()
66 if (unlikely(!node->ih_wq)) { in kfd_interrupt_init()
67 kfifo_free(&node->ih_fifo); in kfd_interrupt_init()
68 dev_err(node->adev->dev, "Failed to allocate KFD IH workqueue\n"); in kfd_interrupt_init()
71 spin_lock_init(&node->interrupt_lock); in kfd_interrupt_init()
73 INIT_WORK(&node->interrupt_work, interrupt_wq); in kfd_interrupt_init()
[all …]
Dkfd_device.c483 static int kfd_gws_init(struct kfd_node *node) in kfd_gws_init() argument
486 struct kfd_dev *kfd = node->kfd; in kfd_gws_init()
487 uint32_t mes_rev = node->adev->mes.sched_version & AMDGPU_MES_VERSION_MASK; in kfd_gws_init()
489 if (node->dqm->sched_policy == KFD_SCHED_POLICY_NO_HWS) in kfd_gws_init()
492 if (hws_gws_support || (KFD_IS_SOC15(node) && in kfd_gws_init()
493 ((KFD_GC_VERSION(node) == IP_VERSION(9, 0, 1) in kfd_gws_init()
495 (KFD_GC_VERSION(node) <= IP_VERSION(9, 4, 0) in kfd_gws_init()
497 (KFD_GC_VERSION(node) == IP_VERSION(9, 4, 1) in kfd_gws_init()
499 (KFD_GC_VERSION(node) == IP_VERSION(9, 4, 2) in kfd_gws_init()
501 (KFD_GC_VERSION(node) == IP_VERSION(9, 4, 3)) || in kfd_gws_init()
[all …]
/drivers/net/wireless/quantenna/qtnfmac/
Dutil.c19 struct qtnf_sta_node *node; in qtnf_sta_list_lookup() local
24 list_for_each_entry(node, &list->head, list) { in qtnf_sta_list_lookup()
25 if (ether_addr_equal(node->mac_addr, mac)) in qtnf_sta_list_lookup()
26 return node; in qtnf_sta_list_lookup()
35 struct qtnf_sta_node *node; in qtnf_sta_list_lookup_index() local
40 list_for_each_entry(node, &list->head, list) { in qtnf_sta_list_lookup_index()
42 return node; in qtnf_sta_list_lookup_index()
52 struct qtnf_sta_node *node; in qtnf_sta_list_add() local
57 node = qtnf_sta_list_lookup(list, mac); in qtnf_sta_list_add()
59 if (node) in qtnf_sta_list_add()
[all …]

12345678910>>...83