Home
last modified time | relevance | path

Searched refs:nodes (Results 1 – 25 of 93) sorted by relevance

1234

/drivers/gpu/drm/selftests/
Dtest-drm_mm.c268 struct drm_mm_node nodes[2]; in igt_debug() local
277 memset(nodes, 0, sizeof(nodes)); in igt_debug()
278 nodes[0].start = 512; in igt_debug()
279 nodes[0].size = 1024; in igt_debug()
280 ret = drm_mm_reserve_node(&mm, &nodes[0]); in igt_debug()
283 nodes[0].start, nodes[0].size); in igt_debug()
287 nodes[1].size = 1024; in igt_debug()
288 nodes[1].start = 4096 - 512 - nodes[1].size; in igt_debug()
289 ret = drm_mm_reserve_node(&mm, &nodes[1]); in igt_debug()
292 nodes[1].start, nodes[1].size); in igt_debug()
[all …]
/drivers/md/persistent-data/
Ddm-btree-spine.c131 s->nodes[0] = NULL; in init_ro_spine()
132 s->nodes[1] = NULL; in init_ro_spine()
140 unlock_block(s->info, s->nodes[i]); in exit_ro_spine()
149 unlock_block(s->info, s->nodes[0]); in ro_step()
150 s->nodes[0] = s->nodes[1]; in ro_step()
154 r = bn_read_lock(s->info, new_child, s->nodes + s->count); in ro_step()
165 unlock_block(s->info, s->nodes[s->count]); in ro_pop()
173 block = s->nodes[s->count - 1]; in ro_node()
191 unlock_block(s->info, s->nodes[i]); in exit_shadow_spine()
203 unlock_block(s->info, s->nodes[0]); in shadow_step()
[all …]
/drivers/interconnect/qcom/
Dsc7180.c161 .nodes = { &ebi },
168 .nodes = { &ebi },
175 .nodes = { &qns_llcc },
182 .nodes = { &qns_mem_noc_hf },
189 .nodes = { &qxm_crypto },
196 .nodes = { &qnm_snoc,
251 .nodes = { &qxm_camnoc_hf0_uncomp,
266 .nodes = { &acm_sys_tcu },
273 .nodes = { &qns_mem_noc_sf },
280 .nodes = { &qup_core_master_1, &qup_core_master_2 },
[all …]
Dosm-l3.c68 const struct qcom_icc_node **nodes; member
93 .nodes = sdm845_osm_l3_nodes,
109 .nodes = sc7180_osm_l3_nodes,
125 .nodes = sm8150_osm_l3_nodes,
141 .nodes = sm8250_epss_l3_nodes,
163 list_for_each_entry(n, &provider->nodes, node_list) in qcom_icc_set()
258 qnodes = desc->nodes; in qcom_osm_l3_probe()
261 data = devm_kzalloc(&pdev->dev, struct_size(data, nodes, num_nodes), GFP_KERNEL); in qcom_osm_l3_probe()
270 INIT_LIST_HEAD(&provider->nodes); in qcom_osm_l3_probe()
296 data->nodes[i] = node; in qcom_osm_l3_probe()
Dsm8150.c206 .nodes = aggre1_noc_nodes,
241 .nodes = aggre2_noc_nodes,
259 .nodes = camnoc_virt_nodes,
276 .nodes = compute_noc_nodes,
344 .nodes = config_noc_nodes,
360 .nodes = dc_noc_nodes,
395 .nodes = gem_noc_nodes,
411 .nodes = ipa_virt_nodes,
428 .nodes = mc_virt_nodes,
458 .nodes = mmss_noc_nodes,
[all …]
Dsm8250.c220 .nodes = aggre1_noc_nodes,
250 .nodes = aggre2_noc_nodes,
267 .nodes = compute_noc_nodes,
333 .nodes = config_noc_nodes,
349 .nodes = dc_noc_nodes,
383 .nodes = gem_noc_nodes,
399 .nodes = ipa_virt_nodes,
416 .nodes = mc_virt_nodes,
446 .nodes = mmss_noc_nodes,
472 .nodes = npu_noc_nodes,
[all …]
Dmsm8974.c223 struct msm8974_icc_node **nodes; member
258 .nodes = msm8974_bimc_nodes,
341 .nodes = msm8974_cnoc_nodes,
394 .nodes = msm8974_mnoc_nodes,
429 .nodes = msm8974_onoc_nodes,
492 .nodes = msm8974_pnoc_nodes,
549 .nodes = msm8974_snoc_nodes,
596 list_for_each_entry(n, &provider->nodes, node_list) in msm8974_icc_set()
668 qnodes = desc->nodes; in msm8974_icc_probe()
675 data = devm_kzalloc(dev, struct_size(data, nodes, num_nodes), in msm8974_icc_probe()
[all …]
Dqcs404.c143 struct qcom_icc_node **nodes; member
244 .nodes = qcs404_bimc_nodes,
298 .nodes = qcs404_pcnoc_nodes,
326 .nodes = qcs404_snoc_nodes,
347 list_for_each_entry(n, &provider->nodes, node_list) in qcom_icc_set()
420 qnodes = desc->nodes; in qnoc_probe()
427 data = devm_kzalloc(dev, struct_size(data, nodes, num_nodes), in qnoc_probe()
447 INIT_LIST_HEAD(&provider->nodes); in qnoc_probe()
480 data->nodes[i] = node; in qnoc_probe()
Dmsm8916.c155 struct msm8916_icc_node **nodes; member
287 .nodes = msm8916_snoc_nodes,
304 .nodes = msm8916_bimc_nodes,
362 .nodes = msm8916_pcnoc_nodes,
380 list_for_each_entry(n, &provider->nodes, node_list) in msm8916_icc_set()
453 qnodes = desc->nodes; in msm8916_qnoc_probe()
460 data = devm_kzalloc(dev, struct_size(data, nodes, num_nodes), in msm8916_qnoc_probe()
480 INIT_LIST_HEAD(&provider->nodes); in msm8916_qnoc_probe()
510 data->nodes[i] = node; in msm8916_qnoc_probe()
Dsdm845.c196 .nodes = aggre1_noc_nodes,
224 .nodes = aggre2_noc_nodes,
285 .nodes = config_noc_nodes,
301 .nodes = dc_noc_nodes,
319 .nodes = gladiator_noc_nodes,
354 .nodes = mem_noc_nodes,
388 .nodes = mmss_noc_nodes,
434 .nodes = system_noc_nodes,
455 qnodes = desc->nodes; in qnoc_probe()
462 data = devm_kzalloc(&pdev->dev, struct_size(data, nodes, num_nodes), in qnoc_probe()
[all …]
/drivers/gpu/drm/amd/amdgpu/
Damdgpu_vram_mgr.c261 struct drm_mm_node *nodes = mem->mm_node; in amdgpu_vram_mgr_bo_visible_size() local
271 for (usage = 0; nodes && pages; pages -= nodes->size, nodes++) in amdgpu_vram_mgr_bo_visible_size()
272 usage += amdgpu_vram_mgr_vis_size(adev, nodes); in amdgpu_vram_mgr_bo_visible_size()
317 struct drm_mm_node *nodes; in amdgpu_vram_mgr_new() local
353 nodes = kvmalloc_array((uint32_t)num_nodes, sizeof(*nodes), in amdgpu_vram_mgr_new()
355 if (!nodes) { in amdgpu_vram_mgr_new()
371 r = drm_mm_insert_node_in_range(mm, &nodes[i], pages, in amdgpu_vram_mgr_new()
378 vis_usage += amdgpu_vram_mgr_vis_size(adev, &nodes[i]); in amdgpu_vram_mgr_new()
379 amdgpu_vram_mgr_virt_start(mem, &nodes[i]); in amdgpu_vram_mgr_new()
390 r = drm_mm_insert_node_in_range(mm, &nodes[i], in amdgpu_vram_mgr_new()
[all …]
Dta_xgmi_if.h92 struct ta_xgmi_node_info nodes[TA_XGMI__MAX_CONNECTED_NODES]; member
97 struct ta_xgmi_node_info nodes[TA_XGMI__MAX_CONNECTED_NODES]; member
102 struct ta_xgmi_node_info nodes[TA_XGMI__MAX_CONNECTED_NODES]; member
/drivers/clk/zynqmp/
Dclk-zynqmp.h39 const struct clock_topology *nodes);
44 const struct clock_topology *nodes);
50 const struct clock_topology *nodes);
55 const struct clock_topology *nodes);
61 const struct clock_topology *nodes);
Ddivider.c247 const struct clock_topology *nodes) in zynqmp_clk_register_divider() argument
262 init.flags = nodes->flag & ~CLK_FRAC; in zynqmp_clk_register_divider()
267 div->is_frac = !!((nodes->flag & CLK_FRAC) | in zynqmp_clk_register_divider()
268 (nodes->custom_type_flag & CUSTOM_FLAG_CLK_FRAC)); in zynqmp_clk_register_divider()
269 div->flags = nodes->type_flag; in zynqmp_clk_register_divider()
272 div->div_type = nodes->type; in zynqmp_clk_register_divider()
278 div->max_div = zynqmp_clk_get_max_divisor(clk_id, nodes->type); in zynqmp_clk_register_divider()
Dclk-mux-zynqmp.c107 const struct clock_topology *nodes) in zynqmp_clk_register_mux() argument
119 if (nodes->type_flag & CLK_MUX_READ_ONLY) in zynqmp_clk_register_mux()
123 init.flags = nodes->flag; in zynqmp_clk_register_mux()
126 mux->flags = nodes->type_flag; in zynqmp_clk_register_mux()
Dclkc.c124 const struct clock_topology *nodes)
288 const struct clock_topology *nodes) in zynqmp_clk_register_fixed_factor() argument
308 nodes->flag, mult, in zynqmp_clk_register_fixed_factor()
565 struct clock_topology *nodes; in zynqmp_register_clk_topology() local
568 nodes = clock[clk_id].node; in zynqmp_register_clk_topology()
579 clk_type_postfix[nodes[j].type]); in zynqmp_register_clk_topology()
584 if (!clk_topology[nodes[j].type]) in zynqmp_register_clk_topology()
587 hw = (*clk_topology[nodes[j].type])(clk_out[j], clk_dev_id, in zynqmp_register_clk_topology()
590 &nodes[j]); in zynqmp_register_clk_topology()
/drivers/dma/
Duniphier-xdmac.c83 struct uniphier_xdmac_desc_node nodes[]; member
140 src_addr = xd->nodes[xd->cur_node].src; in uniphier_xdmac_chan_start()
141 dst_addr = xd->nodes[xd->cur_node].dst; in uniphier_xdmac_chan_start()
142 its = xd->nodes[xd->cur_node].burst_size; in uniphier_xdmac_chan_start()
143 tnum = xd->nodes[xd->cur_node].nr_burst; in uniphier_xdmac_chan_start()
295 xd = kzalloc(struct_size(xd, nodes, nr), GFP_NOWAIT); in uniphier_xdmac_prep_dma_memcpy()
301 xd->nodes[i].src = src; in uniphier_xdmac_prep_dma_memcpy()
302 xd->nodes[i].dst = dst; in uniphier_xdmac_prep_dma_memcpy()
303 xd->nodes[i].burst_size = burst_size; in uniphier_xdmac_prep_dma_memcpy()
304 xd->nodes[i].nr_burst = len / burst_size; in uniphier_xdmac_prep_dma_memcpy()
[all …]
/drivers/interconnect/imx/
Dimx.c169 list_for_each_entry_safe(node, tmp, &provider->nodes, node_list) in imx_icc_unregister_nodes()
192 provider_data->nodes[node->id] = node; in imx_icc_register_nodes()
212 static int get_max_node_id(struct imx_icc_node_desc *nodes, int nodes_count) in get_max_node_id() argument
217 if (nodes[i].id > ret) in get_max_node_id()
218 ret = nodes[i].id; in get_max_node_id()
224 struct imx_icc_node_desc *nodes, int nodes_count) in imx_icc_register() argument
233 num_nodes = get_max_node_id(nodes, nodes_count) + 1; in imx_icc_register()
234 data = devm_kzalloc(dev, struct_size(data, nodes, num_nodes), in imx_icc_register()
256 ret = imx_icc_register_nodes(provider, nodes, nodes_count); in imx_icc_register()
Dimx8mn.c32 static struct imx_icc_node_desc nodes[] = { variable
75 return imx_icc_register(pdev, nodes, ARRAY_SIZE(nodes)); in imx8mn_icc_probe()
Dimx8mm.c35 static struct imx_icc_node_desc nodes[] = { variable
86 return imx_icc_register(pdev, nodes, ARRAY_SIZE(nodes)); in imx8mm_icc_probe()
Dimx8mq.c33 static struct imx_icc_node_desc nodes[] = { variable
85 return imx_icc_register(pdev, nodes, ARRAY_SIZE(nodes)); in imx8mq_icc_probe()
/drivers/gpu/drm/nouveau/nvkm/core/
Dmm.c26 #define node(root, dir) ((root)->nl_entry.dir == &mm->nodes) ? NULL : \
36 list_for_each_entry(node, &mm->nodes, nl_entry) { in nvkm_mm_dump()
246 prev = list_last_entry(&mm->nodes, typeof(*node), nl_entry); in nvkm_mm_init()
255 list_add_tail(&node->nl_entry, &mm->nodes); in nvkm_mm_init()
259 INIT_LIST_HEAD(&mm->nodes); in nvkm_mm_init()
275 list_add_tail(&node->nl_entry, &mm->nodes); in nvkm_mm_init()
286 int nodes = 0; in nvkm_mm_fini() local
291 list_for_each_entry(node, &mm->nodes, nl_entry) { in nvkm_mm_fini()
293 if (++nodes > mm->heap_nodes) { in nvkm_mm_fini()
300 list_for_each_entry_safe(node, temp, &mm->nodes, nl_entry) { in nvkm_mm_fini()
/drivers/staging/media/ipu3/
Dipu3.c120 if (!imgu_pipe->nodes[IMGU_NODE_VF].enabled && in imgu_dummybufs_init()
128 meta = &imgu_pipe->nodes[node].vdev_fmt.fmt.meta; in imgu_dummybufs_init()
129 mpix = &imgu_pipe->nodes[node].vdev_fmt.fmt.pix_mp; in imgu_dummybufs_init()
214 list_for_each_entry(buf, &imgu_pipe->nodes[node].buffers, vid_buf.list) { in imgu_queue_getbuf()
249 !imgu_pipe->nodes[IMGU_NODE_VF].enabled) { in imgu_queue_buffers()
254 imgu_pipe->nodes[node].enabled) { in imgu_queue_buffers()
259 if (list_empty(&imgu_pipe->nodes[node].buffers)) in imgu_queue_buffers()
262 ivb = list_first_entry(&imgu_pipe->nodes[node].buffers, in imgu_queue_buffers()
330 &imgu_pipe->nodes[node].buffers, in imgu_queue_buffers()
354 fmt = &imgu->imgu_pipe[pipe].nodes[IMGU_NODE_IN].pad_fmt; in imgu_powerup()
[all …]
Dipu3-v4l2.c32 imgu_pipe->nodes[IMGU_NODE_IN].vdev_fmt.fmt.pix_mp.width; in imgu_subdev_open()
34 imgu_pipe->nodes[IMGU_NODE_IN].vdev_fmt.fmt.pix_mp.height; in imgu_subdev_open()
43 try_fmt->code = imgu_pipe->nodes[i].pad_fmt.code; in imgu_subdev_open()
79 imgu_pipe->queue_enabled[i] = imgu_pipe->nodes[i].enabled; in imgu_subdev_s_stream()
91 &imgu_pipe->nodes[node].vdev_fmt.fmt.pix_mp : NULL; in imgu_subdev_s_stream()
96 if (imgu_pipe->nodes[IMGU_NODE_VF].enabled) in imgu_subdev_s_stream()
137 fmt->format = imgu_pipe->nodes[pad].pad_fmt; in imgu_subdev_get_fmt()
166 mf = &imgu_pipe->nodes[pad].pad_fmt; in imgu_subdev_set_fmt()
285 imgu_pipe->nodes[pad].enabled = flags & MEDIA_LNK_FL_ENABLED; in imgu_link_setup()
431 node = &imgu->imgu_pipe[p].nodes[i]; in imgu_all_nodes_streaming()
[all …]
/drivers/base/test/
Dproperty-entry-test.c366 static const struct software_node nodes[] = { in pe_test_reference() local
374 .node = &nodes[0], in pe_test_reference()
378 .node = &nodes[1], in pe_test_reference()
385 PROPERTY_ENTRY_REF("ref-1", &nodes[0]), in pe_test_reference()
386 PROPERTY_ENTRY_REF("ref-2", &nodes[1], 1, 2), in pe_test_reference()
395 error = software_node_register_nodes(nodes); in pe_test_reference()
404 KUNIT_EXPECT_PTR_EQ(test, to_software_node(ref.fwnode), &nodes[0]); in pe_test_reference()
415 KUNIT_EXPECT_PTR_EQ(test, to_software_node(ref.fwnode), &nodes[1]); in pe_test_reference()
423 KUNIT_EXPECT_PTR_EQ(test, to_software_node(ref.fwnode), &nodes[1]); in pe_test_reference()
438 KUNIT_EXPECT_PTR_EQ(test, to_software_node(ref.fwnode), &nodes[0]); in pe_test_reference()
[all …]

1234