/drivers/gpu/drm/selftests/ |
D | test-drm_mm.c | 268 struct drm_mm_node nodes[2]; in igt_debug() local 277 memset(nodes, 0, sizeof(nodes)); in igt_debug() 278 nodes[0].start = 512; in igt_debug() 279 nodes[0].size = 1024; in igt_debug() 280 ret = drm_mm_reserve_node(&mm, &nodes[0]); in igt_debug() 283 nodes[0].start, nodes[0].size); in igt_debug() 287 nodes[1].size = 1024; in igt_debug() 288 nodes[1].start = 4096 - 512 - nodes[1].size; in igt_debug() 289 ret = drm_mm_reserve_node(&mm, &nodes[1]); in igt_debug() 292 nodes[1].start, nodes[1].size); in igt_debug() [all …]
|
/drivers/md/persistent-data/ |
D | dm-btree-spine.c | 131 s->nodes[0] = NULL; in init_ro_spine() 132 s->nodes[1] = NULL; in init_ro_spine() 140 unlock_block(s->info, s->nodes[i]); in exit_ro_spine() 149 unlock_block(s->info, s->nodes[0]); in ro_step() 150 s->nodes[0] = s->nodes[1]; in ro_step() 154 r = bn_read_lock(s->info, new_child, s->nodes + s->count); in ro_step() 165 unlock_block(s->info, s->nodes[s->count]); in ro_pop() 173 block = s->nodes[s->count - 1]; in ro_node() 191 unlock_block(s->info, s->nodes[i]); in exit_shadow_spine() 203 unlock_block(s->info, s->nodes[0]); in shadow_step() [all …]
|
/drivers/interconnect/qcom/ |
D | sc7180.c | 161 .nodes = { &ebi }, 168 .nodes = { &ebi }, 175 .nodes = { &qns_llcc }, 182 .nodes = { &qns_mem_noc_hf }, 189 .nodes = { &qxm_crypto }, 196 .nodes = { &qnm_snoc, 251 .nodes = { &qxm_camnoc_hf0_uncomp, 266 .nodes = { &acm_sys_tcu }, 273 .nodes = { &qns_mem_noc_sf }, 280 .nodes = { &qup_core_master_1, &qup_core_master_2 }, [all …]
|
D | osm-l3.c | 68 const struct qcom_icc_node **nodes; member 93 .nodes = sdm845_osm_l3_nodes, 109 .nodes = sc7180_osm_l3_nodes, 125 .nodes = sm8150_osm_l3_nodes, 141 .nodes = sm8250_epss_l3_nodes, 163 list_for_each_entry(n, &provider->nodes, node_list) in qcom_icc_set() 258 qnodes = desc->nodes; in qcom_osm_l3_probe() 261 data = devm_kzalloc(&pdev->dev, struct_size(data, nodes, num_nodes), GFP_KERNEL); in qcom_osm_l3_probe() 270 INIT_LIST_HEAD(&provider->nodes); in qcom_osm_l3_probe() 296 data->nodes[i] = node; in qcom_osm_l3_probe()
|
D | sm8150.c | 206 .nodes = aggre1_noc_nodes, 241 .nodes = aggre2_noc_nodes, 259 .nodes = camnoc_virt_nodes, 276 .nodes = compute_noc_nodes, 344 .nodes = config_noc_nodes, 360 .nodes = dc_noc_nodes, 395 .nodes = gem_noc_nodes, 411 .nodes = ipa_virt_nodes, 428 .nodes = mc_virt_nodes, 458 .nodes = mmss_noc_nodes, [all …]
|
D | sm8250.c | 220 .nodes = aggre1_noc_nodes, 250 .nodes = aggre2_noc_nodes, 267 .nodes = compute_noc_nodes, 333 .nodes = config_noc_nodes, 349 .nodes = dc_noc_nodes, 383 .nodes = gem_noc_nodes, 399 .nodes = ipa_virt_nodes, 416 .nodes = mc_virt_nodes, 446 .nodes = mmss_noc_nodes, 472 .nodes = npu_noc_nodes, [all …]
|
D | msm8974.c | 223 struct msm8974_icc_node **nodes; member 258 .nodes = msm8974_bimc_nodes, 341 .nodes = msm8974_cnoc_nodes, 394 .nodes = msm8974_mnoc_nodes, 429 .nodes = msm8974_onoc_nodes, 492 .nodes = msm8974_pnoc_nodes, 549 .nodes = msm8974_snoc_nodes, 596 list_for_each_entry(n, &provider->nodes, node_list) in msm8974_icc_set() 668 qnodes = desc->nodes; in msm8974_icc_probe() 675 data = devm_kzalloc(dev, struct_size(data, nodes, num_nodes), in msm8974_icc_probe() [all …]
|
D | qcs404.c | 143 struct qcom_icc_node **nodes; member 244 .nodes = qcs404_bimc_nodes, 298 .nodes = qcs404_pcnoc_nodes, 326 .nodes = qcs404_snoc_nodes, 347 list_for_each_entry(n, &provider->nodes, node_list) in qcom_icc_set() 420 qnodes = desc->nodes; in qnoc_probe() 427 data = devm_kzalloc(dev, struct_size(data, nodes, num_nodes), in qnoc_probe() 447 INIT_LIST_HEAD(&provider->nodes); in qnoc_probe() 480 data->nodes[i] = node; in qnoc_probe()
|
D | msm8916.c | 155 struct msm8916_icc_node **nodes; member 287 .nodes = msm8916_snoc_nodes, 304 .nodes = msm8916_bimc_nodes, 362 .nodes = msm8916_pcnoc_nodes, 380 list_for_each_entry(n, &provider->nodes, node_list) in msm8916_icc_set() 453 qnodes = desc->nodes; in msm8916_qnoc_probe() 460 data = devm_kzalloc(dev, struct_size(data, nodes, num_nodes), in msm8916_qnoc_probe() 480 INIT_LIST_HEAD(&provider->nodes); in msm8916_qnoc_probe() 510 data->nodes[i] = node; in msm8916_qnoc_probe()
|
D | sdm845.c | 196 .nodes = aggre1_noc_nodes, 224 .nodes = aggre2_noc_nodes, 285 .nodes = config_noc_nodes, 301 .nodes = dc_noc_nodes, 319 .nodes = gladiator_noc_nodes, 354 .nodes = mem_noc_nodes, 388 .nodes = mmss_noc_nodes, 434 .nodes = system_noc_nodes, 455 qnodes = desc->nodes; in qnoc_probe() 462 data = devm_kzalloc(&pdev->dev, struct_size(data, nodes, num_nodes), in qnoc_probe() [all …]
|
/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_vram_mgr.c | 261 struct drm_mm_node *nodes = mem->mm_node; in amdgpu_vram_mgr_bo_visible_size() local 271 for (usage = 0; nodes && pages; pages -= nodes->size, nodes++) in amdgpu_vram_mgr_bo_visible_size() 272 usage += amdgpu_vram_mgr_vis_size(adev, nodes); in amdgpu_vram_mgr_bo_visible_size() 317 struct drm_mm_node *nodes; in amdgpu_vram_mgr_new() local 353 nodes = kvmalloc_array((uint32_t)num_nodes, sizeof(*nodes), in amdgpu_vram_mgr_new() 355 if (!nodes) { in amdgpu_vram_mgr_new() 371 r = drm_mm_insert_node_in_range(mm, &nodes[i], pages, in amdgpu_vram_mgr_new() 378 vis_usage += amdgpu_vram_mgr_vis_size(adev, &nodes[i]); in amdgpu_vram_mgr_new() 379 amdgpu_vram_mgr_virt_start(mem, &nodes[i]); in amdgpu_vram_mgr_new() 390 r = drm_mm_insert_node_in_range(mm, &nodes[i], in amdgpu_vram_mgr_new() [all …]
|
D | ta_xgmi_if.h | 92 struct ta_xgmi_node_info nodes[TA_XGMI__MAX_CONNECTED_NODES]; member 97 struct ta_xgmi_node_info nodes[TA_XGMI__MAX_CONNECTED_NODES]; member 102 struct ta_xgmi_node_info nodes[TA_XGMI__MAX_CONNECTED_NODES]; member
|
/drivers/clk/zynqmp/ |
D | clk-zynqmp.h | 39 const struct clock_topology *nodes); 44 const struct clock_topology *nodes); 50 const struct clock_topology *nodes); 55 const struct clock_topology *nodes); 61 const struct clock_topology *nodes);
|
D | divider.c | 247 const struct clock_topology *nodes) in zynqmp_clk_register_divider() argument 262 init.flags = nodes->flag & ~CLK_FRAC; in zynqmp_clk_register_divider() 267 div->is_frac = !!((nodes->flag & CLK_FRAC) | in zynqmp_clk_register_divider() 268 (nodes->custom_type_flag & CUSTOM_FLAG_CLK_FRAC)); in zynqmp_clk_register_divider() 269 div->flags = nodes->type_flag; in zynqmp_clk_register_divider() 272 div->div_type = nodes->type; in zynqmp_clk_register_divider() 278 div->max_div = zynqmp_clk_get_max_divisor(clk_id, nodes->type); in zynqmp_clk_register_divider()
|
D | clk-mux-zynqmp.c | 107 const struct clock_topology *nodes) in zynqmp_clk_register_mux() argument 119 if (nodes->type_flag & CLK_MUX_READ_ONLY) in zynqmp_clk_register_mux() 123 init.flags = nodes->flag; in zynqmp_clk_register_mux() 126 mux->flags = nodes->type_flag; in zynqmp_clk_register_mux()
|
D | clkc.c | 124 const struct clock_topology *nodes) 288 const struct clock_topology *nodes) in zynqmp_clk_register_fixed_factor() argument 308 nodes->flag, mult, in zynqmp_clk_register_fixed_factor() 565 struct clock_topology *nodes; in zynqmp_register_clk_topology() local 568 nodes = clock[clk_id].node; in zynqmp_register_clk_topology() 579 clk_type_postfix[nodes[j].type]); in zynqmp_register_clk_topology() 584 if (!clk_topology[nodes[j].type]) in zynqmp_register_clk_topology() 587 hw = (*clk_topology[nodes[j].type])(clk_out[j], clk_dev_id, in zynqmp_register_clk_topology() 590 &nodes[j]); in zynqmp_register_clk_topology()
|
/drivers/dma/ |
D | uniphier-xdmac.c | 83 struct uniphier_xdmac_desc_node nodes[]; member 140 src_addr = xd->nodes[xd->cur_node].src; in uniphier_xdmac_chan_start() 141 dst_addr = xd->nodes[xd->cur_node].dst; in uniphier_xdmac_chan_start() 142 its = xd->nodes[xd->cur_node].burst_size; in uniphier_xdmac_chan_start() 143 tnum = xd->nodes[xd->cur_node].nr_burst; in uniphier_xdmac_chan_start() 295 xd = kzalloc(struct_size(xd, nodes, nr), GFP_NOWAIT); in uniphier_xdmac_prep_dma_memcpy() 301 xd->nodes[i].src = src; in uniphier_xdmac_prep_dma_memcpy() 302 xd->nodes[i].dst = dst; in uniphier_xdmac_prep_dma_memcpy() 303 xd->nodes[i].burst_size = burst_size; in uniphier_xdmac_prep_dma_memcpy() 304 xd->nodes[i].nr_burst = len / burst_size; in uniphier_xdmac_prep_dma_memcpy() [all …]
|
/drivers/interconnect/imx/ |
D | imx.c | 169 list_for_each_entry_safe(node, tmp, &provider->nodes, node_list) in imx_icc_unregister_nodes() 192 provider_data->nodes[node->id] = node; in imx_icc_register_nodes() 212 static int get_max_node_id(struct imx_icc_node_desc *nodes, int nodes_count) in get_max_node_id() argument 217 if (nodes[i].id > ret) in get_max_node_id() 218 ret = nodes[i].id; in get_max_node_id() 224 struct imx_icc_node_desc *nodes, int nodes_count) in imx_icc_register() argument 233 num_nodes = get_max_node_id(nodes, nodes_count) + 1; in imx_icc_register() 234 data = devm_kzalloc(dev, struct_size(data, nodes, num_nodes), in imx_icc_register() 256 ret = imx_icc_register_nodes(provider, nodes, nodes_count); in imx_icc_register()
|
D | imx8mn.c | 32 static struct imx_icc_node_desc nodes[] = { variable 75 return imx_icc_register(pdev, nodes, ARRAY_SIZE(nodes)); in imx8mn_icc_probe()
|
D | imx8mm.c | 35 static struct imx_icc_node_desc nodes[] = { variable 86 return imx_icc_register(pdev, nodes, ARRAY_SIZE(nodes)); in imx8mm_icc_probe()
|
D | imx8mq.c | 33 static struct imx_icc_node_desc nodes[] = { variable 85 return imx_icc_register(pdev, nodes, ARRAY_SIZE(nodes)); in imx8mq_icc_probe()
|
/drivers/gpu/drm/nouveau/nvkm/core/ |
D | mm.c | 26 #define node(root, dir) ((root)->nl_entry.dir == &mm->nodes) ? NULL : \ 36 list_for_each_entry(node, &mm->nodes, nl_entry) { in nvkm_mm_dump() 246 prev = list_last_entry(&mm->nodes, typeof(*node), nl_entry); in nvkm_mm_init() 255 list_add_tail(&node->nl_entry, &mm->nodes); in nvkm_mm_init() 259 INIT_LIST_HEAD(&mm->nodes); in nvkm_mm_init() 275 list_add_tail(&node->nl_entry, &mm->nodes); in nvkm_mm_init() 286 int nodes = 0; in nvkm_mm_fini() local 291 list_for_each_entry(node, &mm->nodes, nl_entry) { in nvkm_mm_fini() 293 if (++nodes > mm->heap_nodes) { in nvkm_mm_fini() 300 list_for_each_entry_safe(node, temp, &mm->nodes, nl_entry) { in nvkm_mm_fini()
|
/drivers/staging/media/ipu3/ |
D | ipu3.c | 120 if (!imgu_pipe->nodes[IMGU_NODE_VF].enabled && in imgu_dummybufs_init() 128 meta = &imgu_pipe->nodes[node].vdev_fmt.fmt.meta; in imgu_dummybufs_init() 129 mpix = &imgu_pipe->nodes[node].vdev_fmt.fmt.pix_mp; in imgu_dummybufs_init() 214 list_for_each_entry(buf, &imgu_pipe->nodes[node].buffers, vid_buf.list) { in imgu_queue_getbuf() 249 !imgu_pipe->nodes[IMGU_NODE_VF].enabled) { in imgu_queue_buffers() 254 imgu_pipe->nodes[node].enabled) { in imgu_queue_buffers() 259 if (list_empty(&imgu_pipe->nodes[node].buffers)) in imgu_queue_buffers() 262 ivb = list_first_entry(&imgu_pipe->nodes[node].buffers, in imgu_queue_buffers() 330 &imgu_pipe->nodes[node].buffers, in imgu_queue_buffers() 354 fmt = &imgu->imgu_pipe[pipe].nodes[IMGU_NODE_IN].pad_fmt; in imgu_powerup() [all …]
|
D | ipu3-v4l2.c | 32 imgu_pipe->nodes[IMGU_NODE_IN].vdev_fmt.fmt.pix_mp.width; in imgu_subdev_open() 34 imgu_pipe->nodes[IMGU_NODE_IN].vdev_fmt.fmt.pix_mp.height; in imgu_subdev_open() 43 try_fmt->code = imgu_pipe->nodes[i].pad_fmt.code; in imgu_subdev_open() 79 imgu_pipe->queue_enabled[i] = imgu_pipe->nodes[i].enabled; in imgu_subdev_s_stream() 91 &imgu_pipe->nodes[node].vdev_fmt.fmt.pix_mp : NULL; in imgu_subdev_s_stream() 96 if (imgu_pipe->nodes[IMGU_NODE_VF].enabled) in imgu_subdev_s_stream() 137 fmt->format = imgu_pipe->nodes[pad].pad_fmt; in imgu_subdev_get_fmt() 166 mf = &imgu_pipe->nodes[pad].pad_fmt; in imgu_subdev_set_fmt() 285 imgu_pipe->nodes[pad].enabled = flags & MEDIA_LNK_FL_ENABLED; in imgu_link_setup() 431 node = &imgu->imgu_pipe[p].nodes[i]; in imgu_all_nodes_streaming() [all …]
|
/drivers/base/test/ |
D | property-entry-test.c | 366 static const struct software_node nodes[] = { in pe_test_reference() local 374 .node = &nodes[0], in pe_test_reference() 378 .node = &nodes[1], in pe_test_reference() 385 PROPERTY_ENTRY_REF("ref-1", &nodes[0]), in pe_test_reference() 386 PROPERTY_ENTRY_REF("ref-2", &nodes[1], 1, 2), in pe_test_reference() 395 error = software_node_register_nodes(nodes); in pe_test_reference() 404 KUNIT_EXPECT_PTR_EQ(test, to_software_node(ref.fwnode), &nodes[0]); in pe_test_reference() 415 KUNIT_EXPECT_PTR_EQ(test, to_software_node(ref.fwnode), &nodes[1]); in pe_test_reference() 423 KUNIT_EXPECT_PTR_EQ(test, to_software_node(ref.fwnode), &nodes[1]); in pe_test_reference() 438 KUNIT_EXPECT_PTR_EQ(test, to_software_node(ref.fwnode), &nodes[0]); in pe_test_reference() [all …]
|