/drivers/gpu/drm/i915/display/ |
D | intel_ddi_buf_trans.c | 31 .num_entries = ARRAY_SIZE(_hsw_trans_dp), 48 .num_entries = ARRAY_SIZE(_hsw_trans_fdi), 69 .num_entries = ARRAY_SIZE(_hsw_trans_hdmi), 87 .num_entries = ARRAY_SIZE(_bdw_trans_edp), 104 .num_entries = ARRAY_SIZE(_bdw_trans_dp), 121 .num_entries = ARRAY_SIZE(_bdw_trans_fdi), 140 .num_entries = ARRAY_SIZE(_bdw_trans_hdmi), 159 .num_entries = ARRAY_SIZE(_skl_trans_dp), 177 .num_entries = ARRAY_SIZE(_skl_u_trans_dp), 195 .num_entries = ARRAY_SIZE(_skl_y_trans_dp), [all …]
|
/drivers/char/agp/ |
D | generic.c | 321 int num_entries; in agp_num_entries() local 328 num_entries = A_SIZE_8(temp)->num_entries; in agp_num_entries() 331 num_entries = A_SIZE_16(temp)->num_entries; in agp_num_entries() 334 num_entries = A_SIZE_32(temp)->num_entries; in agp_num_entries() 337 num_entries = A_SIZE_LVL2(temp)->num_entries; in agp_num_entries() 340 num_entries = A_SIZE_FIX(temp)->num_entries; in agp_num_entries() 343 num_entries = 0; in agp_num_entries() 347 num_entries -= agp_memory_reserved>>PAGE_SHIFT; in agp_num_entries() 348 if (num_entries<0) in agp_num_entries() 349 num_entries = 0; in agp_num_entries() [all …]
|
D | i460-agp.c | 153 values[i].num_entries = (values[i].size << 8) >> (I460_IO_PAGE_SHIFT - 12); in i460_fetch_size() 154 values[i].page_order = ilog2((sizeof(u32)*values[i].num_entries) >> PAGE_SHIFT); in i460_fetch_size() 234 size = current_size->num_entries * sizeof(i460.lp_desc[0]); in i460_configure() 244 int page_order, num_entries, i; in i460_create_gatt_table() local 252 num_entries = A_SIZE_8(temp)->num_entries; in i460_create_gatt_table() 265 for (i = 0; i < num_entries; ++i) in i460_create_gatt_table() 273 int num_entries, i; in i460_free_gatt_table() local 278 num_entries = A_SIZE_8(temp)->num_entries; in i460_free_gatt_table() 280 for (i = 0; i < num_entries; ++i) in i460_free_gatt_table() 282 WR_FLUSH_GATT(num_entries - 1); in i460_free_gatt_table() [all …]
|
D | efficeon-agp.c | 198 int num_entries, l1_pages; in efficeon_create_gatt_table() local 200 num_entries = A_SIZE_LVL2(agp_bridge->current_size)->num_entries; in efficeon_create_gatt_table() 202 printk(KERN_DEBUG PFX "efficeon_create_gatt_table(%d)\n", num_entries); in efficeon_create_gatt_table() 205 BUG_ON(num_entries & 0x3ff); in efficeon_create_gatt_table() 206 l1_pages = num_entries >> 10; in efficeon_create_gatt_table() 238 int i, count = mem->page_count, num_entries; in efficeon_insert_memory() local 245 num_entries = A_SIZE_LVL2(agp_bridge->current_size)->num_entries; in efficeon_insert_memory() 246 if ((pg_start + mem->page_count) > num_entries) in efficeon_insert_memory() 287 int i, count = mem->page_count, num_entries; in efficeon_remove_memory() local 291 num_entries = A_SIZE_LVL2(agp_bridge->current_size)->num_entries; in efficeon_remove_memory() [all …]
|
D | alpha-agp.c | 88 int num_entries, status; in alpha_core_agp_insert_memory() local 95 num_entries = A_SIZE_FIX(temp)->num_entries; in alpha_core_agp_insert_memory() 96 if ((pg_start + mem->page_count) > num_entries) in alpha_core_agp_insert_memory() 169 aper_size->num_entries = agp->aperture.size / PAGE_SIZE; in alpha_core_agp_setup() 170 aper_size->page_order = __ffs(aper_size->num_entries / 1024); in alpha_core_agp_setup()
|
/drivers/gpu/drm/amd/display/dc/dml/dcn321/ |
D | dcn321_fpu.c | 211 unsigned int *num_entries, in dcn321_insert_entry_into_table_sorted() argument 222 if (*num_entries == 0) { in dcn321_insert_entry_into_table_sorted() 224 (*num_entries)++; in dcn321_insert_entry_into_table_sorted() 229 if (index >= *num_entries) in dcn321_insert_entry_into_table_sorted() 233 for (i = *num_entries; i > index; i--) in dcn321_insert_entry_into_table_sorted() 237 (*num_entries)++; in dcn321_insert_entry_into_table_sorted() 241 …ove_entry_from_table_at_index(struct _vcs_dpi_voltage_scaling_st *table, unsigned int *num_entries, in remove_entry_from_table_at_index() argument 246 if (*num_entries == 0) in remove_entry_from_table_at_index() 249 for (i = index; i < *num_entries - 1; i++) { in remove_entry_from_table_at_index() 252 memset(&table[--(*num_entries)], 0, sizeof(struct _vcs_dpi_voltage_scaling_st)); in remove_entry_from_table_at_index() [all …]
|
/drivers/net/wwan/iosm/ |
D | iosm_ipc_coredump.c | 66 u32 byte_read, num_entries, file_size; in ipc_coredump_get_list() local 95 num_entries = le32_to_cpu(cd_table->list.num_entries); in ipc_coredump_get_list() 96 if (num_entries == 0 || num_entries > IOSM_NOF_CD_REGION) { in ipc_coredump_get_list() 101 for (i = 0; i < num_entries; i++) { in ipc_coredump_get_list()
|
/drivers/net/ethernet/netronome/nfp/ |
D | nfp_shared_buf.c | 81 unsigned int i, num_entries, entry_sz; in nfp_shared_buf_register() local 92 num_entries = n; in nfp_shared_buf_register() 95 num_entries * sizeof(pf->shared_bufs[0]), in nfp_shared_buf_register() 100 entry_sz = nfp_cpp_area_size(sb_desc_area) / num_entries; in nfp_shared_buf_register() 102 pf->shared_bufs = kmalloc_array(num_entries, sizeof(pf->shared_bufs[0]), in nfp_shared_buf_register() 109 for (i = 0; i < num_entries; i++) { in nfp_shared_buf_register() 125 pf->num_shared_bufs = num_entries; in nfp_shared_buf_register()
|
/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_bo_list.c | 64 size_t num_entries, struct amdgpu_bo_list **result) in amdgpu_bo_list_create() argument 66 unsigned last_entry = 0, first_userptr = num_entries; in amdgpu_bo_list_create() 74 if (num_entries > (SIZE_MAX - sizeof(struct amdgpu_bo_list)) in amdgpu_bo_list_create() 79 size += num_entries * sizeof(struct amdgpu_bo_list_entry); in amdgpu_bo_list_create() 90 memset(array, 0, num_entries * sizeof(struct amdgpu_bo_list_entry)); in amdgpu_bo_list_create() 92 for (i = 0; i < num_entries; ++i) { in amdgpu_bo_list_create() 135 list->num_entries = num_entries; in amdgpu_bo_list_create() 137 trace_amdgpu_cs_bo_status(list->num_entries, total_size); in amdgpu_bo_list_create() 149 for (i = first_userptr; i < num_entries; ++i) { in amdgpu_bo_list_create()
|
D | amdgpu_bo_list.h | 52 unsigned num_entries; member 70 size_t num_entries, 83 e != amdgpu_bo_list_array_entry(list, (list)->num_entries); \ 88 e != amdgpu_bo_list_array_entry(list, (list)->num_entries); \
|
/drivers/net/dsa/sja1105/ |
D | sja1105_tas.c | 172 int num_entries = 0; in sja1105_init_scheduling() local 213 num_entries += tas_data->offload[port]->num_entries; in sja1105_init_scheduling() 219 num_entries += gating_cfg->num_entries; in sja1105_init_scheduling() 231 table->entries = kcalloc(num_entries, table->ops->unpacked_entry_size, in sja1105_init_scheduling() 235 table->entry_count = num_entries; in sja1105_init_scheduling() 283 schedule_end_idx = k + offload->num_entries - 1; in sja1105_init_scheduling() 310 for (i = 0; i < offload->num_entries; i++, k++) { in sja1105_init_scheduling() 329 schedule_end_idx = k + gating_cfg->num_entries - 1; in sja1105_init_scheduling() 425 i < offload->num_entries; in sja1105_tas_check_conflicts() 431 j < admin->num_entries; in sja1105_tas_check_conflicts() [all …]
|
/drivers/net/ethernet/engleder/ |
D | tsnep_selftests.c | 381 qopt->num_entries = 7; in tsnep_test_taprio() 405 qopt->num_entries = 8; in tsnep_test_taprio() 434 qopt->num_entries = 10; in tsnep_test_taprio() 468 qopt->num_entries = 2; in tsnep_test_taprio_change() 501 qopt->num_entries = 3; in tsnep_test_taprio_change() 513 qopt->num_entries = 2; in tsnep_test_taprio_change() 527 qopt->num_entries = 4; in tsnep_test_taprio_change() 539 qopt->num_entries = 2; in tsnep_test_taprio_change() 551 qopt->num_entries = 3; in tsnep_test_taprio_change() 567 qopt->num_entries = 4; in tsnep_test_taprio_change() [all …]
|
/drivers/iommu/ |
D | io-pgtable-arm-v7s.c | 315 static void __arm_v7s_pte_sync(arm_v7s_iopte *ptep, int num_entries, in __arm_v7s_pte_sync() argument 322 num_entries * sizeof(*ptep), DMA_TO_DEVICE); in __arm_v7s_pte_sync() 325 int num_entries, struct io_pgtable_cfg *cfg) in __arm_v7s_set_pte() argument 329 for (i = 0; i < num_entries; i++) in __arm_v7s_set_pte() 332 __arm_v7s_pte_sync(ptep, num_entries, cfg); in __arm_v7s_set_pte() 433 int lvl, int num_entries, arm_v7s_iopte *ptep) in arm_v7s_init_pte() argument 439 for (i = 0; i < num_entries; i++) in arm_v7s_init_pte() 459 if (num_entries > 1) in arm_v7s_init_pte() 464 __arm_v7s_set_pte(ptep, pte, num_entries, cfg); in arm_v7s_init_pte() 503 int num_entries = size >> ARM_V7S_LVL_SHIFT(lvl); in __arm_v7s_map() local [all …]
|
D | io-pgtable-dart.c | 125 dart_iopte prot, int num_entries, in dart_init_pte() argument 132 for (i = 0; i < num_entries; i++) in dart_init_pte() 146 for (i = 0; i < num_entries; i++) in dart_init_pte() 243 int ret = 0, tbl, num_entries, max_entries, map_idx_start; in dart_map_pages() local 283 num_entries = min_t(int, pgcount, max_entries); in dart_map_pages() 285 ret = dart_init_pte(data, iova, paddr, prot, num_entries, ptep); in dart_map_pages() 287 *mapped += num_entries * pgsize; in dart_map_pages() 304 int i = 0, num_entries, max_entries, unmap_idx_start; in dart_unmap_pages() local 320 num_entries = min_t(int, pgcount, max_entries); in dart_unmap_pages() 322 while (i < num_entries) { in dart_unmap_pages()
|
D | io-pgtable-arm.c | 241 static void __arm_lpae_sync_pte(arm_lpae_iopte *ptep, int num_entries, in __arm_lpae_sync_pte() argument 245 sizeof(*ptep) * num_entries, DMA_TO_DEVICE); in __arm_lpae_sync_pte() 264 int lvl, int num_entries, arm_lpae_iopte *ptep) in __arm_lpae_init_pte() argument 276 for (i = 0; i < num_entries; i++) in __arm_lpae_init_pte() 280 __arm_lpae_sync_pte(ptep, num_entries, cfg); in __arm_lpae_init_pte() 285 arm_lpae_iopte prot, int lvl, int num_entries, in arm_lpae_init_pte() argument 290 for (i = 0; i < num_entries; i++) in arm_lpae_init_pte() 311 __arm_lpae_init_pte(data, paddr, prot, lvl, num_entries, ptep); in arm_lpae_init_pte() 356 int ret = 0, num_entries, max_entries, map_idx_start; in __arm_lpae_map() local 365 num_entries = min_t(int, pgcount, max_entries); in __arm_lpae_map() [all …]
|
/drivers/soc/qcom/ |
D | smsm.c | 81 u32 num_entries; member 449 u32 num_entries; in smsm_get_size_info() member 461 smsm->num_entries = SMSM_DEFAULT_NUM_ENTRIES; in smsm_get_size_info() 466 smsm->num_entries = info->num_entries; in smsm_get_size_info() 471 smsm->num_entries, smsm->num_hosts); in smsm_get_size_info() 499 smsm->num_entries, in qcom_smsm_probe() 534 smsm->num_entries * sizeof(u32)); in qcom_smsm_probe() 548 size = smsm->num_entries * smsm->num_hosts * sizeof(u32); in qcom_smsm_probe() 580 if (ret || id >= smsm->num_entries) { in qcom_smsm_probe() 607 for (id = 0; id < smsm->num_entries; id++) in qcom_smsm_probe() [all …]
|
/drivers/parisc/ |
D | iosapic.c | 222 static struct irt_entry *iosapic_alloc_irt(int num_entries) in iosapic_alloc_irt() argument 224 return kcalloc(num_entries, sizeof(struct irt_entry), GFP_KERNEL); in iosapic_alloc_irt() 259 unsigned long num_entries = 0UL; in iosapic_load_irt() local 266 status = pdc_pat_get_irt_size(&num_entries, cell_num); in iosapic_load_irt() 270 BUG_ON(num_entries == 0); in iosapic_load_irt() 278 table = iosapic_alloc_irt(num_entries); in iosapic_load_irt() 299 status = pdc_pci_irt_size(&num_entries, 0); in iosapic_load_irt() 307 BUG_ON(num_entries == 0); in iosapic_load_irt() 309 table = iosapic_alloc_irt(num_entries); in iosapic_load_irt() 317 status = pdc_pci_irt(num_entries, 0, table); in iosapic_load_irt() [all …]
|
/drivers/net/ethernet/mellanox/mlxsw/ |
D | spectrum_nve.c | 65 unsigned int num_entries; member 311 WARN_ON(mc_record->num_entries); in mlxsw_sp_nve_mc_record_destroy() 323 unsigned int num_entries = mc_record->num_entries; in mlxsw_sp_nve_mc_record_get() local 327 num_entries < nve->num_max_mc_entries[proto]) in mlxsw_sp_nve_mc_record_get() 337 if (mc_record->num_entries != 0) in mlxsw_sp_nve_mc_record_put() 368 unsigned int num_entries = 0; in mlxsw_sp_nve_mc_record_refresh() local 383 next_kvdl_index, mc_record->num_entries); in mlxsw_sp_nve_mc_record_refresh() 393 num_entries++); in mlxsw_sp_nve_mc_record_refresh() 396 WARN_ON(num_entries != mc_record->num_entries); in mlxsw_sp_nve_mc_record_refresh() 449 mc_record->num_entries++; in mlxsw_sp_nve_mc_record_ip_add() [all …]
|
/drivers/gpu/drm/amd/display/dc/dml/dcn31/ |
D | dcn31_fpu.c | 599 ASSERT(clk_table->num_entries); in dcn31_update_bw_bounding_box() 602 for (i = 0; i < clk_table->num_entries; ++i) { in dcn31_update_bw_bounding_box() 609 for (i = 0; i < clk_table->num_entries; i++) { in dcn31_update_bw_bounding_box() 642 if (clk_table->num_entries) { in dcn31_update_bw_bounding_box() 643 dcn3_1_soc.num_states = clk_table->num_entries; in dcn31_update_bw_bounding_box() 679 ASSERT(clk_table->num_entries); in dcn315_update_bw_bounding_box() 682 for (i = 0; i < clk_table->num_entries; ++i) { in dcn315_update_bw_bounding_box() 689 for (i = 0; i < clk_table->num_entries; i++) { in dcn315_update_bw_bounding_box() 708 dcn3_15_soc.num_states = clk_table->num_entries; in dcn315_update_bw_bounding_box() 747 ASSERT(clk_table->num_entries); in dcn316_update_bw_bounding_box() [all …]
|
/drivers/gpu/drm/amd/pm/powerplay/smumgr/ |
D | smu7_smumgr.c | 376 toc->num_entries = 0; in smu7_request_smu_load_fw() 380 UCODE_ID_RLC_G, &toc->entry[toc->num_entries++]), in smu7_request_smu_load_fw() 383 UCODE_ID_CP_CE, &toc->entry[toc->num_entries++]), in smu7_request_smu_load_fw() 386 UCODE_ID_CP_PFP, &toc->entry[toc->num_entries++]), in smu7_request_smu_load_fw() 389 UCODE_ID_CP_ME, &toc->entry[toc->num_entries++]), in smu7_request_smu_load_fw() 392 UCODE_ID_CP_MEC, &toc->entry[toc->num_entries++]), in smu7_request_smu_load_fw() 395 UCODE_ID_CP_MEC_JT1, &toc->entry[toc->num_entries++]), in smu7_request_smu_load_fw() 398 UCODE_ID_CP_MEC_JT2, &toc->entry[toc->num_entries++]), in smu7_request_smu_load_fw() 401 UCODE_ID_SDMA0, &toc->entry[toc->num_entries++]), in smu7_request_smu_load_fw() 404 UCODE_ID_SDMA1, &toc->entry[toc->num_entries++]), in smu7_request_smu_load_fw() [all …]
|
/drivers/platform/x86/intel/ |
D | vsec.c | 56 u8 num_entries; member 212 if (!header->num_entries) { in intel_vsec_add_dev() 226 res = kcalloc(header->num_entries, sizeof(*res), GFP_KERNEL); in intel_vsec_add_dev() 240 for (i = 0, tmp = res; i < header->num_entries; i++, tmp++) { in intel_vsec_add_dev() 249 intel_vsec_dev->num_resources = header->num_entries; in intel_vsec_add_dev() 310 pci_read_config_byte(pdev, pos + INTEL_DVSEC_ENTRIES, &header.num_entries); in intel_vsec_walk_dvsec() 358 pci_read_config_byte(pdev, pos + INTEL_DVSEC_ENTRIES, &header.num_entries); in intel_vsec_walk_vsec() 416 .num_entries = 1,
|
/drivers/platform/x86/intel/pmt/ |
D | telemetry.c | 40 int num_entries; member 99 for (i = 0; i < priv->num_entries; i++) in pmt_telem_remove() 118 struct intel_pmt_entry *entry = &priv->entry[priv->num_entries]; in pmt_telem_probe() 126 priv->num_entries++; in pmt_telem_probe()
|
/drivers/scsi/aic7xxx/aicasm/ |
D | aicasm_symbol.c | 383 symbol_node_t *regnode, u_int num_entries) in aic_print_reg_dump_end() argument 396 if (num_entries != 0) in aic_print_reg_dump_end() 413 num_entries != 0 ? regnode->symbol->name : "NULL", in aic_print_reg_dump_end() 414 num_entries != 0 ? "_parse_table" : "", in aic_print_reg_dump_end() 415 num_entries, in aic_print_reg_dump_end() 548 int num_entries; in symtable_dump() local 550 num_entries = 0; in symtable_dump() 556 if (num_entries == 0) in symtable_dump() 561 num_entries++; in symtable_dump() 565 curnode, num_entries); in symtable_dump()
|
/drivers/regulator/ |
D | ti-abb-regulator.c | 506 int num_entries, min_uV = INT_MAX, max_uV = 0; in ti_abb_init_table() local 514 num_entries = of_property_count_u32_elems(dev->of_node, pname); in ti_abb_init_table() 515 if (num_entries < 0) { in ti_abb_init_table() 517 return num_entries; in ti_abb_init_table() 520 if (!num_entries || (num_entries % num_values)) { in ti_abb_init_table() 525 num_entries /= num_values; in ti_abb_init_table() 527 info = devm_kcalloc(dev, num_entries, sizeof(*info), GFP_KERNEL); in ti_abb_init_table() 533 volt_table = devm_kcalloc(dev, num_entries, sizeof(unsigned int), in ti_abb_init_table() 538 abb->rdesc.n_voltages = num_entries; in ti_abb_init_table() 543 for (i = 0; i < num_entries; i++, info++, volt_table++) { in ti_abb_init_table()
|
/drivers/infiniband/core/ |
D | uverbs_std_types_device.c | 273 size_t num_entries, size_t user_entry_size) in copy_gid_entries_to_user() argument 284 entries, sizeof(*entries) * num_entries); in copy_gid_entries_to_user() 294 for (i = 0; i < num_entries; i++) { in copy_gid_entries_to_user() 319 ssize_t num_entries; in UVERBS_HANDLER() local 352 num_entries = rdma_query_gid_table(ib_dev, entries, max_entries); in UVERBS_HANDLER() 353 if (num_entries < 0) in UVERBS_HANDLER() 356 ret = copy_gid_entries_to_user(attrs, entries, num_entries, in UVERBS_HANDLER() 363 &num_entries, sizeof(num_entries)); in UVERBS_HANDLER()
|