/drivers/char/agp/ |
D | generic.c | 321 int num_entries; in agp_num_entries() local 328 num_entries = A_SIZE_8(temp)->num_entries; in agp_num_entries() 331 num_entries = A_SIZE_16(temp)->num_entries; in agp_num_entries() 334 num_entries = A_SIZE_32(temp)->num_entries; in agp_num_entries() 337 num_entries = A_SIZE_LVL2(temp)->num_entries; in agp_num_entries() 340 num_entries = A_SIZE_FIX(temp)->num_entries; in agp_num_entries() 343 num_entries = 0; in agp_num_entries() 347 num_entries -= agp_memory_reserved>>PAGE_SHIFT; in agp_num_entries() 348 if (num_entries<0) in agp_num_entries() 349 num_entries = 0; in agp_num_entries() [all …]
|
D | i460-agp.c | 153 values[i].num_entries = (values[i].size << 8) >> (I460_IO_PAGE_SHIFT - 12); in i460_fetch_size() 154 values[i].page_order = ilog2((sizeof(u32)*values[i].num_entries) >> PAGE_SHIFT); in i460_fetch_size() 234 size = current_size->num_entries * sizeof(i460.lp_desc[0]); in i460_configure() 244 int page_order, num_entries, i; in i460_create_gatt_table() local 252 num_entries = A_SIZE_8(temp)->num_entries; in i460_create_gatt_table() 265 for (i = 0; i < num_entries; ++i) in i460_create_gatt_table() 273 int num_entries, i; in i460_free_gatt_table() local 278 num_entries = A_SIZE_8(temp)->num_entries; in i460_free_gatt_table() 280 for (i = 0; i < num_entries; ++i) in i460_free_gatt_table() 282 WR_FLUSH_GATT(num_entries - 1); in i460_free_gatt_table() [all …]
|
D | efficeon-agp.c | 198 int num_entries, l1_pages; in efficeon_create_gatt_table() local 200 num_entries = A_SIZE_LVL2(agp_bridge->current_size)->num_entries; in efficeon_create_gatt_table() 202 printk(KERN_DEBUG PFX "efficeon_create_gatt_table(%d)\n", num_entries); in efficeon_create_gatt_table() 205 BUG_ON(num_entries & 0x3ff); in efficeon_create_gatt_table() 206 l1_pages = num_entries >> 10; in efficeon_create_gatt_table() 238 int i, count = mem->page_count, num_entries; in efficeon_insert_memory() local 245 num_entries = A_SIZE_LVL2(agp_bridge->current_size)->num_entries; in efficeon_insert_memory() 246 if ((pg_start + mem->page_count) > num_entries) in efficeon_insert_memory() 287 int i, count = mem->page_count, num_entries; in efficeon_remove_memory() local 291 num_entries = A_SIZE_LVL2(agp_bridge->current_size)->num_entries; in efficeon_remove_memory() [all …]
|
D | alpha-agp.c | 88 int num_entries, status; in alpha_core_agp_insert_memory() local 95 num_entries = A_SIZE_FIX(temp)->num_entries; in alpha_core_agp_insert_memory() 96 if ((pg_start + mem->page_count) > num_entries) in alpha_core_agp_insert_memory() 169 aper_size->num_entries = agp->aperture.size / PAGE_SIZE; in alpha_core_agp_setup() 170 aper_size->page_order = __ffs(aper_size->num_entries / 1024); in alpha_core_agp_setup()
|
D | agp.h | 66 int num_entries; member 73 int num_entries; member 80 int num_entries; member 87 int num_entries; member 93 int num_entries; member
|
D | uninorth-agp.c | 151 int i, num_entries; in uninorth_insert_memory() local 169 num_entries = A_SIZE_32(temp)->num_entries; in uninorth_insert_memory() 171 if ((pg_start + mem->page_count) > num_entries) in uninorth_insert_memory() 374 int num_entries; in uninorth_create_gatt_table() local 386 size = page_order = num_entries = 0; in uninorth_create_gatt_table() 391 num_entries = A_SIZE_32(temp)->num_entries; in uninorth_create_gatt_table() 436 for (i = 0; i < num_entries; i++) in uninorth_create_gatt_table()
|
D | amd-k7-agp.c | 138 retval = amd_create_gatt_pages(value->num_entries / 1024); in amd_create_gatt_table() 157 for (i = 0; i < value->num_entries / 1024; i++, addr += 0x00400000) { in amd_create_gatt_table() 163 for (i = 0; i < value->num_entries; i++) { in amd_create_gatt_table() 287 int i, j, num_entries; in amd_insert_memory() local 291 num_entries = A_SIZE_LVL2(agp_bridge->current_size)->num_entries; in amd_insert_memory() 297 if ((pg_start + mem->page_count) > num_entries) in amd_insert_memory()
|
D | ati-agp.c | 271 int i, j, num_entries; in ati_insert_memory() local 276 num_entries = A_SIZE_LVL2(agp_bridge->current_size)->num_entries; in ati_insert_memory() 285 if ((pg_start + mem->page_count) > num_entries) in ati_insert_memory() 358 retval = ati_create_gatt_pages(value->num_entries / 1024); in ati_create_gatt_table() 394 for (i = 0; i < value->num_entries / 1024; i++, addr += 0x00400000) { in ati_create_gatt_table() 400 for (i = 0; i < value->num_entries; i++) { in ati_create_gatt_table()
|
/drivers/net/ethernet/netronome/nfp/ |
D | nfp_shared_buf.c | 81 unsigned int i, num_entries, entry_sz; in nfp_shared_buf_register() local 92 num_entries = n; in nfp_shared_buf_register() 95 num_entries * sizeof(pf->shared_bufs[0]), in nfp_shared_buf_register() 100 entry_sz = nfp_cpp_area_size(sb_desc_area) / num_entries; in nfp_shared_buf_register() 102 pf->shared_bufs = kmalloc_array(num_entries, sizeof(pf->shared_bufs[0]), in nfp_shared_buf_register() 109 for (i = 0; i < num_entries; i++) { in nfp_shared_buf_register() 125 pf->num_shared_bufs = num_entries; in nfp_shared_buf_register()
|
/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_bo_list.c | 64 unsigned num_entries, struct amdgpu_bo_list **result) in amdgpu_bo_list_create() argument 66 unsigned last_entry = 0, first_userptr = num_entries; in amdgpu_bo_list_create() 74 if (num_entries > (SIZE_MAX - sizeof(struct amdgpu_bo_list)) in amdgpu_bo_list_create() 79 size += num_entries * sizeof(struct amdgpu_bo_list_entry); in amdgpu_bo_list_create() 90 memset(array, 0, num_entries * sizeof(struct amdgpu_bo_list_entry)); in amdgpu_bo_list_create() 92 for (i = 0; i < num_entries; ++i) { in amdgpu_bo_list_create() 135 list->num_entries = num_entries; in amdgpu_bo_list_create() 137 trace_amdgpu_cs_bo_status(list->num_entries, total_size); in amdgpu_bo_list_create() 148 for (i = first_userptr; i < num_entries; ++i) { in amdgpu_bo_list_create()
|
D | amdgpu_bo_list.h | 49 unsigned num_entries; member 63 unsigned num_entries, 76 e != amdgpu_bo_list_array_entry(list, (list)->num_entries); \ 81 e != amdgpu_bo_list_array_entry(list, (list)->num_entries); \
|
/drivers/net/dsa/sja1105/ |
D | sja1105_tas.c | 98 int num_entries = 0; in sja1105_init_scheduling() local 135 num_entries += tas_data->offload[port]->num_entries; in sja1105_init_scheduling() 148 table->entries = kcalloc(num_entries, table->ops->unpacked_entry_size, in sja1105_init_scheduling() 152 table->entry_count = num_entries; in sja1105_init_scheduling() 198 schedule_end_idx = k + offload->num_entries - 1; in sja1105_init_scheduling() 219 for (i = 0; i < offload->num_entries; i++, k++) { in sja1105_init_scheduling() 298 i < offload->num_entries; in sja1105_tas_check_conflicts() 304 j < admin->num_entries; in sja1105_tas_check_conflicts() 374 for (i = 0; i < admin->num_entries; i++) { in sja1105_setup_tc_taprio()
|
/drivers/iommu/ |
D | io-pgtable-arm-v7s.c | 296 static void __arm_v7s_pte_sync(arm_v7s_iopte *ptep, int num_entries, in __arm_v7s_pte_sync() argument 303 num_entries * sizeof(*ptep), DMA_TO_DEVICE); in __arm_v7s_pte_sync() 306 int num_entries, struct io_pgtable_cfg *cfg) in __arm_v7s_set_pte() argument 310 for (i = 0; i < num_entries; i++) in __arm_v7s_set_pte() 313 __arm_v7s_pte_sync(ptep, num_entries, cfg); in __arm_v7s_set_pte() 414 int lvl, int num_entries, arm_v7s_iopte *ptep) in arm_v7s_init_pte() argument 420 for (i = 0; i < num_entries; i++) in arm_v7s_init_pte() 440 if (num_entries > 1) in arm_v7s_init_pte() 445 __arm_v7s_set_pte(ptep, pte, num_entries, cfg); in arm_v7s_init_pte() 479 int num_entries = size >> ARM_V7S_LVL_SHIFT(lvl); in __arm_v7s_map() local [all …]
|
/drivers/soc/qcom/ |
D | smsm.c | 81 u32 num_entries; member 428 u32 num_entries; in smsm_get_size_info() member 440 smsm->num_entries = SMSM_DEFAULT_NUM_ENTRIES; in smsm_get_size_info() 445 smsm->num_entries = info->num_entries; in smsm_get_size_info() 450 smsm->num_entries, smsm->num_hosts); in smsm_get_size_info() 478 smsm->num_entries, in qcom_smsm_probe() 513 smsm->num_entries * sizeof(u32)); in qcom_smsm_probe() 526 size = smsm->num_entries * smsm->num_hosts * sizeof(u32); in qcom_smsm_probe() 556 if (ret || id >= smsm->num_entries) { in qcom_smsm_probe() 581 for (id = 0; id < smsm->num_entries; id++) in qcom_smsm_probe() [all …]
|
/drivers/net/ethernet/mellanox/mlxsw/ |
D | spectrum_nve.c | 65 unsigned int num_entries; member 301 WARN_ON(mc_record->num_entries); in mlxsw_sp_nve_mc_record_destroy() 313 unsigned int num_entries = mc_record->num_entries; in mlxsw_sp_nve_mc_record_get() local 317 num_entries < nve->num_max_mc_entries[proto]) in mlxsw_sp_nve_mc_record_get() 327 if (mc_record->num_entries != 0) in mlxsw_sp_nve_mc_record_put() 358 unsigned int num_entries = 0; in mlxsw_sp_nve_mc_record_refresh() local 373 next_kvdl_index, mc_record->num_entries); in mlxsw_sp_nve_mc_record_refresh() 383 num_entries++); in mlxsw_sp_nve_mc_record_refresh() 386 WARN_ON(num_entries != mc_record->num_entries); in mlxsw_sp_nve_mc_record_refresh() 439 mc_record->num_entries++; in mlxsw_sp_nve_mc_record_ip_add() [all …]
|
/drivers/parisc/ |
D | iosapic.c | 222 static struct irt_entry *iosapic_alloc_irt(int num_entries) in iosapic_alloc_irt() argument 231 a = (unsigned long)kmalloc(sizeof(struct irt_entry) * num_entries + 8, GFP_KERNEL); in iosapic_alloc_irt() 268 unsigned long num_entries = 0UL; in iosapic_load_irt() local 275 status = pdc_pat_get_irt_size(&num_entries, cell_num); in iosapic_load_irt() 279 BUG_ON(num_entries == 0); in iosapic_load_irt() 287 table = iosapic_alloc_irt(num_entries); in iosapic_load_irt() 308 status = pdc_pci_irt_size(&num_entries, 0); in iosapic_load_irt() 316 BUG_ON(num_entries == 0); in iosapic_load_irt() 318 table = iosapic_alloc_irt(num_entries); in iosapic_load_irt() 326 status = pdc_pci_irt(num_entries, 0, table); in iosapic_load_irt() [all …]
|
/drivers/staging/exfat/ |
D | exfat_core.c | 1287 s32 entry, s32 num_entries, in fat_init_ext_entry() argument 1307 if ((--num_entries) > 0) { in fat_init_ext_entry() 1311 for (i = 1; i < num_entries; i++) { in fat_init_ext_entry() 1338 s32 entry, s32 num_entries, in exfat_init_ext_entry() argument 1354 file_ep->num_ext = (u8)(num_entries - 1); in exfat_init_ext_entry() 1366 for (i = 2; i < num_entries; i++) { in exfat_init_ext_entry() 1487 s32 entry, s32 order, s32 num_entries) in fat_delete_dir_entry() argument 1494 for (i = num_entries - 1; i >= order; i--) { in fat_delete_dir_entry() 1505 s32 entry, s32 order, s32 num_entries) in exfat_delete_dir_entry() argument 1512 for (i = order; i < num_entries; i++) { in exfat_delete_dir_entry() [all …]
|
/drivers/gpu/drm/amd/powerplay/smumgr/ |
D | smu7_smumgr.c | 389 toc->num_entries = 0; in smu7_request_smu_load_fw() 393 UCODE_ID_RLC_G, &toc->entry[toc->num_entries++]), in smu7_request_smu_load_fw() 396 UCODE_ID_CP_CE, &toc->entry[toc->num_entries++]), in smu7_request_smu_load_fw() 399 UCODE_ID_CP_PFP, &toc->entry[toc->num_entries++]), in smu7_request_smu_load_fw() 402 UCODE_ID_CP_ME, &toc->entry[toc->num_entries++]), in smu7_request_smu_load_fw() 405 UCODE_ID_CP_MEC, &toc->entry[toc->num_entries++]), in smu7_request_smu_load_fw() 408 UCODE_ID_CP_MEC_JT1, &toc->entry[toc->num_entries++]), in smu7_request_smu_load_fw() 411 UCODE_ID_CP_MEC_JT2, &toc->entry[toc->num_entries++]), in smu7_request_smu_load_fw() 414 UCODE_ID_SDMA0, &toc->entry[toc->num_entries++]), in smu7_request_smu_load_fw() 417 UCODE_ID_SDMA1, &toc->entry[toc->num_entries++]), in smu7_request_smu_load_fw() [all …]
|
/drivers/scsi/aic7xxx/aicasm/ |
D | aicasm_symbol.c | 383 symbol_node_t *regnode, u_int num_entries) in aic_print_reg_dump_end() argument 396 if (num_entries != 0) in aic_print_reg_dump_end() 413 num_entries != 0 ? regnode->symbol->name : "NULL", in aic_print_reg_dump_end() 414 num_entries != 0 ? "_parse_table" : "", in aic_print_reg_dump_end() 415 num_entries, in aic_print_reg_dump_end() 548 int num_entries; in symtable_dump() local 550 num_entries = 0; in symtable_dump() 556 if (num_entries == 0) in symtable_dump() 561 num_entries++; in symtable_dump() 565 curnode, num_entries); in symtable_dump()
|
/drivers/regulator/ |
D | ti-abb-regulator.c | 504 int num_entries, min_uV = INT_MAX, max_uV = 0; in ti_abb_init_table() local 512 num_entries = of_property_count_u32_elems(dev->of_node, pname); in ti_abb_init_table() 513 if (num_entries < 0) { in ti_abb_init_table() 515 return num_entries; in ti_abb_init_table() 518 if (!num_entries || (num_entries % num_values)) { in ti_abb_init_table() 523 num_entries /= num_values; in ti_abb_init_table() 525 info = devm_kcalloc(dev, num_entries, sizeof(*info), GFP_KERNEL); in ti_abb_init_table() 531 volt_table = devm_kcalloc(dev, num_entries, sizeof(unsigned int), in ti_abb_init_table() 536 abb->rdesc.n_voltages = num_entries; in ti_abb_init_table() 541 for (i = 0; i < num_entries; i++, info++, volt_table++) { in ti_abb_init_table()
|
/drivers/gpu/drm/amd/display/modules/power/ |
D | power_helpers.c | 211 unsigned int num_entries = NUM_BL_CURVE_SEGS; in fill_backlight_transform_table() local 216 table->backlight_thresholds[num_entries-1] = 0xFFFF; in fill_backlight_transform_table() 217 table->backlight_offsets[num_entries-1] = in fill_backlight_transform_table() 228 for (i = 1; i+1 < num_entries; i++) { in fill_backlight_transform_table() 229 lut_index = (params.backlight_lut_array_size - 1) * i / (num_entries - 1); in fill_backlight_transform_table() 233 cpu_to_be16(DIV_ROUNDUP((i * 65536), num_entries)); in fill_backlight_transform_table() 243 unsigned int num_entries = NUM_BL_CURVE_SEGS; in fill_backlight_transform_table_v_2_2() local 248 table->backlight_thresholds[num_entries-1] = 0xFFFF; in fill_backlight_transform_table_v_2_2() 249 table->backlight_offsets[num_entries-1] = in fill_backlight_transform_table_v_2_2() 260 for (i = 1; i+1 < num_entries; i++) { in fill_backlight_transform_table_v_2_2() [all …]
|
/drivers/media/platform/vsp1/ |
D | vsp1_dl.c | 116 unsigned int num_entries; member 249 unsigned int num_entries, size_t extra_size) in vsp1_dl_body_pool_create() argument 267 dlb_size = num_entries * sizeof(struct vsp1_dl_entry) + extra_size; in vsp1_dl_body_pool_create() 291 dlb->max_entries = num_entries; in vsp1_dl_body_pool_create() 363 dlb->num_entries = 0; in vsp1_dl_body_put() 382 if (WARN_ONCE(dlb->num_entries >= dlb->max_entries, in vsp1_dl_body_write() 386 dlb->entries[dlb->num_entries].addr = reg; in vsp1_dl_body_write() 387 dlb->entries[dlb->num_entries].data = data; in vsp1_dl_body_write() 388 dlb->num_entries++; in vsp1_dl_body_write() 656 dl->body0->num_entries = 0; in __vsp1_dl_list_put() [all …]
|
/drivers/gpu/drm/i915/ |
D | intel_csr.c | 153 u32 num_entries; member 335 unsigned int num_entries, in find_dmc_fw_offset() argument 342 for (i = 0; i < num_entries; i++) { in find_dmc_fw_offset() 487 u32 num_entries, max_entries, dmc_offset; in parse_csr_fw_package() local 517 num_entries = package_header->num_entries; in parse_csr_fw_package() 518 if (WARN_ON(package_header->num_entries > max_entries)) in parse_csr_fw_package() 519 num_entries = max_entries; in parse_csr_fw_package() 523 dmc_offset = find_dmc_fw_offset(fw_info, num_entries, si, in parse_csr_fw_package()
|
/drivers/net/ethernet/brocade/bna/ |
D | bfa_msgq.c | 205 int num_entries = 0; in __cmd_copy() local 222 num_entries++; in __cmd_copy() 240 if (ntohs(cmd->msg_hdr->num_entries) <= in bfa_msgq_cmdq_ci_update() 464 int num_entries; in bfa_msgq_rspq_pi_update() local 476 num_entries = ntohs(msghdr->num_entries); in bfa_msgq_rspq_pi_update() 483 BFA_MSGQ_INDX_ADD(rspq->consumer_index, num_entries, in bfa_msgq_rspq_pi_update() 626 if (ntohs(cmd->msg_hdr->num_entries) <= in bfa_msgq_cmd_post()
|
/drivers/net/ethernet/mellanox/mlx5/core/diag/ |
D | en_tc_tracepoint.h | 31 f->rule->action.num_entries : 0) 35 f->rule->action.num_entries : 0); 39 f->rule->action.num_entries);
|