/drivers/virt/nitro_enclaves/ |
D | ne_misc_dev.c | 738 struct ne_user_memory_region mem_region) in ne_sanity_check_user_mem_region() argument 745 if (mem_region.memory_size & (NE_MIN_MEM_REGION_SIZE - 1)) { in ne_sanity_check_user_mem_region() 752 if (!IS_ALIGNED(mem_region.userspace_addr, NE_MIN_MEM_REGION_SIZE)) { in ne_sanity_check_user_mem_region() 759 if ((mem_region.userspace_addr & (NE_MIN_MEM_REGION_SIZE - 1)) || in ne_sanity_check_user_mem_region() 760 !access_ok((void __user *)(unsigned long)mem_region.userspace_addr, in ne_sanity_check_user_mem_region() 761 mem_region.memory_size)) { in ne_sanity_check_user_mem_region() 773 if ((userspace_addr <= mem_region.userspace_addr && in ne_sanity_check_user_mem_region() 774 mem_region.userspace_addr < (userspace_addr + memory_size)) || in ne_sanity_check_user_mem_region() 775 (mem_region.userspace_addr <= userspace_addr && in ne_sanity_check_user_mem_region() 776 (mem_region.userspace_addr + mem_region.memory_size) > userspace_addr)) { in ne_sanity_check_user_mem_region() [all …]
|
/drivers/soc/qcom/ |
D | mdt_loader.c | 130 const char *firmware, int pas_id, void *mem_region, in __qcom_mdt_load() argument 151 if (!fw || !mem_region || !mem_phys || !mem_size) in __qcom_mdt_load() 241 ptr = mem_region + offset; in __qcom_mdt_load() 303 const char *firmware, int pas_id, void *mem_region, in qcom_mdt_load() argument 307 return __qcom_mdt_load(dev, fw, firmware, pas_id, mem_region, mem_phys, in qcom_mdt_load() 327 void *mem_region, phys_addr_t mem_phys, in qcom_mdt_load_no_init() argument 330 return __qcom_mdt_load(dev, fw, firmware, pas_id, mem_region, mem_phys, in qcom_mdt_load_no_init()
|
/drivers/staging/android/ |
D | debug_kinfo.c | 96 struct device_node *mem_region; in debug_kinfo_probe() local 101 mem_region = of_parse_phandle(pdev->dev.of_node, "memory-region", 0); in debug_kinfo_probe() 102 if (!mem_region) { in debug_kinfo_probe() 107 rmem = of_reserved_mem_lookup(mem_region); in debug_kinfo_probe()
|
/drivers/net/wireless/ath/ath10k/ |
D | qmi.c | 91 ret = ath10k_qmi_map_msa_permission(qmi, &qmi->mem_region[i]); in ath10k_qmi_setup_msa_permissions() 100 ath10k_qmi_unmap_msa_permission(qmi, &qmi->mem_region[i]); in ath10k_qmi_setup_msa_permissions() 112 ath10k_qmi_unmap_msa_permission(qmi, &qmi->mem_region[i]); in ath10k_qmi_remove_msa_permission() 174 qmi->mem_region[i].addr = resp.mem_region_info[i].region_addr; in ath10k_qmi_msa_mem_info_send_sync_msg() 175 qmi->mem_region[i].size = resp.mem_region_info[i].size; in ath10k_qmi_msa_mem_info_send_sync_msg() 176 qmi->mem_region[i].secure = resp.mem_region_info[i].secure_flag; in ath10k_qmi_msa_mem_info_send_sync_msg() 179 i, &qmi->mem_region[i].addr, in ath10k_qmi_msa_mem_info_send_sync_msg() 180 qmi->mem_region[i].size, in ath10k_qmi_msa_mem_info_send_sync_msg() 181 qmi->mem_region[i].secure); in ath10k_qmi_msa_mem_info_send_sync_msg() 188 memset(&qmi->mem_region[0], 0, sizeof(qmi->mem_region[0]) * i); in ath10k_qmi_msa_mem_info_send_sync_msg()
|
D | coredump.c | 1423 const struct ath10k_mem_region *mem_region; in ath10k_coredump_get_ramdump_size() local 1432 mem_region = &hw->region_table.regions[0]; in ath10k_coredump_get_ramdump_size() 1435 size += mem_region->len; in ath10k_coredump_get_ramdump_size() 1436 mem_region++; in ath10k_coredump_get_ramdump_size()
|
D | qmi.h | 100 struct ath10k_msa_mem_info mem_region[MAX_NUM_MEMORY_REGIONS]; member
|
D | sdio.c | 2284 const struct ath10k_mem_region *mem_region, in ath10k_sdio_dump_memory_section() argument 2291 if (!mem_region || !buf) in ath10k_sdio_dump_memory_section() 2294 cur_section = &mem_region->section_table.sections[0]; in ath10k_sdio_dump_memory_section() 2296 if (mem_region->start > cur_section->start) { in ath10k_sdio_dump_memory_section() 2298 mem_region->start, cur_section->start); in ath10k_sdio_dump_memory_section() 2302 skip_size = cur_section->start - mem_region->start; in ath10k_sdio_dump_memory_section() 2324 if ((i + 1) == mem_region->section_table.size) { in ath10k_sdio_dump_memory_section()
|
D | pci.c | 1469 const struct ath10k_mem_region *mem_region, in ath10k_pci_dump_memory_section() argument 1476 if (!mem_region || !buf) in ath10k_pci_dump_memory_section() 1479 cur_section = &mem_region->section_table.sections[0]; in ath10k_pci_dump_memory_section() 1481 if (mem_region->start > cur_section->start) { in ath10k_pci_dump_memory_section() 1483 mem_region->start, cur_section->start); in ath10k_pci_dump_memory_section() 1487 skip_size = cur_section->start - mem_region->start; in ath10k_pci_dump_memory_section() 1509 if ((i + 1) == mem_region->section_table.size) { in ath10k_pci_dump_memory_section()
|
/drivers/remoteproc/ |
D | qcom_q6v5_adsp.c | 93 void *mem_region; member 172 adsp->mem_region, adsp->mem_phys, in adsp_load() 293 return adsp->mem_region + offset; in adsp_da_to_va() 415 adsp->mem_region = devm_ioremap_wc(adsp->dev, in adsp_alloc_memory_region() 417 if (!adsp->mem_region) { in adsp_alloc_memory_region()
|
D | qcom_wcnss.c | 91 void *mem_region; member 152 wcnss->mem_region, wcnss->mem_phys, in wcnss_load() 307 return wcnss->mem_region + offset; in wcnss_da_to_va() 463 wcnss->mem_region = devm_ioremap_wc(wcnss->dev, wcnss->mem_phys, wcnss->mem_size); in wcnss_alloc_memory_region() 464 if (!wcnss->mem_region) { in wcnss_alloc_memory_region()
|
D | qcom_q6v5_wcss.c | 93 void *mem_region; member 422 return wcss->mem_region + offset; in q6v5_wcss_da_to_va() 431 0, wcss->mem_region, wcss->mem_phys, in q6v5_wcss_load() 529 wcss->mem_region = devm_ioremap_wc(dev, wcss->mem_phys, wcss->mem_size); in q6v5_alloc_memory_region() 530 if (!wcss->mem_region) { in q6v5_alloc_memory_region()
|
D | qcom_q6v5_pas.c | 75 void *mem_region; member 128 adsp->mem_region, adsp->mem_phys, adsp->mem_size, in adsp_load() 245 return adsp->mem_region + offset; in adsp_da_to_va() 374 adsp->mem_region = devm_ioremap_wc(adsp->dev, adsp->mem_phys, adsp->mem_size); in adsp_alloc_memory_region() 375 if (!adsp->mem_region) { in adsp_alloc_memory_region()
|
/drivers/gpu/drm/msm/adreno/ |
D | adreno_gpu.c | 34 void *mem_region = NULL; in zap_shader_load_mdt() local 119 mem_region = memremap(mem_phys, mem_size, MEMREMAP_WC); in zap_shader_load_mdt() 120 if (!mem_region) { in zap_shader_load_mdt() 136 mem_region, mem_phys, mem_size, NULL); in zap_shader_load_mdt() 143 mem_region, mem_phys, mem_size, NULL); in zap_shader_load_mdt() 162 if (mem_region) in zap_shader_load_mdt() 163 memunmap(mem_region); in zap_shader_load_mdt()
|
/drivers/mtd/nand/raw/ |
D | stm32_fmc2_nand.c | 1859 int chip_cs, mem_region, ret, irq; in stm32_fmc2_nfc_probe() local 1891 for (chip_cs = 0, mem_region = start_region; chip_cs < FMC2_MAX_CE; in stm32_fmc2_nfc_probe() 1892 chip_cs++, mem_region += 3) { in stm32_fmc2_nfc_probe() 1896 res = platform_get_resource(pdev, IORESOURCE_MEM, mem_region); in stm32_fmc2_nfc_probe() 1904 mem_region + 1); in stm32_fmc2_nfc_probe() 1910 mem_region + 2); in stm32_fmc2_nfc_probe()
|
/drivers/net/ethernet/chelsio/cxgb3/ |
D | t3_hw.c | 2594 #define mem_region(adap, start, size, reg) \ macro 2644 mem_region(adap, m, (64 << 10) * 64, SG_EGR_CNTX_BADDR); in partition_mem() 2645 mem_region(adap, m, (64 << 10) * 64, SG_CQ_CONTEXT_BADDR); in partition_mem() 2648 mem_region(adap, m, pstructs * 64, TP_CMM_MM_BASE); in partition_mem() 2649 mem_region(adap, m, 64 * (pstructs / 24), TP_CMM_MM_PS_FLST_BASE); in partition_mem() 2650 mem_region(adap, m, 64 * (p->rx_num_pgs / 24), TP_CMM_MM_RX_FLST_BASE); in partition_mem() 2651 mem_region(adap, m, 64 * (p->tx_num_pgs / 24), TP_CMM_MM_TX_FLST_BASE); in partition_mem()
|
/drivers/firmware/ |
D | qcom_scm.c | 793 static int __qcom_scm_assign_mem(struct device *dev, phys_addr_t mem_region, in __qcom_scm_assign_mem() argument 804 .args[0] = mem_region, in __qcom_scm_assign_mem()
|