/drivers/mtd/chips/ |
D | jedec_probe.c | 272 const uint32_t regions[6]; member 304 .regions = { 316 .regions = { 331 .regions = { 346 .regions = { 361 .regions = { 376 .regions = { 392 .regions = { 409 .regions = { 426 .regions = { [all …]
|
D | cfi_util.c | 360 struct mtd_erase_region_info *regions = mtd->eraseregions; in cfi_varsize_frob() local 374 while (i < mtd->numeraseregions && ofs >= regions[i].offset) in cfi_varsize_frob() 384 if (ofs & (regions[i].erasesize-1)) in cfi_varsize_frob() 394 while (i<mtd->numeraseregions && (ofs + len) >= regions[i].offset) in cfi_varsize_frob() 402 if ((ofs + len) & (regions[i].erasesize-1)) in cfi_varsize_frob() 411 int size = regions[i].erasesize; in cfi_varsize_frob() 422 if (ofs == regions[i].offset + size * regions[i].numblocks) in cfi_varsize_frob()
|
D | cfi_cmdset_0020.c | 896 struct mtd_erase_region_info *regions = mtd->eraseregions; in cfi_staa_erase_varsize() local 910 while (i < mtd->numeraseregions && instr->addr >= regions[i].offset) in cfi_staa_erase_varsize() 920 if (instr->addr & (regions[i].erasesize-1)) in cfi_staa_erase_varsize() 930 while (i<mtd->numeraseregions && (instr->addr + instr->len) >= regions[i].offset) in cfi_staa_erase_varsize() 938 if ((instr->addr + instr->len) & (regions[i].erasesize-1)) in cfi_staa_erase_varsize() 953 adr += regions[i].erasesize; in cfi_staa_erase_varsize() 954 len -= regions[i].erasesize; in cfi_staa_erase_varsize() 956 …if (adr % (1<< cfi->chipshift) == (((unsigned long)regions[i].offset + (regions[i].erasesize * reg… in cfi_staa_erase_varsize()
|
/drivers/vfio/platform/ |
D | vfio_platform_common.c | 150 vdev->regions = kcalloc(cnt, sizeof(struct vfio_platform_region), in vfio_platform_regions_init() 152 if (!vdev->regions) in vfio_platform_regions_init() 162 vdev->regions[i].addr = res->start; in vfio_platform_regions_init() 163 vdev->regions[i].size = resource_size(res); in vfio_platform_regions_init() 164 vdev->regions[i].flags = 0; in vfio_platform_regions_init() 168 vdev->regions[i].type = VFIO_PLATFORM_REGION_TYPE_MMIO; in vfio_platform_regions_init() 169 vdev->regions[i].flags |= VFIO_REGION_INFO_FLAG_READ; in vfio_platform_regions_init() 171 vdev->regions[i].flags |= in vfio_platform_regions_init() 178 if (!(vdev->regions[i].addr & ~PAGE_MASK) && in vfio_platform_regions_init() 179 !(vdev->regions[i].size & ~PAGE_MASK)) in vfio_platform_regions_init() [all …]
|
/drivers/staging/fsl-mc/bus/ |
D | fsl-mc-bus.c | 375 struct resource *regions; in fsl_mc_device_get_mmio_regions() local 395 regions = kmalloc_array(obj_desc->region_count, in fsl_mc_device_get_mmio_regions() 396 sizeof(regions[0]), GFP_KERNEL); in fsl_mc_device_get_mmio_regions() 397 if (!regions) in fsl_mc_device_get_mmio_regions() 417 ®ions[i].start); in fsl_mc_device_get_mmio_regions() 426 regions[i].end = regions[i].start + region_desc.size - 1; in fsl_mc_device_get_mmio_regions() 427 regions[i].name = "fsl-mc object MMIO region"; in fsl_mc_device_get_mmio_regions() 428 regions[i].flags = IORESOURCE_IO; in fsl_mc_device_get_mmio_regions() 430 regions[i].flags |= IORESOURCE_CACHEABLE; in fsl_mc_device_get_mmio_regions() 433 mc_dev->regions = regions; in fsl_mc_device_get_mmio_regions() [all …]
|
D | mc-io.c | 245 mc_portal_phys_addr = dpmcp_dev->regions[0].start; in fsl_mc_portal_allocate() 246 mc_portal_size = dpmcp_dev->regions[0].end - in fsl_mc_portal_allocate() 247 dpmcp_dev->regions[0].start + 1; in fsl_mc_portal_allocate()
|
/drivers/soc/qcom/ |
D | smem.c | 244 struct smem_region regions[0]; member 355 header = smem->regions[0].virt_base; in qcom_smem_alloc_global() 434 header = smem->regions[0].virt_base; in qcom_smem_get_global() 442 area = &smem->regions[i]; in qcom_smem_get_global() 545 header = __smem->regions[0].virt_base; in qcom_smem_get_free_space() 582 ptable = smem->regions[0].virt_base + smem->regions[0].size - SZ_4K; in qcom_smem_enumerate_partitions() 626 header = smem->regions[0].virt_base + le32_to_cpu(entry->offset); in qcom_smem_enumerate_partitions() 685 smem->regions[i].aux_base = (u32)r.start; in qcom_smem_map_memory() 686 smem->regions[i].size = resource_size(&r); in qcom_smem_map_memory() 687 smem->regions[i].virt_base = devm_ioremap_wc(dev, r.start, resource_size(&r)); in qcom_smem_map_memory() [all …]
|
/drivers/staging/android/ |
D | TODO | 23 - Add dt-binding to expose cma regions as ion heaps, with the rule that any 24 such cma regions must already be used by some device for dma. I.e. ion only 25 exposes existing cma regions and doesn't reserve unecessarily memory when 38 region. This will cause false wakeups in regions with a large number of 41 - Add debugfs support for examining the permissions of regions.
|
D | vsoc.c | 108 struct vsoc_device_region *regions; member 217 return &vsoc_dev.regions[iminor(inode)]; in vsoc_region_from_inode() 286 if (&vsoc_dev.regions[managed_region_p->managed_by] != region_p) in do_create_fd_scoped_permission() 854 vsoc_dev.regions = (struct vsoc_device_region *) in vsoc_probe_device() 887 const struct vsoc_device_region *region = vsoc_dev.regions + i; in vsoc_probe_device() 917 const struct vsoc_device_region *region = vsoc_dev.regions + i; in vsoc_probe_device() 1008 vsoc_dev.regions = 0; in vsoc_remove_device() 1094 &vsoc_dev.regions[owner_region_p->managed_by]; in vsoc_release()
|
/drivers/gpu/drm/nouveau/nvkm/subdev/fb/ |
D | base.c | 133 for (i = 0; i < fb->tile.regions; i++) in nvkm_fb_init() 154 for (i = 0; i < fb->tile.regions; i++) in nvkm_fb_dtor() 178 fb->tile.regions = fb->func->tile.regions; in nvkm_fb_ctor()
|
D | nv1a.c | 31 .tile.regions = 8,
|
D | nv4e.c | 32 .tile.regions = 12,
|
D | nv47.c | 32 .tile.regions = 15,
|
D | nv49.c | 32 .tile.regions = 15,
|
D | ramnv50.c | 501 while (!list_empty(&mem->regions)) { in __nv50_ram_put() 502 this = list_first_entry(&mem->regions, typeof(*this), rl_entry); in __nv50_ram_put() 562 INIT_LIST_HEAD(&mem->regions); in nv50_ram_get() 578 list_add_tail(&r->rl_entry, &mem->regions); in nv50_ram_get() 583 r = list_first_entry(&mem->regions, struct nvkm_mm_node, rl_entry); in nv50_ram_get()
|
D | nv46.c | 46 .tile.regions = 15,
|
/drivers/video/console/ |
D | sticore.c | 460 REGION_OFFSET_TO_PHYS(sti->regions[i], newhpa); in sti_init_glob_cfg() 462 len = sti->regions[i].region_desc.length * 4096; in sti_init_glob_cfg() 470 sti->regions[i].region_desc.btlb, in sti_init_glob_cfg() 471 sti->regions[i].region_desc.sys_only, in sti_init_glob_cfg() 472 sti->regions[i].region_desc.cache, in sti_init_glob_cfg() 473 sti->regions[i].region_desc.last)); in sti_init_glob_cfg() 476 if (sti->regions[i].region_desc.last) in sti_init_glob_cfg() 480 if (++i<8 && sti->regions[i].region) in sti_init_glob_cfg() 482 __FILE__, sti->regions[i].region); in sti_init_glob_cfg() 754 memcpy(sti->regions, ((void *)raw)+raw->region_list, sizeof(sti->regions)); in sti_read_rom()
|
/drivers/char/ |
D | ps3flash.c | 102 dev->regions[dev->region_idx].size*dev->blk_size); in ps3flash_llseek() 119 size = dev->regions[dev->region_idx].size*dev->blk_size; in ps3flash_read() 188 size = dev->regions[dev->region_idx].size*dev->blk_size; in ps3flash_write() 349 tmp = dev->regions[dev->region_idx].start*dev->blk_size; in ps3flash_probe() 356 tmp = dev->regions[dev->region_idx].size*dev->blk_size; in ps3flash_probe()
|
/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/ |
D | gk20a.c | 312 if (unlikely(list_empty(&node->base.mem.regions))) in gk20a_instobj_dtor_iommu() 323 r = list_first_entry(&node->base.mem.regions, struct nvkm_mm_node, in gk20a_instobj_dtor_iommu() 408 INIT_LIST_HEAD(&node->base.mem.regions); in gk20a_instobj_ctor_dma() 409 list_add_tail(&node->r.rl_entry, &node->base.mem.regions); in gk20a_instobj_ctor_dma() 488 INIT_LIST_HEAD(&node->base.mem.regions); in gk20a_instobj_ctor_iommu() 489 list_add_tail(&r->rl_entry, &node->base.mem.regions); in gk20a_instobj_ctor_iommu()
|
/drivers/vfio/platform/reset/ |
D | vfio_platform_amdxgbe.c | 61 struct vfio_platform_region *xgmac_regs = &vdev->regions[0]; in vfio_platform_amdxgbe_reset() 62 struct vfio_platform_region *xpcs_regs = &vdev->regions[1]; in vfio_platform_amdxgbe_reset()
|
/drivers/soc/ti/ |
D | knav_qmss.h | 303 struct list_head regions; member 349 list_for_each_entry(region, &kdev->regions, list) 352 list_first_entry_or_null(&kdev->regions, \
|
D | knav_qmss_queue.c | 1065 struct device_node *regions) in knav_queue_setup_regions() argument 1073 for_each_child_of_node(regions, child) { in knav_queue_setup_regions() 1107 list_add_tail(®ion->list, &kdev->regions); in knav_queue_setup_regions() 1109 if (list_empty(&kdev->regions)) { in knav_queue_setup_regions() 1693 struct device_node *qmgrs, *queue_pools, *regions, *pdsps; in knav_queue_probe() local 1714 INIT_LIST_HEAD(&kdev->regions); in knav_queue_probe() 1787 regions = of_get_child_by_name(node, "descriptor-regions"); in knav_queue_probe() 1788 if (!regions) { in knav_queue_probe() 1792 ret = knav_queue_setup_regions(kdev, regions); in knav_queue_probe() 1793 of_node_put(regions); in knav_queue_probe()
|
/drivers/staging/fsl-mc/ |
D | README.txt | 78 the MC and are done directly using memory mapped regions in 110 regions and interrupts). 175 supports and a summary of key resources of the object (MMIO regions 181 -MMIO regions: none 190 -MMIO regions: none 204 -MMIO regions: queue operations, buffer management 211 -MMIO regions: none 218 -MMIO regions: MC command portal
|
/drivers/gpu/drm/nouveau/include/nvkm/subdev/ |
D | fb.h | 32 struct list_head regions; member 56 int regions; member
|
/drivers/ps3/ |
D | ps3stor_lib.c | 134 dev->region_idx, dev->regions[dev->region_idx].start, in ps3stor_probe_access() 135 dev->regions[dev->region_idx].size); in ps3stor_probe_access() 279 unsigned int region_id = dev->regions[dev->region_idx].id; in ps3stor_read_write_sectors()
|