Home
last modified time | relevance | path

Searched refs:regions (Results 1 – 25 of 63) sorted by relevance

123

/drivers/mtd/chips/
Djedec_probe.c272 const uint32_t regions[6]; member
304 .regions = {
316 .regions = {
331 .regions = {
346 .regions = {
361 .regions = {
376 .regions = {
392 .regions = {
409 .regions = {
426 .regions = {
[all …]
Dcfi_util.c360 struct mtd_erase_region_info *regions = mtd->eraseregions; in cfi_varsize_frob() local
374 while (i < mtd->numeraseregions && ofs >= regions[i].offset) in cfi_varsize_frob()
384 if (ofs & (regions[i].erasesize-1)) in cfi_varsize_frob()
394 while (i<mtd->numeraseregions && (ofs + len) >= regions[i].offset) in cfi_varsize_frob()
402 if ((ofs + len) & (regions[i].erasesize-1)) in cfi_varsize_frob()
411 int size = regions[i].erasesize; in cfi_varsize_frob()
422 if (ofs == regions[i].offset + size * regions[i].numblocks) in cfi_varsize_frob()
Dcfi_cmdset_0020.c896 struct mtd_erase_region_info *regions = mtd->eraseregions; in cfi_staa_erase_varsize() local
910 while (i < mtd->numeraseregions && instr->addr >= regions[i].offset) in cfi_staa_erase_varsize()
920 if (instr->addr & (regions[i].erasesize-1)) in cfi_staa_erase_varsize()
930 while (i<mtd->numeraseregions && (instr->addr + instr->len) >= regions[i].offset) in cfi_staa_erase_varsize()
938 if ((instr->addr + instr->len) & (regions[i].erasesize-1)) in cfi_staa_erase_varsize()
953 adr += regions[i].erasesize; in cfi_staa_erase_varsize()
954 len -= regions[i].erasesize; in cfi_staa_erase_varsize()
956 …if (adr % (1<< cfi->chipshift) == (((unsigned long)regions[i].offset + (regions[i].erasesize * reg… in cfi_staa_erase_varsize()
/drivers/vfio/platform/
Dvfio_platform_common.c150 vdev->regions = kcalloc(cnt, sizeof(struct vfio_platform_region), in vfio_platform_regions_init()
152 if (!vdev->regions) in vfio_platform_regions_init()
162 vdev->regions[i].addr = res->start; in vfio_platform_regions_init()
163 vdev->regions[i].size = resource_size(res); in vfio_platform_regions_init()
164 vdev->regions[i].flags = 0; in vfio_platform_regions_init()
168 vdev->regions[i].type = VFIO_PLATFORM_REGION_TYPE_MMIO; in vfio_platform_regions_init()
169 vdev->regions[i].flags |= VFIO_REGION_INFO_FLAG_READ; in vfio_platform_regions_init()
171 vdev->regions[i].flags |= in vfio_platform_regions_init()
178 if (!(vdev->regions[i].addr & ~PAGE_MASK) && in vfio_platform_regions_init()
179 !(vdev->regions[i].size & ~PAGE_MASK)) in vfio_platform_regions_init()
[all …]
/drivers/staging/fsl-mc/bus/
Dfsl-mc-bus.c375 struct resource *regions; in fsl_mc_device_get_mmio_regions() local
395 regions = kmalloc_array(obj_desc->region_count, in fsl_mc_device_get_mmio_regions()
396 sizeof(regions[0]), GFP_KERNEL); in fsl_mc_device_get_mmio_regions()
397 if (!regions) in fsl_mc_device_get_mmio_regions()
417 &regions[i].start); in fsl_mc_device_get_mmio_regions()
426 regions[i].end = regions[i].start + region_desc.size - 1; in fsl_mc_device_get_mmio_regions()
427 regions[i].name = "fsl-mc object MMIO region"; in fsl_mc_device_get_mmio_regions()
428 regions[i].flags = IORESOURCE_IO; in fsl_mc_device_get_mmio_regions()
430 regions[i].flags |= IORESOURCE_CACHEABLE; in fsl_mc_device_get_mmio_regions()
433 mc_dev->regions = regions; in fsl_mc_device_get_mmio_regions()
[all …]
Dmc-io.c245 mc_portal_phys_addr = dpmcp_dev->regions[0].start; in fsl_mc_portal_allocate()
246 mc_portal_size = dpmcp_dev->regions[0].end - in fsl_mc_portal_allocate()
247 dpmcp_dev->regions[0].start + 1; in fsl_mc_portal_allocate()
/drivers/soc/qcom/
Dsmem.c244 struct smem_region regions[0]; member
355 header = smem->regions[0].virt_base; in qcom_smem_alloc_global()
434 header = smem->regions[0].virt_base; in qcom_smem_get_global()
442 area = &smem->regions[i]; in qcom_smem_get_global()
545 header = __smem->regions[0].virt_base; in qcom_smem_get_free_space()
582 ptable = smem->regions[0].virt_base + smem->regions[0].size - SZ_4K; in qcom_smem_enumerate_partitions()
626 header = smem->regions[0].virt_base + le32_to_cpu(entry->offset); in qcom_smem_enumerate_partitions()
685 smem->regions[i].aux_base = (u32)r.start; in qcom_smem_map_memory()
686 smem->regions[i].size = resource_size(&r); in qcom_smem_map_memory()
687 smem->regions[i].virt_base = devm_ioremap_wc(dev, r.start, resource_size(&r)); in qcom_smem_map_memory()
[all …]
/drivers/staging/android/
DTODO23 - Add dt-binding to expose cma regions as ion heaps, with the rule that any
24 such cma regions must already be used by some device for dma. I.e. ion only
25 exposes existing cma regions and doesn't reserve unecessarily memory when
38 region. This will cause false wakeups in regions with a large number of
41 - Add debugfs support for examining the permissions of regions.
Dvsoc.c108 struct vsoc_device_region *regions; member
217 return &vsoc_dev.regions[iminor(inode)]; in vsoc_region_from_inode()
286 if (&vsoc_dev.regions[managed_region_p->managed_by] != region_p) in do_create_fd_scoped_permission()
854 vsoc_dev.regions = (struct vsoc_device_region *) in vsoc_probe_device()
887 const struct vsoc_device_region *region = vsoc_dev.regions + i; in vsoc_probe_device()
917 const struct vsoc_device_region *region = vsoc_dev.regions + i; in vsoc_probe_device()
1008 vsoc_dev.regions = 0; in vsoc_remove_device()
1094 &vsoc_dev.regions[owner_region_p->managed_by]; in vsoc_release()
/drivers/gpu/drm/nouveau/nvkm/subdev/fb/
Dbase.c133 for (i = 0; i < fb->tile.regions; i++) in nvkm_fb_init()
154 for (i = 0; i < fb->tile.regions; i++) in nvkm_fb_dtor()
178 fb->tile.regions = fb->func->tile.regions; in nvkm_fb_ctor()
Dnv1a.c31 .tile.regions = 8,
Dnv4e.c32 .tile.regions = 12,
Dnv47.c32 .tile.regions = 15,
Dnv49.c32 .tile.regions = 15,
Dramnv50.c501 while (!list_empty(&mem->regions)) { in __nv50_ram_put()
502 this = list_first_entry(&mem->regions, typeof(*this), rl_entry); in __nv50_ram_put()
562 INIT_LIST_HEAD(&mem->regions); in nv50_ram_get()
578 list_add_tail(&r->rl_entry, &mem->regions); in nv50_ram_get()
583 r = list_first_entry(&mem->regions, struct nvkm_mm_node, rl_entry); in nv50_ram_get()
Dnv46.c46 .tile.regions = 15,
/drivers/video/console/
Dsticore.c460 REGION_OFFSET_TO_PHYS(sti->regions[i], newhpa); in sti_init_glob_cfg()
462 len = sti->regions[i].region_desc.length * 4096; in sti_init_glob_cfg()
470 sti->regions[i].region_desc.btlb, in sti_init_glob_cfg()
471 sti->regions[i].region_desc.sys_only, in sti_init_glob_cfg()
472 sti->regions[i].region_desc.cache, in sti_init_glob_cfg()
473 sti->regions[i].region_desc.last)); in sti_init_glob_cfg()
476 if (sti->regions[i].region_desc.last) in sti_init_glob_cfg()
480 if (++i<8 && sti->regions[i].region) in sti_init_glob_cfg()
482 __FILE__, sti->regions[i].region); in sti_init_glob_cfg()
754 memcpy(sti->regions, ((void *)raw)+raw->region_list, sizeof(sti->regions)); in sti_read_rom()
/drivers/char/
Dps3flash.c102 dev->regions[dev->region_idx].size*dev->blk_size); in ps3flash_llseek()
119 size = dev->regions[dev->region_idx].size*dev->blk_size; in ps3flash_read()
188 size = dev->regions[dev->region_idx].size*dev->blk_size; in ps3flash_write()
349 tmp = dev->regions[dev->region_idx].start*dev->blk_size; in ps3flash_probe()
356 tmp = dev->regions[dev->region_idx].size*dev->blk_size; in ps3flash_probe()
/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/
Dgk20a.c312 if (unlikely(list_empty(&node->base.mem.regions))) in gk20a_instobj_dtor_iommu()
323 r = list_first_entry(&node->base.mem.regions, struct nvkm_mm_node, in gk20a_instobj_dtor_iommu()
408 INIT_LIST_HEAD(&node->base.mem.regions); in gk20a_instobj_ctor_dma()
409 list_add_tail(&node->r.rl_entry, &node->base.mem.regions); in gk20a_instobj_ctor_dma()
488 INIT_LIST_HEAD(&node->base.mem.regions); in gk20a_instobj_ctor_iommu()
489 list_add_tail(&r->rl_entry, &node->base.mem.regions); in gk20a_instobj_ctor_iommu()
/drivers/vfio/platform/reset/
Dvfio_platform_amdxgbe.c61 struct vfio_platform_region *xgmac_regs = &vdev->regions[0]; in vfio_platform_amdxgbe_reset()
62 struct vfio_platform_region *xpcs_regs = &vdev->regions[1]; in vfio_platform_amdxgbe_reset()
/drivers/soc/ti/
Dknav_qmss.h303 struct list_head regions; member
349 list_for_each_entry(region, &kdev->regions, list)
352 list_first_entry_or_null(&kdev->regions, \
Dknav_qmss_queue.c1065 struct device_node *regions) in knav_queue_setup_regions() argument
1073 for_each_child_of_node(regions, child) { in knav_queue_setup_regions()
1107 list_add_tail(&region->list, &kdev->regions); in knav_queue_setup_regions()
1109 if (list_empty(&kdev->regions)) { in knav_queue_setup_regions()
1693 struct device_node *qmgrs, *queue_pools, *regions, *pdsps; in knav_queue_probe() local
1714 INIT_LIST_HEAD(&kdev->regions); in knav_queue_probe()
1787 regions = of_get_child_by_name(node, "descriptor-regions"); in knav_queue_probe()
1788 if (!regions) { in knav_queue_probe()
1792 ret = knav_queue_setup_regions(kdev, regions); in knav_queue_probe()
1793 of_node_put(regions); in knav_queue_probe()
/drivers/staging/fsl-mc/
DREADME.txt78 the MC and are done directly using memory mapped regions in
110 regions and interrupts).
175 supports and a summary of key resources of the object (MMIO regions
181 -MMIO regions: none
190 -MMIO regions: none
204 -MMIO regions: queue operations, buffer management
211 -MMIO regions: none
218 -MMIO regions: MC command portal
/drivers/gpu/drm/nouveau/include/nvkm/subdev/
Dfb.h32 struct list_head regions; member
56 int regions; member
/drivers/ps3/
Dps3stor_lib.c134 dev->region_idx, dev->regions[dev->region_idx].start, in ps3stor_probe_access()
135 dev->regions[dev->region_idx].size); in ps3stor_probe_access()
279 unsigned int region_id = dev->regions[dev->region_idx].id; in ps3stor_read_write_sectors()

123