/drivers/soc/ti/ |
D | omap_prm.c | 60 const struct omap_prm_domain_map *dmap; member 159 .pwrstctrl = 0x0, .pwrstst = 0x4, .dmap = &omap_prm_reton, 163 .pwrstctrl = 0x0, .pwrstst = 0x4, .dmap = &omap_prm_noinact, 168 .pwrstctrl = 0, .pwrstst = 0x4, .dmap = &omap_prm_all, 172 .pwrstctrl = 0x0, .pwrstst = 0x4, .dmap = &omap_prm_alwon, 176 .pwrstctrl = 0x0, .pwrstst = 0x4, .dmap = &omap_prm_reton, 183 .pwrstctrl = 0x0, .pwrstst = 0x4, .dmap = &omap_prm_noinact, 188 .pwrstctrl = 0x0, .pwrstst = 0x4, .dmap = &omap_prm_onoff_noauto, 192 .pwrstctrl = 0x0, .pwrstst = 0x4, .dmap = &omap_prm_noinact 196 .pwrstctrl = 0x0, .pwrstst = 0x4, .dmap = &omap_prm_onoff_noauto [all …]
|
/drivers/md/ |
D | dm-clone-metadata.c | 127 struct dirty_map dmap[2]; member 473 static int __dirty_map_init(struct dirty_map *dmap, unsigned long nr_words, in __dirty_map_init() argument 476 dmap->changed = 0; in __dirty_map_init() 478 dmap->dirty_words = kvzalloc(bitmap_size(nr_words), GFP_KERNEL); in __dirty_map_init() 479 if (!dmap->dirty_words) in __dirty_map_init() 482 dmap->dirty_regions = kvzalloc(bitmap_size(nr_regions), GFP_KERNEL); in __dirty_map_init() 483 if (!dmap->dirty_regions) { in __dirty_map_init() 484 kvfree(dmap->dirty_words); in __dirty_map_init() 491 static void __dirty_map_exit(struct dirty_map *dmap) in __dirty_map_exit() argument 493 kvfree(dmap->dirty_words); in __dirty_map_exit() [all …]
|
D | dm-zoned-metadata.c | 1703 struct dmz_map *dmap; in dmz_load_mapping() local 1722 dmap = (struct dmz_map *) dmap_mblk->data; in dmz_load_mapping() 1728 dzone_id = le32_to_cpu(dmap[e].dzone_id); in dmz_load_mapping() 1756 bzone_id = le32_to_cpu(dmap[e].bzone_id); in dmz_load_mapping() 1853 struct dmz_map *dmap = (struct dmz_map *) dmap_mblk->data; in dmz_set_chunk_mapping() local 1856 dmap[map_idx].dzone_id = cpu_to_le32(dzone_id); in dmz_set_chunk_mapping() 1857 dmap[map_idx].bzone_id = cpu_to_le32(bzone_id); in dmz_set_chunk_mapping() 2065 struct dmz_map *dmap = (struct dmz_map *) dmap_mblk->data; in dmz_get_chunk_mapping() local 2075 dzone_id = le32_to_cpu(dmap[dmap_idx].dzone_id); in dmz_get_chunk_mapping()
|
/drivers/gpu/drm/ttm/ |
D | ttm_resource.c | 163 struct dma_buf_map *dmap, in ttm_kmap_iter_iomap_map_local() argument 190 dma_buf_map_set_vaddr_iomem(dmap, addr); in ttm_kmap_iter_iomap_map_local() 245 struct dma_buf_map *dmap, in ttm_kmap_iter_linear_io_map_local() argument 251 *dmap = iter_io->dmap; in ttm_kmap_iter_linear_io_map_local() 252 dma_buf_map_incr(dmap, i * PAGE_SIZE); in ttm_kmap_iter_linear_io_map_local() 288 dma_buf_map_set_vaddr(&iter_io->dmap, mem->bus.addr); in ttm_kmap_iter_linear_io_init() 294 memset(&iter_io->dmap, 0, sizeof(iter_io->dmap)); in ttm_kmap_iter_linear_io_init() 296 dma_buf_map_set_vaddr_iomem(&iter_io->dmap, in ttm_kmap_iter_linear_io_init() 300 dma_buf_map_set_vaddr(&iter_io->dmap, in ttm_kmap_iter_linear_io_init() 307 if (dma_buf_map_is_null(&iter_io->dmap)) in ttm_kmap_iter_linear_io_init() [all …]
|
D | ttm_tt.c | 437 struct dma_buf_map *dmap, in ttm_kmap_iter_tt_map_local() argument 443 dma_buf_map_set_vaddr(dmap, kmap_local_page_prot(iter_tt->tt->pages[i], in ttm_kmap_iter_tt_map_local()
|
/drivers/misc/sgi-gru/ |
D | grufault.c | 143 struct gru_tlb_fault_map *dmap) in get_clear_fault_map() argument 158 dmap->fault_bits[i] = k; in get_clear_fault_map() 522 struct gru_tlb_fault_map imap, dmap; in gru_intr() local 536 get_clear_fault_map(gru, &imap, &dmap); in gru_intr() 541 dmap.fault_bits[0], dmap.fault_bits[1]); in gru_intr() 543 for_each_cbr_in_tfm(cbrnum, dmap.fault_bits) { in gru_intr()
|
/drivers/staging/media/ipu3/ |
D | ipu3.c | 76 &imgu_pipe->queues[i].dmap); in imgu_dummybufs_cleanup() 97 &imgu_pipe->queues[i].dmap, size)) { in imgu_dummybufs_preallocate() 137 &imgu_pipe->queues[i].dmap, in imgu_dummybufs_init() 145 imgu_pipe->queues[i].dmap.daddr); in imgu_dummybufs_init() 162 if (WARN_ON(!imgu_pipe->queues[queue].dmap.vaddr)) in imgu_dummybufs_get() 175 imgu_pipe->queues[queue].dmap.daddr); in imgu_dummybufs_get()
|
D | ipu3.h | 105 struct imgu_css_map dmap; member
|
/drivers/edac/ |
D | pnd2_edac.c | 439 static struct d_cr_dmap dmap[DNV_NUM_CHANNELS]; variable 495 RD_REGP(&dmap[i], d_cr_dmap, dnv_dports[i]) || in dnv_get_registers() 1000 daddr->rank = dnv_get_bit(pmiaddr, dmap[pmiidx].rs0 + 13, 0); in dnv_pmi2mem() 1002 daddr->rank |= dnv_get_bit(pmiaddr, dmap[pmiidx].rs1 + 13, 1); in dnv_pmi2mem() 1010 daddr->bank = dnv_get_bit(pmiaddr, dmap[pmiidx].ba0 + 6, 0); in dnv_pmi2mem() 1011 daddr->bank |= dnv_get_bit(pmiaddr, dmap[pmiidx].ba1 + 6, 1); in dnv_pmi2mem() 1012 daddr->bank |= dnv_get_bit(pmiaddr, dmap[pmiidx].bg0 + 6, 2); in dnv_pmi2mem() 1014 daddr->bank |= dnv_get_bit(pmiaddr, dmap[pmiidx].bg1 + 6, 3); in dnv_pmi2mem()
|
/drivers/iommu/ |
D | tegra-smmu.c | 564 dma_addr_t *dmap) in tegra_smmu_pte_lookup() argument 576 *dmap = smmu_pde_to_dma(smmu, pd[pd_index]); in tegra_smmu_pte_lookup() 582 dma_addr_t *dmap, struct page *page) in as_get_pte() argument 609 *dmap = dma; in as_get_pte() 613 *dmap = smmu_pde_to_dma(smmu, pd[pde]); in as_get_pte()
|
/drivers/net/ethernet/apple/ |
D | bmac.c | 179 dbdma_continue(volatile struct dbdma_regs __iomem *dmap) in dbdma_continue() argument 181 dbdma_st32(&dmap->control, in dbdma_continue() 187 dbdma_reset(volatile struct dbdma_regs __iomem *dmap) in dbdma_reset() argument 189 dbdma_st32(&dmap->control, in dbdma_reset() 192 while (dbdma_ld32(&dmap->status) & RUN) in dbdma_reset()
|
/drivers/media/platform/xilinx/ |
D | xilinx-vipp.c | 497 struct xvip_dma *dmap; in xvip_graph_cleanup() local 503 list_for_each_entry_safe(dma, dmap, &xdev->dmas, list) { in xvip_graph_cleanup()
|
/drivers/gpu/drm/vmwgfx/device_include/ |
D | svga3d_types.h | 605 uint32 dmap : 1; member
|