Home
last modified time | relevance | path

Searched refs:mapping (Results 1 – 25 of 1219) sorted by relevance

12345678910>>...49

/kernel/linux/linux-5.10/drivers/staging/gasket/
Dgasket_sysfs.c77 static void put_mapping(struct gasket_sysfs_mapping *mapping) in put_mapping() argument
84 if (!mapping) { in put_mapping()
89 mutex_lock(&mapping->mutex); in put_mapping()
90 if (kref_put(&mapping->refcount, release_entry)) { in put_mapping()
91 dev_dbg(mapping->device, "Removing Gasket sysfs mapping\n"); in put_mapping()
99 device = mapping->device; in put_mapping()
100 num_files_to_remove = mapping->attribute_count; in put_mapping()
107 mapping->attributes[i].attr; in put_mapping()
111 kfree(mapping->attributes); in put_mapping()
112 mapping->attributes = NULL; in put_mapping()
[all …]
/kernel/linux/linux-5.10/mm/
Dtruncate.c34 static inline void __clear_shadow_entry(struct address_space *mapping, in __clear_shadow_entry() argument
37 XA_STATE(xas, &mapping->i_pages, index); in __clear_shadow_entry()
43 mapping->nrexceptional--; in __clear_shadow_entry()
46 static void clear_shadow_entry(struct address_space *mapping, pgoff_t index, in clear_shadow_entry() argument
49 xa_lock_irq(&mapping->i_pages); in clear_shadow_entry()
50 __clear_shadow_entry(mapping, index, entry); in clear_shadow_entry()
51 xa_unlock_irq(&mapping->i_pages); in clear_shadow_entry()
59 static void truncate_exceptional_pvec_entries(struct address_space *mapping, in truncate_exceptional_pvec_entries() argument
67 if (shmem_mapping(mapping)) in truncate_exceptional_pvec_entries()
77 dax = dax_mapping(mapping); in truncate_exceptional_pvec_entries()
[all …]
Dfilemap.c120 static void page_cache_delete(struct address_space *mapping, in page_cache_delete() argument
123 XA_STATE(xas, &mapping->i_pages, page->index); in page_cache_delete()
126 mapping_set_update(&xas, mapping); in page_cache_delete()
141 page->mapping = NULL; in page_cache_delete()
145 mapping->nrexceptional += nr; in page_cache_delete()
154 mapping->nrpages -= nr; in page_cache_delete()
157 static void unaccount_page_cache_page(struct address_space *mapping, in unaccount_page_cache_page() argument
170 cleancache_invalidate_page(mapping, page); in unaccount_page_cache_page()
184 if (mapping_exiting(mapping) && in unaccount_page_cache_page()
210 filemap_nr_thps_dec(mapping); in unaccount_page_cache_page()
[all …]
Dreadahead.c34 file_ra_state_init(struct file_ra_state *ra, struct address_space *mapping) in file_ra_state_init() argument
36 ra->ra_pages = inode_to_bdi(mapping->host)->ra_pages; in file_ra_state_init()
48 static void read_cache_pages_invalidate_page(struct address_space *mapping, in read_cache_pages_invalidate_page() argument
54 page->mapping = mapping; in read_cache_pages_invalidate_page()
56 page->mapping = NULL; in read_cache_pages_invalidate_page()
65 static void read_cache_pages_invalidate_pages(struct address_space *mapping, in read_cache_pages_invalidate_pages() argument
73 read_cache_pages_invalidate_page(mapping, victim); in read_cache_pages_invalidate_pages()
89 int read_cache_pages(struct address_space *mapping, struct list_head *pages, in read_cache_pages() argument
98 if (add_to_page_cache_lru(page, mapping, page->index, in read_cache_pages()
99 readahead_gfp_mask(mapping))) { in read_cache_pages()
[all …]
Dpage-writeback.c1872 void balance_dirty_pages_ratelimited(struct address_space *mapping) in balance_dirty_pages_ratelimited() argument
1874 struct inode *inode = mapping->host; in balance_dirty_pages_ratelimited()
2110 void tag_pages_for_writeback(struct address_space *mapping, in tag_pages_for_writeback() argument
2113 XA_STATE(xas, &mapping->i_pages, start); in tag_pages_for_writeback()
2163 int write_cache_pages(struct address_space *mapping, in write_cache_pages() argument
2180 index = mapping->writeback_index; /* prev offset */ in write_cache_pages()
2189 tag_pages_for_writeback(mapping, index, end); in write_cache_pages()
2198 nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, end, in write_cache_pages()
2218 if (unlikely(page->mapping != mapping)) { in write_cache_pages()
2240 trace_wbc_writepage(wbc, inode_to_bdi(mapping->host)); in write_cache_pages()
[all …]
/kernel/linux/linux-5.10/include/linux/
Dpagemap.h49 static inline void mapping_set_error(struct address_space *mapping, int error) in mapping_set_error() argument
55 __filemap_set_wb_err(mapping, error); in mapping_set_error()
58 if (mapping->host) in mapping_set_error()
59 errseq_set(&mapping->host->i_sb->s_wb_err, error); in mapping_set_error()
63 set_bit(AS_ENOSPC, &mapping->flags); in mapping_set_error()
65 set_bit(AS_EIO, &mapping->flags); in mapping_set_error()
68 static inline void mapping_set_unevictable(struct address_space *mapping) in mapping_set_unevictable() argument
70 set_bit(AS_UNEVICTABLE, &mapping->flags); in mapping_set_unevictable()
73 static inline void mapping_clear_unevictable(struct address_space *mapping) in mapping_clear_unevictable() argument
75 clear_bit(AS_UNEVICTABLE, &mapping->flags); in mapping_clear_unevictable()
[all …]
Dio-mapping.h58 io_mapping_fini(struct io_mapping *mapping) in io_mapping_fini() argument
60 iomap_free(mapping->base, mapping->size); in io_mapping_fini()
65 io_mapping_map_atomic_wc(struct io_mapping *mapping, in io_mapping_map_atomic_wc() argument
70 BUG_ON(offset >= mapping->size); in io_mapping_map_atomic_wc()
71 phys_addr = mapping->base + offset; in io_mapping_map_atomic_wc()
72 return iomap_atomic_prot_pfn(PHYS_PFN(phys_addr), mapping->prot); in io_mapping_map_atomic_wc()
82 io_mapping_map_wc(struct io_mapping *mapping, in io_mapping_map_wc() argument
88 BUG_ON(offset >= mapping->size); in io_mapping_map_wc()
89 phys_addr = mapping->base + offset; in io_mapping_map_wc()
128 io_mapping_fini(struct io_mapping *mapping) in io_mapping_fini() argument
[all …]
Dtpm_eventlog.h166 void *mapping = NULL; in __calc_tpm2_event_size() local
186 mapping = TPM_MEMREMAP((unsigned long)marker_start, in __calc_tpm2_event_size()
188 if (!mapping) { in __calc_tpm2_event_size()
193 mapping = marker_start; in __calc_tpm2_event_size()
196 event = (struct tcg_pcr_event2_head *)mapping; in __calc_tpm2_event_size()
233 TPM_MEMUNMAP(mapping, mapping_size); in __calc_tpm2_event_size()
235 mapping = TPM_MEMREMAP((unsigned long)marker, in __calc_tpm2_event_size()
237 if (!mapping) { in __calc_tpm2_event_size()
242 mapping = marker; in __calc_tpm2_event_size()
245 memcpy(&halg, mapping, halg_size); in __calc_tpm2_event_size()
[all …]
Dcleancache.h52 static inline bool cleancache_fs_enabled_mapping(struct address_space *mapping) in cleancache_fs_enabled_mapping() argument
54 return mapping->host->i_sb->cleancache_poolid >= 0; in cleancache_fs_enabled_mapping()
58 return cleancache_fs_enabled_mapping(page->mapping); in cleancache_fs_enabled()
104 static inline void cleancache_invalidate_page(struct address_space *mapping, in cleancache_invalidate_page() argument
108 if (cleancache_enabled && cleancache_fs_enabled_mapping(mapping)) in cleancache_invalidate_page()
109 __cleancache_invalidate_page(mapping, page); in cleancache_invalidate_page()
112 static inline void cleancache_invalidate_inode(struct address_space *mapping) in cleancache_invalidate_inode() argument
114 if (cleancache_enabled && cleancache_fs_enabled_mapping(mapping)) in cleancache_invalidate_inode()
115 __cleancache_invalidate_inode(mapping); in cleancache_invalidate_inode()
/kernel/linux/linux-5.10/drivers/gpu/drm/panfrost/
Dpanfrost_gem.c59 struct panfrost_gem_mapping *iter, *mapping = NULL; in panfrost_gem_mapping_get() local
65 mapping = iter; in panfrost_gem_mapping_get()
71 return mapping; in panfrost_gem_mapping_get()
75 panfrost_gem_teardown_mapping(struct panfrost_gem_mapping *mapping) in panfrost_gem_teardown_mapping() argument
77 if (mapping->active) in panfrost_gem_teardown_mapping()
78 panfrost_mmu_unmap(mapping); in panfrost_gem_teardown_mapping()
80 spin_lock(&mapping->mmu->mm_lock); in panfrost_gem_teardown_mapping()
81 if (drm_mm_node_allocated(&mapping->mmnode)) in panfrost_gem_teardown_mapping()
82 drm_mm_remove_node(&mapping->mmnode); in panfrost_gem_teardown_mapping()
83 spin_unlock(&mapping->mmu->mm_lock); in panfrost_gem_teardown_mapping()
[all …]
/kernel/liteos_a/kernel/base/vm/
Dlos_vm_filemap.c72 STATIC VOID OsPageCacheAdd(LosFilePage *page, struct page_mapping *mapping, VM_OFFSET_T pgoff) in OsPageCacheAdd() argument
76 LOS_DL_LIST_FOR_EACH_ENTRY(fpage, &mapping->page_list, LosFilePage, node) { in OsPageCacheAdd()
83 LOS_ListTailInsert(&mapping->page_list, &page->node); in OsPageCacheAdd()
86 mapping->nrpages++; in OsPageCacheAdd()
89 VOID OsAddToPageacheLru(LosFilePage *page, struct page_mapping *mapping, VM_OFFSET_T pgoff) in OsAddToPageacheLru() argument
91 OsPageCacheAdd(page, mapping, pgoff); in OsAddToPageacheLru()
99 fpage->mapping->nrpages--; in OsPageCacheDel()
176 struct page_mapping *mapping = NULL; in OsVmmFileRemove() local
185 mapping = &vnode->mapping; in OsVmmFileRemove()
196 LOS_SpinLockSave(&mapping->list_lock, &intSave); in OsVmmFileRemove()
[all …]
/kernel/linux/linux-5.10/drivers/media/usb/uvc/
Duvc_ctrl.c365 static s32 uvc_ctrl_get_zoom(struct uvc_control_mapping *mapping, in uvc_ctrl_get_zoom() argument
383 static void uvc_ctrl_set_zoom(struct uvc_control_mapping *mapping, in uvc_ctrl_set_zoom() argument
390 static s32 uvc_ctrl_get_rel_speed(struct uvc_control_mapping *mapping, in uvc_ctrl_get_rel_speed() argument
393 unsigned int first = mapping->offset / 8; in uvc_ctrl_get_rel_speed()
410 static void uvc_ctrl_set_rel_speed(struct uvc_control_mapping *mapping, in uvc_ctrl_set_rel_speed() argument
413 unsigned int first = mapping->offset / 8; in uvc_ctrl_set_rel_speed()
764 static s32 uvc_get_le_value(struct uvc_control_mapping *mapping, in uvc_get_le_value() argument
767 int bits = mapping->size; in uvc_get_le_value()
768 int offset = mapping->offset; in uvc_get_le_value()
789 if (mapping->data_type == UVC_CTRL_DATA_TYPE_SIGNED) in uvc_get_le_value()
[all …]
/kernel/linux/linux-5.10/drivers/gpu/drm/exynos/
Dexynos_drm_dma.c67 ret = arm_iommu_attach_device(subdrv_dev, priv->mapping); in drm_iommu_attach_device()
69 ret = iommu_attach_device(priv->mapping, subdrv_dev); in drm_iommu_attach_device()
93 iommu_detach_device(priv->mapping, subdrv_dev); in drm_iommu_detach_device()
110 if (!priv->mapping) { in exynos_drm_register_dma()
111 void *mapping; in exynos_drm_register_dma() local
114 mapping = arm_iommu_create_mapping(&platform_bus_type, in exynos_drm_register_dma()
117 mapping = iommu_get_domain_for_dev(priv->dma_dev); in exynos_drm_register_dma()
119 mapping = ERR_PTR(-ENODEV); in exynos_drm_register_dma()
121 if (IS_ERR(mapping)) in exynos_drm_register_dma()
122 return PTR_ERR(mapping); in exynos_drm_register_dma()
[all …]
/kernel/linux/linux-5.10/fs/
Ddax.c342 static void dax_associate_entry(void *entry, struct address_space *mapping, in dax_associate_entry() argument
355 WARN_ON_ONCE(page->mapping); in dax_associate_entry()
356 page->mapping = mapping; in dax_associate_entry()
361 static void dax_disassociate_entry(void *entry, struct address_space *mapping, in dax_disassociate_entry() argument
373 WARN_ON_ONCE(page->mapping && page->mapping != mapping); in dax_disassociate_entry()
374 page->mapping = NULL; in dax_disassociate_entry()
408 struct address_space *mapping = READ_ONCE(page->mapping); in dax_lock_page() local
411 if (!mapping || !dax_mapping(mapping)) in dax_lock_page()
422 if (S_ISCHR(mapping->host->i_mode)) in dax_lock_page()
425 xas.xa = &mapping->i_pages; in dax_lock_page()
[all …]
/kernel/linux/linux-5.10/arch/arm/mm/
Dflush.c199 void __flush_dcache_page(struct address_space *mapping, struct page *page) in __flush_dcache_page() argument
232 if (mapping && cache_is_vipt_aliasing()) in __flush_dcache_page()
237 static void __flush_dcache_aliases(struct address_space *mapping, struct page *page) in __flush_dcache_aliases() argument
251 flush_dcache_mmap_lock(mapping); in __flush_dcache_aliases()
252 vma_interval_tree_foreach(mpnt, &mapping->i_mmap, pgoff, pgoff) { in __flush_dcache_aliases()
265 flush_dcache_mmap_unlock(mapping); in __flush_dcache_aliases()
273 struct address_space *mapping; in __sync_icache_dcache() local
284 mapping = page_mapping_file(page); in __sync_icache_dcache()
286 mapping = NULL; in __sync_icache_dcache()
289 __flush_dcache_page(mapping, page); in __sync_icache_dcache()
[all …]
Ddma-mapping.c1099 static int extend_iommu_mapping(struct dma_iommu_mapping *mapping);
1101 static inline dma_addr_t __alloc_iova(struct dma_iommu_mapping *mapping, in __alloc_iova() argument
1107 size_t mapping_size = mapping->bits << PAGE_SHIFT; in __alloc_iova()
1118 spin_lock_irqsave(&mapping->lock, flags); in __alloc_iova()
1119 for (i = 0; i < mapping->nr_bitmaps; i++) { in __alloc_iova()
1120 start = bitmap_find_next_zero_area(mapping->bitmaps[i], in __alloc_iova()
1121 mapping->bits, 0, count, align); in __alloc_iova()
1123 if (start > mapping->bits) in __alloc_iova()
1126 bitmap_set(mapping->bitmaps[i], start, count); in __alloc_iova()
1135 if (i == mapping->nr_bitmaps) { in __alloc_iova()
[all …]
/kernel/linux/linux-5.10/drivers/gpu/drm/etnaviv/
Detnaviv_gem.c229 struct etnaviv_vram_mapping *mapping; in etnaviv_gem_get_vram_mapping() local
231 list_for_each_entry(mapping, &obj->vram_list, obj_node) { in etnaviv_gem_get_vram_mapping()
232 if (mapping->context == context) in etnaviv_gem_get_vram_mapping()
233 return mapping; in etnaviv_gem_get_vram_mapping()
239 void etnaviv_gem_mapping_unreference(struct etnaviv_vram_mapping *mapping) in etnaviv_gem_mapping_unreference() argument
241 struct etnaviv_gem_object *etnaviv_obj = mapping->object; in etnaviv_gem_mapping_unreference()
244 WARN_ON(mapping->use == 0); in etnaviv_gem_mapping_unreference()
245 mapping->use -= 1; in etnaviv_gem_mapping_unreference()
256 struct etnaviv_vram_mapping *mapping; in etnaviv_gem_mapping_get() local
261 mapping = etnaviv_gem_get_vram_mapping(etnaviv_obj, mmu_context); in etnaviv_gem_mapping_get()
[all …]
Detnaviv_mmu.c123 struct etnaviv_vram_mapping *mapping) in etnaviv_iommu_remove_mapping() argument
125 struct etnaviv_gem_object *etnaviv_obj = mapping->object; in etnaviv_iommu_remove_mapping()
127 etnaviv_iommu_unmap(context, mapping->vram_node.start, in etnaviv_iommu_remove_mapping()
129 drm_mm_remove_node(&mapping->vram_node); in etnaviv_iommu_remove_mapping()
226 struct etnaviv_vram_mapping *mapping, u64 va) in etnaviv_iommu_map_gem() argument
243 mapping->iova = iova; in etnaviv_iommu_map_gem()
244 list_add_tail(&mapping->mmu_node, &context->mappings); in etnaviv_iommu_map_gem()
250 node = &mapping->vram_node; in etnaviv_iommu_map_gem()
261 mapping->iova = node->start; in etnaviv_iommu_map_gem()
270 list_add_tail(&mapping->mmu_node, &context->mappings); in etnaviv_iommu_map_gem()
[all …]
/kernel/linux/linux-5.10/arch/nios2/mm/
Dcacheflush.c73 static void flush_aliases(struct address_space *mapping, struct page *page) in flush_aliases() argument
81 flush_dcache_mmap_lock(mapping); in flush_aliases()
82 vma_interval_tree_foreach(mpnt, &mapping->i_mmap, pgoff, pgoff) { in flush_aliases()
94 flush_dcache_mmap_unlock(mapping); in flush_aliases()
160 void __flush_dcache_page(struct address_space *mapping, struct page *page) in __flush_dcache_page() argument
174 struct address_space *mapping; in flush_dcache_page() local
183 mapping = page_mapping_file(page); in flush_dcache_page()
186 if (mapping && !mapping_mapped(mapping)) { in flush_dcache_page()
189 __flush_dcache_page(mapping, page); in flush_dcache_page()
190 if (mapping) { in flush_dcache_page()
[all …]
/kernel/linux/linux-5.10/arch/csky/abiv1/
Dcacheflush.c18 struct address_space *mapping; in flush_dcache_page() local
23 mapping = page_mapping_file(page); in flush_dcache_page()
25 if (mapping && !page_mapcount(page)) in flush_dcache_page()
29 if (mapping) in flush_dcache_page()
60 struct address_space *mapping; in flush_kernel_dcache_page() local
62 mapping = page_mapping_file(page); in flush_kernel_dcache_page()
64 if (!mapping || mapping_mapped(mapping)) in flush_kernel_dcache_page()
/kernel/linux/linux-5.10/drivers/net/wireless/marvell/mwifiex/
Dutil.h69 struct mwifiex_dma_mapping *mapping) in mwifiex_store_mapping() argument
73 memcpy(&cb->dma_mapping, mapping, sizeof(*mapping)); in mwifiex_store_mapping()
77 struct mwifiex_dma_mapping *mapping) in mwifiex_get_mapping() argument
81 memcpy(mapping, &cb->dma_mapping, sizeof(*mapping)); in mwifiex_get_mapping()
86 struct mwifiex_dma_mapping mapping; in MWIFIEX_SKB_DMA_ADDR() local
88 mwifiex_get_mapping(skb, &mapping); in MWIFIEX_SKB_DMA_ADDR()
90 return mapping.addr; in MWIFIEX_SKB_DMA_ADDR()
/kernel/linux/linux-5.10/drivers/sh/clk/
Dcore.c340 struct clk_mapping *mapping = clk->mapping; in clk_establish_mapping() local
345 if (!mapping) { in clk_establish_mapping()
352 clk->mapping = &dummy_mapping; in clk_establish_mapping()
361 mapping = clkp->mapping; in clk_establish_mapping()
362 BUG_ON(!mapping); in clk_establish_mapping()
368 if (!mapping->base && mapping->phys) { in clk_establish_mapping()
369 kref_init(&mapping->ref); in clk_establish_mapping()
371 mapping->base = ioremap(mapping->phys, mapping->len); in clk_establish_mapping()
372 if (unlikely(!mapping->base)) in clk_establish_mapping()
374 } else if (mapping->base) { in clk_establish_mapping()
[all …]
/kernel/linux/linux-5.10/fs/gfs2/
Daops.c91 struct inode *inode = page->mapping->host; in gfs2_writepage()
120 struct inode * const inode = page->mapping->host; in gfs2_write_jdata_page()
153 struct inode *inode = page->mapping->host; in __gfs2_jdata_writepage()
179 struct inode *inode = page->mapping->host; in gfs2_jdata_writepage()
203 static int gfs2_writepages(struct address_space *mapping, in gfs2_writepages() argument
206 struct gfs2_sbd *sdp = gfs2_mapping2sbd(mapping); in gfs2_writepages()
216 ret = iomap_writepages(mapping, wbc, &wpc, &gfs2_writeback_ops); in gfs2_writepages()
233 static int gfs2_write_jdata_pagevec(struct address_space *mapping, in gfs2_write_jdata_pagevec() argument
239 struct inode *inode = mapping->host; in gfs2_write_jdata_pagevec()
256 if (unlikely(page->mapping != mapping)) { in gfs2_write_jdata_pagevec()
[all …]
/kernel/linux/linux-5.10/drivers/net/ethernet/broadcom/bnxt/
Dbnxt_xdp.c25 dma_addr_t mapping, u32 len) in bnxt_xmit_bd() argument
40 txbd->tx_bd_haddr = cpu_to_le64(mapping); in bnxt_xmit_bd()
48 dma_addr_t mapping, u32 len, u16 rx_prod) in __bnxt_xmit_xdp() argument
52 tx_buf = bnxt_xmit_bd(bp, txr, mapping, len); in __bnxt_xmit_xdp()
59 dma_addr_t mapping, u32 len, in __bnxt_xmit_xdp_redirect() argument
64 tx_buf = bnxt_xmit_bd(bp, txr, mapping, len); in __bnxt_xmit_xdp_redirect()
67 dma_unmap_addr_set(tx_buf, mapping, mapping); in __bnxt_xmit_xdp_redirect()
88 dma_unmap_addr(tx_buf, mapping), in bnxt_tx_int_xdp()
119 dma_addr_t mapping; in bnxt_rx_xdp() local
132 mapping = rx_buf->mapping - bp->rx_dma_offset; in bnxt_rx_xdp()
[all …]
/kernel/linux/linux-5.10/fs/afs/
Dwrite.c77 int afs_write_begin(struct file *file, struct address_space *mapping, in afs_write_begin() argument
93 page = grab_cache_page_write_begin(mapping, index, flags); in afs_write_begin()
162 int afs_write_end(struct file *file, struct address_space *mapping, in afs_write_end() argument
238 static void afs_kill_pages(struct address_space *mapping, in afs_kill_pages() argument
241 struct afs_vnode *vnode = AFS_FS_I(mapping->host); in afs_kill_pages()
256 pv.nr = find_get_pages_contig(mapping, first, count, pv.pages); in afs_kill_pages()
267 generic_error_remove_page(mapping, page); in afs_kill_pages()
281 struct address_space *mapping, in afs_redirty_pages() argument
284 struct afs_vnode *vnode = AFS_FS_I(mapping->host); in afs_redirty_pages()
299 pv.nr = find_get_pages_contig(mapping, first, count, pv.pages); in afs_redirty_pages()
[all …]

12345678910>>...49