Home
last modified time | relevance | path

Searched refs:alloc (Results 1 – 25 of 163) sorted by relevance

1234567

/device/soc/rockchip/common/kernel/drivers/gpu/arm/mali400/mali/linux/
Dmali_memory_ump.c22 mali_mem_allocation *alloc; in mali_mem_ump_map() local
34 alloc = mem_backend->mali_allocation; in mali_mem_ump_map()
35 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_ump_map()
37 session = alloc->session; in mali_mem_ump_map()
63 err = mali_mem_mali_map_prepare(alloc); in mali_mem_ump_map()
73 u32 virt = alloc->mali_vma_node.vm_node.start + offset; in mali_mem_ump_map()
83 if (alloc->flags & _MALI_MAP_EXTERNAL_MAP_GUARD_PAGE) { in mali_mem_ump_map()
84 u32 virt = alloc->mali_vma_node.vm_node.start + offset; in mali_mem_ump_map()
98 static void mali_mem_ump_unmap(mali_mem_allocation *alloc) in mali_mem_ump_unmap() argument
101 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_ump_unmap()
[all …]
Dmali_memory_external.c20 mali_mem_allocation *alloc; in mali_mem_unbind_ext_buf() local
23 alloc = mem_backend->mali_allocation; in mali_mem_unbind_ext_buf()
24 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_unbind_ext_buf()
27 session = alloc->session; in mali_mem_unbind_ext_buf()
30 mali_mem_mali_map_free(session, alloc->psize, alloc->mali_vma_node.vm_node.start, in mali_mem_unbind_ext_buf()
31 alloc->flags); in mali_mem_unbind_ext_buf()
35 _mali_osk_errcode_t mali_mem_bind_ext_buf(mali_mem_allocation *alloc, in mali_mem_bind_ext_buf() argument
44 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_bind_ext_buf()
45 size = alloc->psize; in mali_mem_bind_ext_buf()
46 session = (struct mali_session_data *)(uintptr_t)alloc->session; in mali_mem_bind_ext_buf()
[all …]
Dmali_memory_dma_buf.c37 mali_mem_allocation *alloc; in mali_dma_buf_map() local
48 alloc = mem_backend->mali_allocation; in mali_dma_buf_map()
49 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_dma_buf_map()
54 session = alloc->session; in mali_dma_buf_map()
58 virt = alloc->mali_vma_node.vm_node.start; in mali_dma_buf_map()
59 flags = alloc->flags; in mali_dma_buf_map()
79 err = mali_mem_mali_map_prepare(alloc); in mali_dma_buf_map()
123 static void mali_dma_buf_unmap(mali_mem_allocation *alloc, struct mali_dma_buf_attachment *mem) in mali_dma_buf_unmap() argument
125 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_dma_buf_unmap()
129 MALI_DEBUG_ASSERT_POINTER(alloc->session); in mali_dma_buf_unmap()
[all …]
Dmali_memory_cow.c292 mali_mem_allocation *alloc = NULL; in mali_memory_cow_modify_range() local
305 alloc = backend->mali_allocation; in mali_memory_cow_modify_range()
306 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_memory_cow_modify_range()
308 session = alloc->session; in mali_memory_cow_modify_range()
375 MALI_DEBUG_ASSERT(MALI_MEM_COW == alloc->type); in mali_memory_cow_modify_range()
378 if (NULL != alloc->cpu_mapping.vma) { in mali_memory_cow_modify_range()
379 MALI_DEBUG_ASSERT(0 != alloc->backend_handle); in mali_memory_cow_modify_range()
380 MALI_DEBUG_ASSERT(NULL != alloc->cpu_mapping.vma); in mali_memory_cow_modify_range()
381 …MALI_DEBUG_ASSERT(alloc->cpu_mapping.vma->vm_end - alloc->cpu_mapping.vma->vm_start >= range_size); in mali_memory_cow_modify_range()
384 zap_vma_ptes(alloc->cpu_mapping.vma, alloc->cpu_mapping.vma->vm_start + range_start, range_size); in mali_memory_cow_modify_range()
[all …]
Dmali_memory_secure.c93 void mali_mem_secure_mali_unmap(mali_mem_allocation *alloc) in mali_mem_secure_mali_unmap() argument
96 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_secure_mali_unmap()
97 session = alloc->session; in mali_mem_secure_mali_unmap()
101 mali_mem_mali_map_free(session, alloc->psize, alloc->mali_vma_node.vm_node.start, in mali_mem_secure_mali_unmap()
102 alloc->flags); in mali_mem_secure_mali_unmap()
149 mali_mem_allocation *alloc = mem_bkend->mali_allocation; in mali_mem_secure_release() local
158 mali_mem_secure_mali_unmap(alloc); in mali_mem_secure_release()
Dmali_memory_util.c133 u32 mali_allocation_unref(struct mali_mem_allocation **alloc) in mali_allocation_unref() argument
136 mali_mem_allocation *mali_alloc = *alloc; in mali_allocation_unref()
137 *alloc = NULL; in mali_allocation_unref()
144 void mali_allocation_ref(struct mali_mem_allocation *alloc) in mali_allocation_ref() argument
146 _mali_osk_atomic_inc(&alloc->mem_alloc_refcount); in mali_allocation_ref()
/device/soc/rockchip/common/vendor/drivers/gpu/arm/mali400/mali/linux/
Dmali_memory_ump.c23 mali_mem_allocation *alloc; in mali_mem_ump_map() local
35 alloc = mem_backend->mali_allocation; in mali_mem_ump_map()
36 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_ump_map()
38 session = alloc->session; in mali_mem_ump_map()
64 err = mali_mem_mali_map_prepare(alloc); in mali_mem_ump_map()
74 u32 virt = alloc->mali_vma_node.vm_node.start + offset; in mali_mem_ump_map()
84 if (alloc->flags & MALI_MAP_EXTERNAL_MAP_GUARD_PAGE) { in mali_mem_ump_map()
85 u32 virt = alloc->mali_vma_node.vm_node.start + offset; in mali_mem_ump_map()
99 static void mali_mem_ump_unmap(mali_mem_allocation *alloc) in mali_mem_ump_unmap() argument
102 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_ump_unmap()
[all …]
Dmali_memory_external.c21 mali_mem_allocation *alloc; in mali_mem_unbind_ext_buf() local
24 alloc = mem_backend->mali_allocation; in mali_mem_unbind_ext_buf()
25 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_unbind_ext_buf()
28 session = alloc->session; in mali_mem_unbind_ext_buf()
31 mali_mem_mali_map_free(session, alloc->psize, alloc->mali_vma_node.vm_node.start, alloc->flags); in mali_mem_unbind_ext_buf()
35 mali_osk_errcode_t mali_mem_bind_ext_buf(mali_mem_allocation *alloc, mali_mem_backend *mem_backend,… in mali_mem_bind_ext_buf() argument
42 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_bind_ext_buf()
43 size = alloc->psize; in mali_mem_bind_ext_buf()
44 session = (struct mali_session_data *)(uintptr_t)alloc->session; in mali_mem_bind_ext_buf()
64 alloc->flags |= MALI_MEM_FLAG_MALI_GUARD_PAGE; in mali_mem_bind_ext_buf()
[all …]
Dmali_memory_dma_buf.c38 mali_mem_allocation *alloc; in mali_dma_buf_map() local
49 alloc = mem_backend->mali_allocation; in mali_dma_buf_map()
50 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_dma_buf_map()
55 session = alloc->session; in mali_dma_buf_map()
59 virt = alloc->mali_vma_node.vm_node.start; in mali_dma_buf_map()
60 flags = alloc->flags; in mali_dma_buf_map()
80 err = mali_mem_mali_map_prepare(alloc); in mali_dma_buf_map()
125 static void mali_dma_buf_unmap(mali_mem_allocation *alloc, struct mali_dma_buf_attachment *mem) in mali_dma_buf_unmap() argument
127 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_dma_buf_unmap()
131 MALI_DEBUG_ASSERT_POINTER(alloc->session); in mali_dma_buf_unmap()
[all …]
Dmali_memory_cow.c276 mali_mem_allocation *alloc = NULL; in mali_memory_cow_modify_range() local
293 alloc = backend->mali_allocation; in mali_memory_cow_modify_range()
294 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_memory_cow_modify_range()
296 session = alloc->session; in mali_memory_cow_modify_range()
362 MALI_DEBUG_ASSERT(MALI_MEM_COW == alloc->type); in mali_memory_cow_modify_range()
365 if (alloc->cpu_mapping.vma != NULL) { in mali_memory_cow_modify_range()
366 MALI_DEBUG_ASSERT(alloc->backend_handle != 0); in mali_memory_cow_modify_range()
367 MALI_DEBUG_ASSERT(alloc->cpu_mapping.vma != NULL); in mali_memory_cow_modify_range()
368 …MALI_DEBUG_ASSERT(alloc->cpu_mapping.vma->vm_end - alloc->cpu_mapping.vma->vm_start >= range_size); in mali_memory_cow_modify_range()
371 … zap_vma_ptes(alloc->cpu_mapping.vma, alloc->cpu_mapping.vma->vm_start + range_start, range_size); in mali_memory_cow_modify_range()
[all …]
Dmali_memory_secure.c97 void mali_mem_secure_mali_unmap(mali_mem_allocation *alloc) in mali_mem_secure_mali_unmap() argument
100 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_secure_mali_unmap()
101 session = alloc->session; in mali_mem_secure_mali_unmap()
105 mali_mem_mali_map_free(session, alloc->psize, alloc->mali_vma_node.vm_node.start, alloc->flags); in mali_mem_secure_mali_unmap()
151 mali_mem_allocation *alloc = mem_bkend->mali_allocation; in mali_mem_secure_release() local
160 mali_mem_secure_mali_unmap(alloc); in mali_mem_secure_release()
/device/soc/rockchip/common/vendor/drivers/gpu/arm/midgard/
Dmali_kbase_mem.h67 struct kbase_mem_phy_alloc *alloc; member
87 struct kbase_mem_phy_alloc *alloc; /* NULL for special, non-NULL for native */ member
175 static inline void kbase_mem_phy_alloc_gpu_mapped(struct kbase_mem_phy_alloc *alloc) in kbase_mem_phy_alloc_gpu_mapped() argument
177 KBASE_DEBUG_ASSERT(alloc); in kbase_mem_phy_alloc_gpu_mapped()
179 if (alloc->type == KBASE_MEM_TYPE_NATIVE) { in kbase_mem_phy_alloc_gpu_mapped()
180 atomic_inc(&alloc->gpu_mappings); in kbase_mem_phy_alloc_gpu_mapped()
184 static inline void kbase_mem_phy_alloc_gpu_unmapped(struct kbase_mem_phy_alloc *alloc) in kbase_mem_phy_alloc_gpu_unmapped() argument
186 KBASE_DEBUG_ASSERT(alloc); in kbase_mem_phy_alloc_gpu_unmapped()
188 if (alloc->type == KBASE_MEM_TYPE_NATIVE) { in kbase_mem_phy_alloc_gpu_unmapped()
189 if (0 > atomic_dec_return(&alloc->gpu_mappings)) { in kbase_mem_phy_alloc_gpu_unmapped()
[all …]
Dmali_kbase_mem.c878 struct kbase_mem_phy_alloc *alloc; in kbase_gpu_mmap() local
880 alloc = reg->gpu_alloc; in kbase_gpu_mmap()
881 stride = alloc->imported.alias.stride; in kbase_gpu_mmap()
882 KBASE_DEBUG_ASSERT(alloc->imported.alias.aliased); in kbase_gpu_mmap()
883 for (i = 0; i < alloc->imported.alias.nents; i++) { in kbase_gpu_mmap()
884 if (alloc->imported.alias.aliased[i].alloc) { in kbase_gpu_mmap()
886 alloc->imported.alias.aliased[i].alloc->pages + in kbase_gpu_mmap()
887 alloc->imported.alias.aliased[i].offset, in kbase_gpu_mmap()
888 alloc->imported.alias.aliased[i].length, reg->flags); in kbase_gpu_mmap()
893 kbase_mem_phy_alloc_gpu_mapped(alloc->imported.alias.aliased[i].alloc); in kbase_gpu_mmap()
[all …]
/device/soc/rockchip/common/kernel/drivers/gpu/arm/midgard/
Dmali_kbase_mem.h69 struct kbase_mem_phy_alloc *alloc; member
89 struct kbase_mem_phy_alloc *alloc; /* NULL for special, non-NULL for native */ member
177 static inline void kbase_mem_phy_alloc_gpu_mapped(struct kbase_mem_phy_alloc *alloc) in kbase_mem_phy_alloc_gpu_mapped() argument
179 KBASE_DEBUG_ASSERT(alloc); in kbase_mem_phy_alloc_gpu_mapped()
181 if (alloc->type == KBASE_MEM_TYPE_NATIVE) in kbase_mem_phy_alloc_gpu_mapped()
182 atomic_inc(&alloc->gpu_mappings); in kbase_mem_phy_alloc_gpu_mapped()
185 static inline void kbase_mem_phy_alloc_gpu_unmapped(struct kbase_mem_phy_alloc *alloc) in kbase_mem_phy_alloc_gpu_unmapped() argument
187 KBASE_DEBUG_ASSERT(alloc); in kbase_mem_phy_alloc_gpu_unmapped()
189 if (alloc->type == KBASE_MEM_TYPE_NATIVE) in kbase_mem_phy_alloc_gpu_unmapped()
190 if (0 > atomic_dec_return(&alloc->gpu_mappings)) { in kbase_mem_phy_alloc_gpu_unmapped()
[all …]
Dmali_kbase_mem.c906 struct kbase_mem_phy_alloc *alloc; in kbase_gpu_mmap() local
908 alloc = reg->gpu_alloc; in kbase_gpu_mmap()
909 stride = alloc->imported.alias.stride; in kbase_gpu_mmap()
910 KBASE_DEBUG_ASSERT(alloc->imported.alias.aliased); in kbase_gpu_mmap()
911 for (i = 0; i < alloc->imported.alias.nents; i++) { in kbase_gpu_mmap()
912 if (alloc->imported.alias.aliased[i].alloc) { in kbase_gpu_mmap()
915 alloc->imported.alias.aliased[i].alloc->pages + alloc->imported.alias.aliased[i].offset, in kbase_gpu_mmap()
916 alloc->imported.alias.aliased[i].length, in kbase_gpu_mmap()
921 kbase_mem_phy_alloc_gpu_mapped(alloc->imported.alias.aliased[i].alloc); in kbase_gpu_mmap()
926 alloc->imported.alias.aliased[i].length, in kbase_gpu_mmap()
[all …]
/device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/
Dmali_kbase_mem.h68 struct kbase_mem_phy_alloc *alloc; member
87 struct kbase_mem_phy_alloc *alloc; /* NULL for special, non-NULL for native */ member
207 static inline void kbase_mem_phy_alloc_gpu_mapped(struct kbase_mem_phy_alloc *alloc) in kbase_mem_phy_alloc_gpu_mapped() argument
209 KBASE_DEBUG_ASSERT(alloc); in kbase_mem_phy_alloc_gpu_mapped()
211 if (alloc->type == KBASE_MEM_TYPE_NATIVE) in kbase_mem_phy_alloc_gpu_mapped()
212 atomic_inc(&alloc->gpu_mappings); in kbase_mem_phy_alloc_gpu_mapped()
215 static inline void kbase_mem_phy_alloc_gpu_unmapped(struct kbase_mem_phy_alloc *alloc) in kbase_mem_phy_alloc_gpu_unmapped() argument
217 KBASE_DEBUG_ASSERT(alloc); in kbase_mem_phy_alloc_gpu_unmapped()
219 if (alloc->type == KBASE_MEM_TYPE_NATIVE) in kbase_mem_phy_alloc_gpu_unmapped()
220 if (atomic_dec_return(&alloc->gpu_mappings) < 0) { in kbase_mem_phy_alloc_gpu_unmapped()
[all …]
Dmali_kbase_mem.c1512 struct kbase_mem_phy_alloc *alloc; in kbase_gpu_mmap() local
1532 alloc = reg->gpu_alloc; in kbase_gpu_mmap()
1533 group_id = alloc->group_id; in kbase_gpu_mmap()
1536 u64 const stride = alloc->imported.alias.stride; in kbase_gpu_mmap()
1538 KBASE_DEBUG_ASSERT(alloc->imported.alias.aliased); in kbase_gpu_mmap()
1539 for (i = 0; i < alloc->imported.alias.nents; i++) { in kbase_gpu_mmap()
1540 if (alloc->imported.alias.aliased[i].alloc) { in kbase_gpu_mmap()
1544 alloc->imported.alias.aliased[i] in kbase_gpu_mmap()
1545 .alloc->pages + in kbase_gpu_mmap()
1546 alloc->imported.alias.aliased[i] in kbase_gpu_mmap()
[all …]
Dmali_kbase_trace_gpu_mem.c163 struct kbase_mem_phy_alloc *alloc) in kbase_remove_dma_buf_usage() argument
171 kctx, alloc->imported.umm.dma_buf, &kbdev->dma_buf_root); in kbase_remove_dma_buf_usage()
174 kctx, alloc->imported.umm.dma_buf, &kctx->kprcs->dma_buf_root); in kbase_remove_dma_buf_usage()
180 kbdev->total_gpu_pages -= alloc->nents; in kbase_remove_dma_buf_usage()
183 kctx->kprcs->total_gpu_pages -= alloc->nents; in kbase_remove_dma_buf_usage()
193 struct kbase_mem_phy_alloc *alloc) in kbase_add_dma_buf_usage() argument
202 kctx, alloc->imported.umm.dma_buf, &kbdev->dma_buf_root); in kbase_add_dma_buf_usage()
205 kctx, alloc->imported.umm.dma_buf, &kctx->kprcs->dma_buf_root); in kbase_add_dma_buf_usage()
211 kbdev->total_gpu_pages += alloc->nents; in kbase_add_dma_buf_usage()
214 kctx->kprcs->total_gpu_pages += alloc->nents; in kbase_add_dma_buf_usage()
Dmali_kbase_mem_linux.c217 struct kbase_mem_phy_alloc *alloc) in kbase_phy_alloc_mapping_term() argument
219 WARN_ON(!alloc->permanent_map); in kbase_phy_alloc_mapping_term()
220 kbase_vunmap_phy_pages(kctx, alloc->permanent_map); in kbase_phy_alloc_mapping_term()
221 kfree(alloc->permanent_map); in kbase_phy_alloc_mapping_term()
223 alloc->permanent_map = NULL; in kbase_phy_alloc_mapping_term()
229 WARN_ON(alloc->nents > atomic_read(&kctx->permanent_mapped_pages)); in kbase_phy_alloc_mapping_term()
230 atomic_sub(alloc->nents, &kctx->permanent_mapped_pages); in kbase_phy_alloc_mapping_term()
678 struct kbase_mem_phy_alloc *alloc; in kbase_mem_evictable_reclaim_scan_objects() local
686 list_for_each_entry_safe(alloc, tmp, &kctx->evict_list, evict_node) { in kbase_mem_evictable_reclaim_scan_objects()
689 err = kbase_mem_shrink_gpu_mapping(kctx, alloc->reg, in kbase_mem_evictable_reclaim_scan_objects()
[all …]
/device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/
Dmali_kbase_mem.h67 struct kbase_mem_phy_alloc *alloc; member
85 struct kbase_mem_phy_alloc *alloc; /* NULL for special, non-NULL for native */ member
198 static inline void kbase_mem_phy_alloc_gpu_mapped(struct kbase_mem_phy_alloc *alloc) in kbase_mem_phy_alloc_gpu_mapped() argument
200 KBASE_DEBUG_ASSERT(alloc); in kbase_mem_phy_alloc_gpu_mapped()
202 if (alloc->type == KBASE_MEM_TYPE_NATIVE) { in kbase_mem_phy_alloc_gpu_mapped()
203 atomic_inc(&alloc->gpu_mappings); in kbase_mem_phy_alloc_gpu_mapped()
207 static inline void kbase_mem_phy_alloc_gpu_unmapped(struct kbase_mem_phy_alloc *alloc) in kbase_mem_phy_alloc_gpu_unmapped() argument
209 KBASE_DEBUG_ASSERT(alloc); in kbase_mem_phy_alloc_gpu_unmapped()
211 if (alloc->type == KBASE_MEM_TYPE_NATIVE) { in kbase_mem_phy_alloc_gpu_unmapped()
212 if (0 > atomic_dec_return(&alloc->gpu_mappings)) { in kbase_mem_phy_alloc_gpu_unmapped()
[all …]
Dmali_kbase_mem.c1235 struct kbase_mem_phy_alloc *alloc; in kbase_gpu_mmap() local
1257 alloc = reg->gpu_alloc; in kbase_gpu_mmap()
1258 group_id = alloc->group_id; in kbase_gpu_mmap()
1261 u64 const stride = alloc->imported.alias.stride; in kbase_gpu_mmap()
1263 KBASE_DEBUG_ASSERT(alloc->imported.alias.aliased); in kbase_gpu_mmap()
1264 for (i = 0; i < alloc->imported.alias.nents; i++) { in kbase_gpu_mmap()
1265 if (alloc->imported.alias.aliased[i].alloc) { in kbase_gpu_mmap()
1268alloc->imported.alias.aliased[i].alloc->pages + alloc->imported.alias.aliased[i].offset, in kbase_gpu_mmap()
1269alloc->imported.alias.aliased[i].length, reg->flags & gwt_mask, kctx->as_nr, group_id); in kbase_gpu_mmap()
1274 kbase_mem_phy_alloc_gpu_mapped(alloc->imported.alias.aliased[i].alloc); in kbase_gpu_mmap()
[all …]
Dmali_kbase_trace_gpu_mem.c160 void kbase_remove_dma_buf_usage(struct kbase_context *kctx, struct kbase_mem_phy_alloc *alloc) in kbase_remove_dma_buf_usage() argument
167 …dev_mapping_removed = kbase_delete_dma_buf_mapping(kctx, alloc->imported.umm.dma_buf, &kbdev->dma_… in kbase_remove_dma_buf_usage()
169 …prcs_mapping_removed = kbase_delete_dma_buf_mapping(kctx, alloc->imported.umm.dma_buf, &kctx->kprc… in kbase_remove_dma_buf_usage()
175 kbdev->total_gpu_pages -= alloc->nents; in kbase_remove_dma_buf_usage()
179 kctx->kprcs->total_gpu_pages -= alloc->nents; in kbase_remove_dma_buf_usage()
190 void kbase_add_dma_buf_usage(struct kbase_context *kctx, struct kbase_mem_phy_alloc *alloc) in kbase_add_dma_buf_usage() argument
198 …unique_dev_dmabuf = kbase_capture_dma_buf_mapping(kctx, alloc->imported.umm.dma_buf, &kbdev->dma_b… in kbase_add_dma_buf_usage()
200 …unique_prcs_dmabuf = kbase_capture_dma_buf_mapping(kctx, alloc->imported.umm.dma_buf, &kctx->kprcs… in kbase_add_dma_buf_usage()
206 kbdev->total_gpu_pages += alloc->nents; in kbase_add_dma_buf_usage()
210 kctx->kprcs->total_gpu_pages += alloc->nents; in kbase_add_dma_buf_usage()
Dmali_kbase_mem_linux.c229 struct kbase_mem_phy_alloc *alloc) in kbase_phy_alloc_mapping_term() argument
231 WARN_ON(!alloc->permanent_map); in kbase_phy_alloc_mapping_term()
232 kbase_vunmap_phy_pages(kctx, alloc->permanent_map); in kbase_phy_alloc_mapping_term()
233 kfree(alloc->permanent_map); in kbase_phy_alloc_mapping_term()
235 alloc->permanent_map = NULL; in kbase_phy_alloc_mapping_term()
241 WARN_ON(alloc->nents > atomic_read(&kctx->permanent_mapped_pages)); in kbase_phy_alloc_mapping_term()
242 atomic_sub(alloc->nents, &kctx->permanent_mapped_pages); in kbase_phy_alloc_mapping_term()
687 struct kbase_mem_phy_alloc *alloc; in kbase_mem_evictable_reclaim_count_objects() local
694 list_for_each_entry(alloc, &kctx->evict_list, evict_node) pages += in kbase_mem_evictable_reclaim_count_objects()
695 alloc->nents; in kbase_mem_evictable_reclaim_count_objects()
[all …]
/device/soc/esp/esp32/components/xtensa/esp32/include/xtensa/config/
Dtie-asm.h76 .macro xchal_ncp_store ptr at1 at2 at3 at4 continue=0 ofs=-1 select=XTHAL_SAS_ALL alloc=0
84 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_GLOB) & ~(\alloc)) == 0
96 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_CALR) & ~(\alloc)) == 0
116 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_NOCC | XTHAL_SAS_CALR) & ~(\alloc)) == 0
130 .elseif ((XTHAL_SAS_TIE | XTHAL_SAS_NOCC | XTHAL_SAS_CALR) & ~(\alloc)) == 0
156 .macro xchal_ncp_load ptr at1 at2 at3 at4 continue=0 ofs=-1 select=XTHAL_SAS_ALL alloc=0
164 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_GLOB) & ~(\alloc)) == 0
176 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_CALR) & ~(\alloc)) == 0
196 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_NOCC | XTHAL_SAS_CALR) & ~(\alloc)) == 0
210 .elseif ((XTHAL_SAS_TIE | XTHAL_SAS_NOCC | XTHAL_SAS_CALR) & ~(\alloc)) == 0
[all …]
/device/soc/esp/esp32/components/osal/port/xtensa/
Dxtensa_context.S486 …a5, a6, a7 continue=0 ofs=-1 select=XTHAL_SAS_TIE|XTHAL_SAS_NOCC|XTHAL_SAS_CALE alloc=XTHAL_SAS_ALL
494 …a5, a6, a7 continue=0 ofs=-1 select=XTHAL_SAS_TIE|XTHAL_SAS_NOCC|XTHAL_SAS_CALE alloc=XTHAL_SAS_ALL
502 …a5, a6, a7 continue=0 ofs=-1 select=XTHAL_SAS_TIE|XTHAL_SAS_NOCC|XTHAL_SAS_CALE alloc=XTHAL_SAS_ALL
510 …a5, a6, a7 continue=0 ofs=-1 select=XTHAL_SAS_TIE|XTHAL_SAS_NOCC|XTHAL_SAS_CALE alloc=XTHAL_SAS_ALL
518 …a5, a6, a7 continue=0 ofs=-1 select=XTHAL_SAS_TIE|XTHAL_SAS_NOCC|XTHAL_SAS_CALE alloc=XTHAL_SAS_ALL
526 …a5, a6, a7 continue=0 ofs=-1 select=XTHAL_SAS_TIE|XTHAL_SAS_NOCC|XTHAL_SAS_CALE alloc=XTHAL_SAS_ALL
534 …a5, a6, a7 continue=0 ofs=-1 select=XTHAL_SAS_TIE|XTHAL_SAS_NOCC|XTHAL_SAS_CALE alloc=XTHAL_SAS_ALL
542 …a5, a6, a7 continue=0 ofs=-1 select=XTHAL_SAS_TIE|XTHAL_SAS_NOCC|XTHAL_SAS_CALE alloc=XTHAL_SAS_ALL
594 …a5, a6, a7 continue=0 ofs=-1 select=XTHAL_SAS_TIE|XTHAL_SAS_NOCC|XTHAL_SAS_CALE alloc=XTHAL_SAS_ALL
602 …a5, a6, a7 continue=0 ofs=-1 select=XTHAL_SAS_TIE|XTHAL_SAS_NOCC|XTHAL_SAS_CALE alloc=XTHAL_SAS_ALL
[all …]

1234567