| /device/soc/rockchip/common/kernel/drivers/gpu/arm/mali400/mali/linux/ |
| D | mali_memory_ump.c | 22 mali_mem_allocation *alloc; in mali_mem_ump_map() local 34 alloc = mem_backend->mali_allocation; in mali_mem_ump_map() 35 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_ump_map() 37 session = alloc->session; in mali_mem_ump_map() 63 err = mali_mem_mali_map_prepare(alloc); in mali_mem_ump_map() 73 u32 virt = alloc->mali_vma_node.vm_node.start + offset; in mali_mem_ump_map() 83 if (alloc->flags & _MALI_MAP_EXTERNAL_MAP_GUARD_PAGE) { in mali_mem_ump_map() 84 u32 virt = alloc->mali_vma_node.vm_node.start + offset; in mali_mem_ump_map() 98 static void mali_mem_ump_unmap(mali_mem_allocation *alloc) in mali_mem_ump_unmap() argument 101 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_ump_unmap() [all …]
|
| D | mali_memory_external.c | 20 mali_mem_allocation *alloc; in mali_mem_unbind_ext_buf() local 23 alloc = mem_backend->mali_allocation; in mali_mem_unbind_ext_buf() 24 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_unbind_ext_buf() 27 session = alloc->session; in mali_mem_unbind_ext_buf() 30 mali_mem_mali_map_free(session, alloc->psize, alloc->mali_vma_node.vm_node.start, in mali_mem_unbind_ext_buf() 31 alloc->flags); in mali_mem_unbind_ext_buf() 35 _mali_osk_errcode_t mali_mem_bind_ext_buf(mali_mem_allocation *alloc, in mali_mem_bind_ext_buf() argument 44 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_bind_ext_buf() 45 size = alloc->psize; in mali_mem_bind_ext_buf() 46 session = (struct mali_session_data *)(uintptr_t)alloc->session; in mali_mem_bind_ext_buf() [all …]
|
| D | mali_memory_dma_buf.c | 37 mali_mem_allocation *alloc; in mali_dma_buf_map() local 48 alloc = mem_backend->mali_allocation; in mali_dma_buf_map() 49 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_dma_buf_map() 54 session = alloc->session; in mali_dma_buf_map() 58 virt = alloc->mali_vma_node.vm_node.start; in mali_dma_buf_map() 59 flags = alloc->flags; in mali_dma_buf_map() 79 err = mali_mem_mali_map_prepare(alloc); in mali_dma_buf_map() 123 static void mali_dma_buf_unmap(mali_mem_allocation *alloc, struct mali_dma_buf_attachment *mem) in mali_dma_buf_unmap() argument 125 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_dma_buf_unmap() 129 MALI_DEBUG_ASSERT_POINTER(alloc->session); in mali_dma_buf_unmap() [all …]
|
| D | mali_memory_cow.c | 292 mali_mem_allocation *alloc = NULL; in mali_memory_cow_modify_range() local 305 alloc = backend->mali_allocation; in mali_memory_cow_modify_range() 306 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_memory_cow_modify_range() 308 session = alloc->session; in mali_memory_cow_modify_range() 375 MALI_DEBUG_ASSERT(MALI_MEM_COW == alloc->type); in mali_memory_cow_modify_range() 378 if (NULL != alloc->cpu_mapping.vma) { in mali_memory_cow_modify_range() 379 MALI_DEBUG_ASSERT(0 != alloc->backend_handle); in mali_memory_cow_modify_range() 380 MALI_DEBUG_ASSERT(NULL != alloc->cpu_mapping.vma); in mali_memory_cow_modify_range() 381 …MALI_DEBUG_ASSERT(alloc->cpu_mapping.vma->vm_end - alloc->cpu_mapping.vma->vm_start >= range_size); in mali_memory_cow_modify_range() 384 zap_vma_ptes(alloc->cpu_mapping.vma, alloc->cpu_mapping.vma->vm_start + range_start, range_size); in mali_memory_cow_modify_range() [all …]
|
| D | mali_memory_secure.c | 93 void mali_mem_secure_mali_unmap(mali_mem_allocation *alloc) in mali_mem_secure_mali_unmap() argument 96 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_secure_mali_unmap() 97 session = alloc->session; in mali_mem_secure_mali_unmap() 101 mali_mem_mali_map_free(session, alloc->psize, alloc->mali_vma_node.vm_node.start, in mali_mem_secure_mali_unmap() 102 alloc->flags); in mali_mem_secure_mali_unmap() 149 mali_mem_allocation *alloc = mem_bkend->mali_allocation; in mali_mem_secure_release() local 158 mali_mem_secure_mali_unmap(alloc); in mali_mem_secure_release()
|
| D | mali_memory_util.c | 133 u32 mali_allocation_unref(struct mali_mem_allocation **alloc) in mali_allocation_unref() argument 136 mali_mem_allocation *mali_alloc = *alloc; in mali_allocation_unref() 137 *alloc = NULL; in mali_allocation_unref() 144 void mali_allocation_ref(struct mali_mem_allocation *alloc) in mali_allocation_ref() argument 146 _mali_osk_atomic_inc(&alloc->mem_alloc_refcount); in mali_allocation_ref()
|
| /device/soc/rockchip/common/vendor/drivers/gpu/arm/mali400/mali/linux/ |
| D | mali_memory_ump.c | 23 mali_mem_allocation *alloc; in mali_mem_ump_map() local 35 alloc = mem_backend->mali_allocation; in mali_mem_ump_map() 36 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_ump_map() 38 session = alloc->session; in mali_mem_ump_map() 64 err = mali_mem_mali_map_prepare(alloc); in mali_mem_ump_map() 74 u32 virt = alloc->mali_vma_node.vm_node.start + offset; in mali_mem_ump_map() 84 if (alloc->flags & MALI_MAP_EXTERNAL_MAP_GUARD_PAGE) { in mali_mem_ump_map() 85 u32 virt = alloc->mali_vma_node.vm_node.start + offset; in mali_mem_ump_map() 99 static void mali_mem_ump_unmap(mali_mem_allocation *alloc) in mali_mem_ump_unmap() argument 102 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_ump_unmap() [all …]
|
| D | mali_memory_external.c | 21 mali_mem_allocation *alloc; in mali_mem_unbind_ext_buf() local 24 alloc = mem_backend->mali_allocation; in mali_mem_unbind_ext_buf() 25 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_unbind_ext_buf() 28 session = alloc->session; in mali_mem_unbind_ext_buf() 31 mali_mem_mali_map_free(session, alloc->psize, alloc->mali_vma_node.vm_node.start, alloc->flags); in mali_mem_unbind_ext_buf() 35 mali_osk_errcode_t mali_mem_bind_ext_buf(mali_mem_allocation *alloc, mali_mem_backend *mem_backend,… in mali_mem_bind_ext_buf() argument 42 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_bind_ext_buf() 43 size = alloc->psize; in mali_mem_bind_ext_buf() 44 session = (struct mali_session_data *)(uintptr_t)alloc->session; in mali_mem_bind_ext_buf() 64 alloc->flags |= MALI_MEM_FLAG_MALI_GUARD_PAGE; in mali_mem_bind_ext_buf() [all …]
|
| D | mali_memory_dma_buf.c | 38 mali_mem_allocation *alloc; in mali_dma_buf_map() local 49 alloc = mem_backend->mali_allocation; in mali_dma_buf_map() 50 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_dma_buf_map() 55 session = alloc->session; in mali_dma_buf_map() 59 virt = alloc->mali_vma_node.vm_node.start; in mali_dma_buf_map() 60 flags = alloc->flags; in mali_dma_buf_map() 80 err = mali_mem_mali_map_prepare(alloc); in mali_dma_buf_map() 125 static void mali_dma_buf_unmap(mali_mem_allocation *alloc, struct mali_dma_buf_attachment *mem) in mali_dma_buf_unmap() argument 127 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_dma_buf_unmap() 131 MALI_DEBUG_ASSERT_POINTER(alloc->session); in mali_dma_buf_unmap() [all …]
|
| D | mali_memory_cow.c | 276 mali_mem_allocation *alloc = NULL; in mali_memory_cow_modify_range() local 293 alloc = backend->mali_allocation; in mali_memory_cow_modify_range() 294 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_memory_cow_modify_range() 296 session = alloc->session; in mali_memory_cow_modify_range() 362 MALI_DEBUG_ASSERT(MALI_MEM_COW == alloc->type); in mali_memory_cow_modify_range() 365 if (alloc->cpu_mapping.vma != NULL) { in mali_memory_cow_modify_range() 366 MALI_DEBUG_ASSERT(alloc->backend_handle != 0); in mali_memory_cow_modify_range() 367 MALI_DEBUG_ASSERT(alloc->cpu_mapping.vma != NULL); in mali_memory_cow_modify_range() 368 …MALI_DEBUG_ASSERT(alloc->cpu_mapping.vma->vm_end - alloc->cpu_mapping.vma->vm_start >= range_size); in mali_memory_cow_modify_range() 371 … zap_vma_ptes(alloc->cpu_mapping.vma, alloc->cpu_mapping.vma->vm_start + range_start, range_size); in mali_memory_cow_modify_range() [all …]
|
| D | mali_memory_secure.c | 97 void mali_mem_secure_mali_unmap(mali_mem_allocation *alloc) in mali_mem_secure_mali_unmap() argument 100 MALI_DEBUG_ASSERT_POINTER(alloc); in mali_mem_secure_mali_unmap() 101 session = alloc->session; in mali_mem_secure_mali_unmap() 105 mali_mem_mali_map_free(session, alloc->psize, alloc->mali_vma_node.vm_node.start, alloc->flags); in mali_mem_secure_mali_unmap() 151 mali_mem_allocation *alloc = mem_bkend->mali_allocation; in mali_mem_secure_release() local 160 mali_mem_secure_mali_unmap(alloc); in mali_mem_secure_release()
|
| /device/soc/rockchip/common/vendor/drivers/gpu/arm/midgard/ |
| D | mali_kbase_mem.h | 67 struct kbase_mem_phy_alloc *alloc; member 87 struct kbase_mem_phy_alloc *alloc; /* NULL for special, non-NULL for native */ member 175 static inline void kbase_mem_phy_alloc_gpu_mapped(struct kbase_mem_phy_alloc *alloc) in kbase_mem_phy_alloc_gpu_mapped() argument 177 KBASE_DEBUG_ASSERT(alloc); in kbase_mem_phy_alloc_gpu_mapped() 179 if (alloc->type == KBASE_MEM_TYPE_NATIVE) { in kbase_mem_phy_alloc_gpu_mapped() 180 atomic_inc(&alloc->gpu_mappings); in kbase_mem_phy_alloc_gpu_mapped() 184 static inline void kbase_mem_phy_alloc_gpu_unmapped(struct kbase_mem_phy_alloc *alloc) in kbase_mem_phy_alloc_gpu_unmapped() argument 186 KBASE_DEBUG_ASSERT(alloc); in kbase_mem_phy_alloc_gpu_unmapped() 188 if (alloc->type == KBASE_MEM_TYPE_NATIVE) { in kbase_mem_phy_alloc_gpu_unmapped() 189 if (0 > atomic_dec_return(&alloc->gpu_mappings)) { in kbase_mem_phy_alloc_gpu_unmapped() [all …]
|
| D | mali_kbase_mem.c | 878 struct kbase_mem_phy_alloc *alloc; in kbase_gpu_mmap() local 880 alloc = reg->gpu_alloc; in kbase_gpu_mmap() 881 stride = alloc->imported.alias.stride; in kbase_gpu_mmap() 882 KBASE_DEBUG_ASSERT(alloc->imported.alias.aliased); in kbase_gpu_mmap() 883 for (i = 0; i < alloc->imported.alias.nents; i++) { in kbase_gpu_mmap() 884 if (alloc->imported.alias.aliased[i].alloc) { in kbase_gpu_mmap() 886 alloc->imported.alias.aliased[i].alloc->pages + in kbase_gpu_mmap() 887 alloc->imported.alias.aliased[i].offset, in kbase_gpu_mmap() 888 alloc->imported.alias.aliased[i].length, reg->flags); in kbase_gpu_mmap() 893 kbase_mem_phy_alloc_gpu_mapped(alloc->imported.alias.aliased[i].alloc); in kbase_gpu_mmap() [all …]
|
| /device/soc/rockchip/common/kernel/drivers/gpu/arm/midgard/ |
| D | mali_kbase_mem.h | 69 struct kbase_mem_phy_alloc *alloc; member 89 struct kbase_mem_phy_alloc *alloc; /* NULL for special, non-NULL for native */ member 177 static inline void kbase_mem_phy_alloc_gpu_mapped(struct kbase_mem_phy_alloc *alloc) in kbase_mem_phy_alloc_gpu_mapped() argument 179 KBASE_DEBUG_ASSERT(alloc); in kbase_mem_phy_alloc_gpu_mapped() 181 if (alloc->type == KBASE_MEM_TYPE_NATIVE) in kbase_mem_phy_alloc_gpu_mapped() 182 atomic_inc(&alloc->gpu_mappings); in kbase_mem_phy_alloc_gpu_mapped() 185 static inline void kbase_mem_phy_alloc_gpu_unmapped(struct kbase_mem_phy_alloc *alloc) in kbase_mem_phy_alloc_gpu_unmapped() argument 187 KBASE_DEBUG_ASSERT(alloc); in kbase_mem_phy_alloc_gpu_unmapped() 189 if (alloc->type == KBASE_MEM_TYPE_NATIVE) in kbase_mem_phy_alloc_gpu_unmapped() 190 if (0 > atomic_dec_return(&alloc->gpu_mappings)) { in kbase_mem_phy_alloc_gpu_unmapped() [all …]
|
| D | mali_kbase_mem.c | 906 struct kbase_mem_phy_alloc *alloc; in kbase_gpu_mmap() local 908 alloc = reg->gpu_alloc; in kbase_gpu_mmap() 909 stride = alloc->imported.alias.stride; in kbase_gpu_mmap() 910 KBASE_DEBUG_ASSERT(alloc->imported.alias.aliased); in kbase_gpu_mmap() 911 for (i = 0; i < alloc->imported.alias.nents; i++) { in kbase_gpu_mmap() 912 if (alloc->imported.alias.aliased[i].alloc) { in kbase_gpu_mmap() 915 alloc->imported.alias.aliased[i].alloc->pages + alloc->imported.alias.aliased[i].offset, in kbase_gpu_mmap() 916 alloc->imported.alias.aliased[i].length, in kbase_gpu_mmap() 921 kbase_mem_phy_alloc_gpu_mapped(alloc->imported.alias.aliased[i].alloc); in kbase_gpu_mmap() 926 alloc->imported.alias.aliased[i].length, in kbase_gpu_mmap() [all …]
|
| /device/soc/rockchip/common/kernel/drivers/gpu/arm/bifrost/ |
| D | mali_kbase_mem.h | 68 struct kbase_mem_phy_alloc *alloc; member 87 struct kbase_mem_phy_alloc *alloc; /* NULL for special, non-NULL for native */ member 207 static inline void kbase_mem_phy_alloc_gpu_mapped(struct kbase_mem_phy_alloc *alloc) in kbase_mem_phy_alloc_gpu_mapped() argument 209 KBASE_DEBUG_ASSERT(alloc); in kbase_mem_phy_alloc_gpu_mapped() 211 if (alloc->type == KBASE_MEM_TYPE_NATIVE) in kbase_mem_phy_alloc_gpu_mapped() 212 atomic_inc(&alloc->gpu_mappings); in kbase_mem_phy_alloc_gpu_mapped() 215 static inline void kbase_mem_phy_alloc_gpu_unmapped(struct kbase_mem_phy_alloc *alloc) in kbase_mem_phy_alloc_gpu_unmapped() argument 217 KBASE_DEBUG_ASSERT(alloc); in kbase_mem_phy_alloc_gpu_unmapped() 219 if (alloc->type == KBASE_MEM_TYPE_NATIVE) in kbase_mem_phy_alloc_gpu_unmapped() 220 if (atomic_dec_return(&alloc->gpu_mappings) < 0) { in kbase_mem_phy_alloc_gpu_unmapped() [all …]
|
| D | mali_kbase_mem.c | 1512 struct kbase_mem_phy_alloc *alloc; in kbase_gpu_mmap() local 1532 alloc = reg->gpu_alloc; in kbase_gpu_mmap() 1533 group_id = alloc->group_id; in kbase_gpu_mmap() 1536 u64 const stride = alloc->imported.alias.stride; in kbase_gpu_mmap() 1538 KBASE_DEBUG_ASSERT(alloc->imported.alias.aliased); in kbase_gpu_mmap() 1539 for (i = 0; i < alloc->imported.alias.nents; i++) { in kbase_gpu_mmap() 1540 if (alloc->imported.alias.aliased[i].alloc) { in kbase_gpu_mmap() 1544 alloc->imported.alias.aliased[i] in kbase_gpu_mmap() 1545 .alloc->pages + in kbase_gpu_mmap() 1546 alloc->imported.alias.aliased[i] in kbase_gpu_mmap() [all …]
|
| D | mali_kbase_trace_gpu_mem.c | 163 struct kbase_mem_phy_alloc *alloc) in kbase_remove_dma_buf_usage() argument 171 kctx, alloc->imported.umm.dma_buf, &kbdev->dma_buf_root); in kbase_remove_dma_buf_usage() 174 kctx, alloc->imported.umm.dma_buf, &kctx->kprcs->dma_buf_root); in kbase_remove_dma_buf_usage() 180 kbdev->total_gpu_pages -= alloc->nents; in kbase_remove_dma_buf_usage() 183 kctx->kprcs->total_gpu_pages -= alloc->nents; in kbase_remove_dma_buf_usage() 193 struct kbase_mem_phy_alloc *alloc) in kbase_add_dma_buf_usage() argument 202 kctx, alloc->imported.umm.dma_buf, &kbdev->dma_buf_root); in kbase_add_dma_buf_usage() 205 kctx, alloc->imported.umm.dma_buf, &kctx->kprcs->dma_buf_root); in kbase_add_dma_buf_usage() 211 kbdev->total_gpu_pages += alloc->nents; in kbase_add_dma_buf_usage() 214 kctx->kprcs->total_gpu_pages += alloc->nents; in kbase_add_dma_buf_usage()
|
| D | mali_kbase_mem_linux.c | 217 struct kbase_mem_phy_alloc *alloc) in kbase_phy_alloc_mapping_term() argument 219 WARN_ON(!alloc->permanent_map); in kbase_phy_alloc_mapping_term() 220 kbase_vunmap_phy_pages(kctx, alloc->permanent_map); in kbase_phy_alloc_mapping_term() 221 kfree(alloc->permanent_map); in kbase_phy_alloc_mapping_term() 223 alloc->permanent_map = NULL; in kbase_phy_alloc_mapping_term() 229 WARN_ON(alloc->nents > atomic_read(&kctx->permanent_mapped_pages)); in kbase_phy_alloc_mapping_term() 230 atomic_sub(alloc->nents, &kctx->permanent_mapped_pages); in kbase_phy_alloc_mapping_term() 678 struct kbase_mem_phy_alloc *alloc; in kbase_mem_evictable_reclaim_scan_objects() local 686 list_for_each_entry_safe(alloc, tmp, &kctx->evict_list, evict_node) { in kbase_mem_evictable_reclaim_scan_objects() 689 err = kbase_mem_shrink_gpu_mapping(kctx, alloc->reg, in kbase_mem_evictable_reclaim_scan_objects() [all …]
|
| /device/soc/rockchip/common/vendor/drivers/gpu/arm/bifrost/ |
| D | mali_kbase_mem.h | 67 struct kbase_mem_phy_alloc *alloc; member 85 struct kbase_mem_phy_alloc *alloc; /* NULL for special, non-NULL for native */ member 198 static inline void kbase_mem_phy_alloc_gpu_mapped(struct kbase_mem_phy_alloc *alloc) in kbase_mem_phy_alloc_gpu_mapped() argument 200 KBASE_DEBUG_ASSERT(alloc); in kbase_mem_phy_alloc_gpu_mapped() 202 if (alloc->type == KBASE_MEM_TYPE_NATIVE) { in kbase_mem_phy_alloc_gpu_mapped() 203 atomic_inc(&alloc->gpu_mappings); in kbase_mem_phy_alloc_gpu_mapped() 207 static inline void kbase_mem_phy_alloc_gpu_unmapped(struct kbase_mem_phy_alloc *alloc) in kbase_mem_phy_alloc_gpu_unmapped() argument 209 KBASE_DEBUG_ASSERT(alloc); in kbase_mem_phy_alloc_gpu_unmapped() 211 if (alloc->type == KBASE_MEM_TYPE_NATIVE) { in kbase_mem_phy_alloc_gpu_unmapped() 212 if (0 > atomic_dec_return(&alloc->gpu_mappings)) { in kbase_mem_phy_alloc_gpu_unmapped() [all …]
|
| D | mali_kbase_mem.c | 1235 struct kbase_mem_phy_alloc *alloc; in kbase_gpu_mmap() local 1257 alloc = reg->gpu_alloc; in kbase_gpu_mmap() 1258 group_id = alloc->group_id; in kbase_gpu_mmap() 1261 u64 const stride = alloc->imported.alias.stride; in kbase_gpu_mmap() 1263 KBASE_DEBUG_ASSERT(alloc->imported.alias.aliased); in kbase_gpu_mmap() 1264 for (i = 0; i < alloc->imported.alias.nents; i++) { in kbase_gpu_mmap() 1265 if (alloc->imported.alias.aliased[i].alloc) { in kbase_gpu_mmap() 1268 … alloc->imported.alias.aliased[i].alloc->pages + alloc->imported.alias.aliased[i].offset, in kbase_gpu_mmap() 1269 … alloc->imported.alias.aliased[i].length, reg->flags & gwt_mask, kctx->as_nr, group_id); in kbase_gpu_mmap() 1274 kbase_mem_phy_alloc_gpu_mapped(alloc->imported.alias.aliased[i].alloc); in kbase_gpu_mmap() [all …]
|
| D | mali_kbase_trace_gpu_mem.c | 160 void kbase_remove_dma_buf_usage(struct kbase_context *kctx, struct kbase_mem_phy_alloc *alloc) in kbase_remove_dma_buf_usage() argument 167 …dev_mapping_removed = kbase_delete_dma_buf_mapping(kctx, alloc->imported.umm.dma_buf, &kbdev->dma_… in kbase_remove_dma_buf_usage() 169 …prcs_mapping_removed = kbase_delete_dma_buf_mapping(kctx, alloc->imported.umm.dma_buf, &kctx->kprc… in kbase_remove_dma_buf_usage() 175 kbdev->total_gpu_pages -= alloc->nents; in kbase_remove_dma_buf_usage() 179 kctx->kprcs->total_gpu_pages -= alloc->nents; in kbase_remove_dma_buf_usage() 190 void kbase_add_dma_buf_usage(struct kbase_context *kctx, struct kbase_mem_phy_alloc *alloc) in kbase_add_dma_buf_usage() argument 198 …unique_dev_dmabuf = kbase_capture_dma_buf_mapping(kctx, alloc->imported.umm.dma_buf, &kbdev->dma_b… in kbase_add_dma_buf_usage() 200 …unique_prcs_dmabuf = kbase_capture_dma_buf_mapping(kctx, alloc->imported.umm.dma_buf, &kctx->kprcs… in kbase_add_dma_buf_usage() 206 kbdev->total_gpu_pages += alloc->nents; in kbase_add_dma_buf_usage() 210 kctx->kprcs->total_gpu_pages += alloc->nents; in kbase_add_dma_buf_usage()
|
| D | mali_kbase_mem_linux.c | 229 struct kbase_mem_phy_alloc *alloc) in kbase_phy_alloc_mapping_term() argument 231 WARN_ON(!alloc->permanent_map); in kbase_phy_alloc_mapping_term() 232 kbase_vunmap_phy_pages(kctx, alloc->permanent_map); in kbase_phy_alloc_mapping_term() 233 kfree(alloc->permanent_map); in kbase_phy_alloc_mapping_term() 235 alloc->permanent_map = NULL; in kbase_phy_alloc_mapping_term() 241 WARN_ON(alloc->nents > atomic_read(&kctx->permanent_mapped_pages)); in kbase_phy_alloc_mapping_term() 242 atomic_sub(alloc->nents, &kctx->permanent_mapped_pages); in kbase_phy_alloc_mapping_term() 687 struct kbase_mem_phy_alloc *alloc; in kbase_mem_evictable_reclaim_count_objects() local 694 list_for_each_entry(alloc, &kctx->evict_list, evict_node) pages += in kbase_mem_evictable_reclaim_count_objects() 695 alloc->nents; in kbase_mem_evictable_reclaim_count_objects() [all …]
|
| /device/soc/esp/esp32/components/xtensa/esp32/include/xtensa/config/ |
| D | tie-asm.h | 76 .macro xchal_ncp_store ptr at1 at2 at3 at4 continue=0 ofs=-1 select=XTHAL_SAS_ALL alloc=0 84 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_GLOB) & ~(\alloc)) == 0 96 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_CALR) & ~(\alloc)) == 0 116 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_NOCC | XTHAL_SAS_CALR) & ~(\alloc)) == 0 130 .elseif ((XTHAL_SAS_TIE | XTHAL_SAS_NOCC | XTHAL_SAS_CALR) & ~(\alloc)) == 0 156 .macro xchal_ncp_load ptr at1 at2 at3 at4 continue=0 ofs=-1 select=XTHAL_SAS_ALL alloc=0 164 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_GLOB) & ~(\alloc)) == 0 176 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_CALR) & ~(\alloc)) == 0 196 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_NOCC | XTHAL_SAS_CALR) & ~(\alloc)) == 0 210 .elseif ((XTHAL_SAS_TIE | XTHAL_SAS_NOCC | XTHAL_SAS_CALR) & ~(\alloc)) == 0 [all …]
|
| /device/soc/esp/esp32/components/osal/port/xtensa/ |
| D | xtensa_context.S | 486 …a5, a6, a7 continue=0 ofs=-1 select=XTHAL_SAS_TIE|XTHAL_SAS_NOCC|XTHAL_SAS_CALE alloc=XTHAL_SAS_ALL 494 …a5, a6, a7 continue=0 ofs=-1 select=XTHAL_SAS_TIE|XTHAL_SAS_NOCC|XTHAL_SAS_CALE alloc=XTHAL_SAS_ALL 502 …a5, a6, a7 continue=0 ofs=-1 select=XTHAL_SAS_TIE|XTHAL_SAS_NOCC|XTHAL_SAS_CALE alloc=XTHAL_SAS_ALL 510 …a5, a6, a7 continue=0 ofs=-1 select=XTHAL_SAS_TIE|XTHAL_SAS_NOCC|XTHAL_SAS_CALE alloc=XTHAL_SAS_ALL 518 …a5, a6, a7 continue=0 ofs=-1 select=XTHAL_SAS_TIE|XTHAL_SAS_NOCC|XTHAL_SAS_CALE alloc=XTHAL_SAS_ALL 526 …a5, a6, a7 continue=0 ofs=-1 select=XTHAL_SAS_TIE|XTHAL_SAS_NOCC|XTHAL_SAS_CALE alloc=XTHAL_SAS_ALL 534 …a5, a6, a7 continue=0 ofs=-1 select=XTHAL_SAS_TIE|XTHAL_SAS_NOCC|XTHAL_SAS_CALE alloc=XTHAL_SAS_ALL 542 …a5, a6, a7 continue=0 ofs=-1 select=XTHAL_SAS_TIE|XTHAL_SAS_NOCC|XTHAL_SAS_CALE alloc=XTHAL_SAS_ALL 594 …a5, a6, a7 continue=0 ofs=-1 select=XTHAL_SAS_TIE|XTHAL_SAS_NOCC|XTHAL_SAS_CALE alloc=XTHAL_SAS_ALL 602 …a5, a6, a7 continue=0 ofs=-1 select=XTHAL_SAS_TIE|XTHAL_SAS_NOCC|XTHAL_SAS_CALE alloc=XTHAL_SAS_ALL [all …]
|