| /kernel/linux/linux-5.10/tools/testing/selftests/kvm/lib/aarch64/ |
| D | processor.c | 19 static uint64_t page_align(struct kvm_vm *vm, uint64_t v) in page_align() argument 21 return (v + vm->page_size) & ~(vm->page_size - 1); in page_align() 24 static uint64_t pgd_index(struct kvm_vm *vm, vm_vaddr_t gva) in pgd_index() argument 26 unsigned int shift = (vm->pgtable_levels - 1) * (vm->page_shift - 3) + vm->page_shift; in pgd_index() 27 uint64_t mask = (1UL << (vm->va_bits - shift)) - 1; in pgd_index() 32 static uint64_t pud_index(struct kvm_vm *vm, vm_vaddr_t gva) in pud_index() argument 34 unsigned int shift = 2 * (vm->page_shift - 3) + vm->page_shift; in pud_index() 35 uint64_t mask = (1UL << (vm->page_shift - 3)) - 1; in pud_index() 37 TEST_ASSERT(vm->pgtable_levels == 4, in pud_index() 38 "Mode %d does not have 4 page table levels", vm->mode); in pud_index() [all …]
|
| /kernel/linux/linux-6.6/drivers/virtio/ |
| D | virtio_mem.c | 278 static void virtio_mem_retry(struct virtio_mem *vm); 279 static int virtio_mem_create_resource(struct virtio_mem *vm); 280 static void virtio_mem_delete_resource(struct virtio_mem *vm); 286 static int register_virtio_mem_device(struct virtio_mem *vm) in register_virtio_mem_device() argument 295 list_add_rcu(&vm->next, &virtio_mem_devices); in register_virtio_mem_device() 305 static void unregister_virtio_mem_device(struct virtio_mem *vm) in unregister_virtio_mem_device() argument 309 list_del_rcu(&vm->next); in unregister_virtio_mem_device() 336 static unsigned long virtio_mem_phys_to_bb_id(struct virtio_mem *vm, in virtio_mem_phys_to_bb_id() argument 339 return addr / vm->bbm.bb_size; in virtio_mem_phys_to_bb_id() 345 static uint64_t virtio_mem_bb_id_to_phys(struct virtio_mem *vm, in virtio_mem_bb_id_to_phys() argument [all …]
|
| /kernel/linux/linux-5.10/drivers/virtio/ |
| D | virtio_mem.c | 168 static int register_virtio_mem_device(struct virtio_mem *vm) in register_virtio_mem_device() argument 177 list_add_rcu(&vm->next, &virtio_mem_devices); in register_virtio_mem_device() 187 static void unregister_virtio_mem_device(struct virtio_mem *vm) in unregister_virtio_mem_device() argument 191 list_del_rcu(&vm->next); in unregister_virtio_mem_device() 218 static unsigned long virtio_mem_phys_to_sb_id(struct virtio_mem *vm, in virtio_mem_phys_to_sb_id() argument 224 return (addr - mb_addr) / vm->subblock_size; in virtio_mem_phys_to_sb_id() 230 static void virtio_mem_mb_set_state(struct virtio_mem *vm, unsigned long mb_id, in virtio_mem_mb_set_state() argument 233 const unsigned long idx = mb_id - vm->first_mb_id; in virtio_mem_mb_set_state() 236 old_state = vm->mb_state[idx]; in virtio_mem_mb_set_state() 237 vm->mb_state[idx] = state; in virtio_mem_mb_set_state() [all …]
|
| /kernel/linux/linux-6.6/tools/testing/selftests/kvm/lib/aarch64/ |
| D | processor.c | 20 static uint64_t page_align(struct kvm_vm *vm, uint64_t v) in page_align() argument 22 return (v + vm->page_size) & ~(vm->page_size - 1); in page_align() 25 static uint64_t pgd_index(struct kvm_vm *vm, vm_vaddr_t gva) in pgd_index() argument 27 unsigned int shift = (vm->pgtable_levels - 1) * (vm->page_shift - 3) + vm->page_shift; in pgd_index() 28 uint64_t mask = (1UL << (vm->va_bits - shift)) - 1; in pgd_index() 33 static uint64_t pud_index(struct kvm_vm *vm, vm_vaddr_t gva) in pud_index() argument 35 unsigned int shift = 2 * (vm->page_shift - 3) + vm->page_shift; in pud_index() 36 uint64_t mask = (1UL << (vm->page_shift - 3)) - 1; in pud_index() 38 TEST_ASSERT(vm->pgtable_levels == 4, in pud_index() 39 "Mode %d does not have 4 page table levels", vm->mode); in pud_index() [all …]
|
| /kernel/linux/linux-5.10/tools/testing/selftests/kvm/include/ |
| D | kvm_util.h | 22 * structure kvm_util is using to maintain the state of a VM. 65 int vm_enable_cap(struct kvm_vm *vm, struct kvm_enable_cap *cap); 66 int vcpu_enable_cap(struct kvm_vm *vm, uint32_t vcpu_id, 68 void vm_enable_dirty_ring(struct kvm_vm *vm, uint32_t ring_size); 74 void kvm_vm_get_dirty_log(struct kvm_vm *vm, int slot, void *log); 75 void kvm_vm_clear_dirty_log(struct kvm_vm *vm, int slot, void *log, 78 int kvm_memcmp_hva_gva(void *hva, struct kvm_vm *vm, const vm_vaddr_t gva, 81 void kvm_vm_elf_load(struct kvm_vm *vm, const char *filename, 84 void vm_dump(FILE *stream, struct kvm_vm *vm, uint8_t indent); 87 * VM VCPU Dump [all …]
|
| /kernel/linux/linux-5.10/tools/testing/selftests/kvm/lib/ |
| D | kvm_util.c | 66 /* VM Enable Capability 69 * vm - Virtual Machine 76 * Enables a capability (KVM_CAP_*) on the VM. 78 int vm_enable_cap(struct kvm_vm *vm, struct kvm_enable_cap *cap) in vm_enable_cap() argument 82 ret = ioctl(vm->fd, KVM_ENABLE_CAP, cap); in vm_enable_cap() 92 * vm - Virtual Machine 102 int vcpu_enable_cap(struct kvm_vm *vm, uint32_t vcpu_id, in vcpu_enable_cap() argument 105 struct vcpu *vcpu = vcpu_find(vm, vcpu_id); in vcpu_enable_cap() 117 static void vm_open(struct kvm_vm *vm, int perm) in vm_open() argument 119 vm->kvm_fd = open(KVM_DEV_PATH, perm); in vm_open() [all …]
|
| /kernel/linux/linux-5.10/tools/testing/selftests/kvm/lib/s390x/ |
| D | processor.c | 18 void virt_pgd_alloc(struct kvm_vm *vm, uint32_t memslot) in virt_pgd_alloc() argument 22 TEST_ASSERT(vm->page_size == 4096, "Unsupported page size: 0x%x", in virt_pgd_alloc() 23 vm->page_size); in virt_pgd_alloc() 25 if (vm->pgd_created) in virt_pgd_alloc() 28 paddr = vm_phy_pages_alloc(vm, PAGES_PER_REGION, in virt_pgd_alloc() 30 memset(addr_gpa2hva(vm, paddr), 0xff, PAGES_PER_REGION * vm->page_size); in virt_pgd_alloc() 32 vm->pgd = paddr; in virt_pgd_alloc() 33 vm->pgd_created = true; in virt_pgd_alloc() 41 static uint64_t virt_alloc_region(struct kvm_vm *vm, int ri, uint32_t memslot) in virt_alloc_region() argument 45 taddr = vm_phy_pages_alloc(vm, ri < 4 ? PAGES_PER_REGION : 1, in virt_alloc_region() [all …]
|
| /kernel/linux/linux-6.6/drivers/gpu/drm/amd/amdgpu/ |
| D | amdgpu_vm.c | 119 * @vm: pointer to the amdgpu_vm structure to set the fence sequence on 121 struct amdgpu_vm *vm; member 130 * amdgpu_vm_set_pasid - manage pasid and vm ptr mapping 133 * @vm: amdgpu_vm pointer 134 * @pasid: the pasid the VM is using on this GPU 136 * Set the pasid this VM is using on this GPU, can also be used to remove the 140 int amdgpu_vm_set_pasid(struct amdgpu_device *adev, struct amdgpu_vm *vm, in amdgpu_vm_set_pasid() argument 145 if (vm->pasid == pasid) in amdgpu_vm_set_pasid() 148 if (vm->pasid) { in amdgpu_vm_set_pasid() 149 r = xa_err(xa_erase_irq(&adev->vm_manager.pasids, vm->pasid)); in amdgpu_vm_set_pasid() [all …]
|
| /kernel/linux/linux-6.6/drivers/gpu/drm/lima/ |
| D | lima_vm.c | 18 struct lima_vm *vm; member 35 static void lima_vm_unmap_range(struct lima_vm *vm, u32 start, u32 end) in lima_vm_unmap_range() argument 43 vm->bts[pbe].cpu[bte] = 0; in lima_vm_unmap_range() 47 static int lima_vm_map_page(struct lima_vm *vm, dma_addr_t pa, u32 va) in lima_vm_map_page() argument 52 if (!vm->bts[pbe].cpu) { in lima_vm_map_page() 57 vm->bts[pbe].cpu = dma_alloc_wc( in lima_vm_map_page() 58 vm->dev->dev, LIMA_PAGE_SIZE << LIMA_VM_NUM_PT_PER_BT_SHIFT, in lima_vm_map_page() 59 &vm->bts[pbe].dma, GFP_KERNEL | __GFP_NOWARN | __GFP_ZERO); in lima_vm_map_page() 60 if (!vm->bts[pbe].cpu) in lima_vm_map_page() 63 pts = vm->bts[pbe].dma; in lima_vm_map_page() [all …]
|
| /kernel/linux/linux-5.10/drivers/gpu/drm/lima/ |
| D | lima_vm.c | 18 struct lima_vm *vm; member 35 static void lima_vm_unmap_range(struct lima_vm *vm, u32 start, u32 end) in lima_vm_unmap_range() argument 43 vm->bts[pbe].cpu[bte] = 0; in lima_vm_unmap_range() 47 static int lima_vm_map_page(struct lima_vm *vm, dma_addr_t pa, u32 va) in lima_vm_map_page() argument 52 if (!vm->bts[pbe].cpu) { in lima_vm_map_page() 57 vm->bts[pbe].cpu = dma_alloc_wc( in lima_vm_map_page() 58 vm->dev->dev, LIMA_PAGE_SIZE << LIMA_VM_NUM_PT_PER_BT_SHIFT, in lima_vm_map_page() 59 &vm->bts[pbe].dma, GFP_KERNEL | __GFP_NOWARN | __GFP_ZERO); in lima_vm_map_page() 60 if (!vm->bts[pbe].cpu) in lima_vm_map_page() 63 pts = vm->bts[pbe].dma; in lima_vm_map_page() [all …]
|
| /kernel/linux/linux-6.6/drivers/virt/acrn/ |
| D | vm.c | 21 * is wrote in VM creation ioctl. Use the rwlock mechanism to protect it. 25 struct acrn_vm *acrn_vm_create(struct acrn_vm *vm, in acrn_vm_create() argument 33 "Failed to create VM! Error: %d\n", ret); in acrn_vm_create() 37 mutex_init(&vm->regions_mapping_lock); in acrn_vm_create() 38 INIT_LIST_HEAD(&vm->ioreq_clients); in acrn_vm_create() 39 spin_lock_init(&vm->ioreq_clients_lock); in acrn_vm_create() 40 vm->vmid = vm_param->vmid; in acrn_vm_create() 41 vm->vcpu_num = vm_param->vcpu_num; in acrn_vm_create() 43 if (acrn_ioreq_init(vm, vm_param->ioreq_buf) < 0) { in acrn_vm_create() 45 vm->vmid = ACRN_INVALID_VMID; in acrn_vm_create() [all …]
|
| D | irqfd.c | 23 * @vm: Associated VM pointer 27 * @list: Entry within &acrn_vm.irqfds of irqfds of a VM 32 struct acrn_vm *vm; member 43 struct acrn_vm *vm = irqfd->vm; in acrn_irqfd_inject() local 45 acrn_msi_inject(vm, irqfd->msi.msi_addr, in acrn_irqfd_inject() 53 lockdep_assert_held(&irqfd->vm->irqfds_lock); in hsm_irqfd_shutdown() 65 struct acrn_vm *vm; in hsm_irqfd_shutdown_work() local 68 vm = irqfd->vm; in hsm_irqfd_shutdown_work() 69 mutex_lock(&vm->irqfds_lock); in hsm_irqfd_shutdown_work() 72 mutex_unlock(&vm->irqfds_lock); in hsm_irqfd_shutdown_work() [all …]
|
| D | acrn_drv.h | 28 * @user_vm_pa: Physical address of User VM to be mapped. 29 * @service_vm_pa: Physical address of Service VM to be mapped. 33 * to manage the EPT mappings of a single memory region of the User VM. Several 47 * @vmid: A User VM ID. 54 * multiple memory regions of a User VM. A &struct vm_memory_region_batch 67 * struct vm_memory_mapping - Memory map between a User VM and the Service VM 68 * @pages: Pages in Service VM kernel. 70 * @service_vm_va: Virtual address in Service VM kernel. 71 * @user_vm_pa: Physical address in User VM. 74 * HSM maintains memory mappings between a User VM GPA and the Service VM [all …]
|
| /kernel/linux/linux-6.6/tools/testing/selftests/kvm/lib/ |
| D | kvm_util.c | 128 void vm_enable_dirty_ring(struct kvm_vm *vm, uint32_t ring_size) in vm_enable_dirty_ring() argument 130 if (vm_check_cap(vm, KVM_CAP_DIRTY_LOG_RING_ACQ_REL)) in vm_enable_dirty_ring() 131 vm_enable_cap(vm, KVM_CAP_DIRTY_LOG_RING_ACQ_REL, ring_size); in vm_enable_dirty_ring() 133 vm_enable_cap(vm, KVM_CAP_DIRTY_LOG_RING, ring_size); in vm_enable_dirty_ring() 134 vm->dirty_ring_size = ring_size; in vm_enable_dirty_ring() 137 static void vm_open(struct kvm_vm *vm) in vm_open() argument 139 vm->kvm_fd = _open_kvm_dev_path_or_exit(O_RDWR); in vm_open() 143 vm->fd = __kvm_ioctl(vm->kvm_fd, KVM_CREATE_VM, (void *)vm->type); in vm_open() 144 TEST_ASSERT(vm->fd >= 0, KVM_IOCTL_ERROR(KVM_CREATE_VM, vm->fd)); in vm_open() 195 * Initializes vm->vpages_valid to match the canonical VA space of the [all …]
|
| /kernel/linux/linux-6.6/drivers/gpu/drm/i915/gt/ |
| D | intel_ggtt.c | 52 struct drm_i915_private *i915 = ggtt->vm.i915; in ggtt_init_hw() 54 i915_address_space_init(&ggtt->vm, VM_CLASS_GGTT); in ggtt_init_hw() 56 ggtt->vm.is_ggtt = true; in ggtt_init_hw() 59 ggtt->vm.has_read_only = IS_VALLEYVIEW(i915); in ggtt_init_hw() 62 ggtt->vm.mm.color_adjust = i915_ggtt_color_adjust; in ggtt_init_hw() 68 ggtt->vm.cleanup(&ggtt->vm); in ggtt_init_hw() 103 * i915_ggtt_suspend_vm - Suspend the memory mappings for a GGTT or DPT VM 104 * @vm: The VM to suspend the mappings for 109 void i915_ggtt_suspend_vm(struct i915_address_space *vm) in i915_ggtt_suspend_vm() argument 114 drm_WARN_ON(&vm->i915->drm, !vm->is_ggtt && !vm->is_dpt); in i915_ggtt_suspend_vm() [all …]
|
| D | gen8_ppgtt.c | 89 struct drm_i915_private *i915 = ppgtt->vm.i915; in gen8_ppgtt_notify_vgt() 90 struct intel_uncore *uncore = ppgtt->vm.gt->uncore; in gen8_ppgtt_notify_vgt() 101 if (i915_vm_is_4lvl(&ppgtt->vm)) { in gen8_ppgtt_notify_vgt() 179 static unsigned int gen8_pd_top_count(const struct i915_address_space *vm) in gen8_pd_top_count() argument 181 unsigned int shift = __gen8_pte_shift(vm->top); in gen8_pd_top_count() 183 return (vm->total + (1ull << shift) - 1) >> shift; in gen8_pd_top_count() 187 gen8_pdp_for_page_index(struct i915_address_space * const vm, const u64 idx) in gen8_pdp_for_page_index() argument 189 struct i915_ppgtt * const ppgtt = i915_vm_to_ppgtt(vm); in gen8_pdp_for_page_index() 191 if (vm->top == 2) in gen8_pdp_for_page_index() 194 return i915_pd_entry(ppgtt->pd, gen8_pd_index(idx, vm->top)); in gen8_pdp_for_page_index() [all …]
|
| /kernel/linux/linux-6.6/tools/testing/selftests/kvm/lib/s390x/ |
| D | processor.c | 13 void virt_arch_pgd_alloc(struct kvm_vm *vm) in virt_arch_pgd_alloc() argument 17 TEST_ASSERT(vm->page_size == 4096, "Unsupported page size: 0x%x", in virt_arch_pgd_alloc() 18 vm->page_size); in virt_arch_pgd_alloc() 20 if (vm->pgd_created) in virt_arch_pgd_alloc() 23 paddr = vm_phy_pages_alloc(vm, PAGES_PER_REGION, in virt_arch_pgd_alloc() 25 vm->memslots[MEM_REGION_PT]); in virt_arch_pgd_alloc() 26 memset(addr_gpa2hva(vm, paddr), 0xff, PAGES_PER_REGION * vm->page_size); in virt_arch_pgd_alloc() 28 vm->pgd = paddr; in virt_arch_pgd_alloc() 29 vm->pgd_created = true; in virt_arch_pgd_alloc() 37 static uint64_t virt_alloc_region(struct kvm_vm *vm, int ri) in virt_alloc_region() argument [all …]
|
| /kernel/linux/linux-5.10/drivers/gpu/drm/radeon/ |
| D | radeon_vm.c | 37 * for the entire GPU, there are multiple VM page tables active 38 * at any given time. The VM page tables can contain a mix 42 * Each VM has an ID associated with it and there is a page table 78 * radeon_vm_manager_init - init the vm manager 82 * Init the vm manager (cayman+). 100 * radeon_vm_manager_fini - tear down the vm manager 104 * Tear down the VM manager (cayman+). 120 * radeon_vm_get_bos - add the vm BOs to a validation list 122 * @vm: vm providing the BOs 129 struct radeon_vm *vm, in radeon_vm_get_bos() argument [all …]
|
| /kernel/linux/linux-5.10/drivers/gpu/drm/i915/selftests/ |
| D | mock_gtt.c | 27 static void mock_insert_page(struct i915_address_space *vm, in mock_insert_page() argument 35 static void mock_insert_entries(struct i915_address_space *vm, in mock_insert_entries() argument 41 static void mock_bind_ppgtt(struct i915_address_space *vm, in mock_bind_ppgtt() argument 51 static void mock_unbind_ppgtt(struct i915_address_space *vm, in mock_unbind_ppgtt() argument 56 static void mock_cleanup(struct i915_address_space *vm) in mock_cleanup() argument 60 static void mock_clear_range(struct i915_address_space *vm, in mock_clear_range() argument 73 ppgtt->vm.gt = &i915->gt; in mock_ppgtt() 74 ppgtt->vm.i915 = i915; in mock_ppgtt() 75 ppgtt->vm.total = round_down(U64_MAX, PAGE_SIZE); in mock_ppgtt() 76 ppgtt->vm.file = ERR_PTR(-ENODEV); in mock_ppgtt() [all …]
|
| /kernel/linux/linux-6.6/drivers/gpu/drm/radeon/ |
| D | radeon_vm.c | 37 * for the entire GPU, there are multiple VM page tables active 38 * at any given time. The VM page tables can contain a mix 42 * Each VM has an ID associated with it and there is a page table 78 * radeon_vm_manager_init - init the vm manager 82 * Init the vm manager (cayman+). 100 * radeon_vm_manager_fini - tear down the vm manager 104 * Tear down the VM manager (cayman+). 120 * radeon_vm_get_bos - add the vm BOs to a validation list 123 * @vm: vm providing the BOs 130 struct radeon_vm *vm, in radeon_vm_get_bos() argument [all …]
|
| /kernel/linux/linux-5.10/drivers/gpu/drm/amd/amdgpu/ |
| D | amdgpu_vm.c | 46 * for the entire GPU, there are multiple VM page tables active 47 * at any given time. The VM page tables can contain a mix 51 * Each VM has an ID associated with it and there is a page table 88 * vm eviction_lock can be taken in MMU notifiers. Make sure no reclaim-FS 92 static inline void amdgpu_vm_eviction_lock(struct amdgpu_vm *vm) in amdgpu_vm_eviction_lock() argument 94 mutex_lock(&vm->eviction_lock); in amdgpu_vm_eviction_lock() 95 vm->saved_flags = memalloc_nofs_save(); in amdgpu_vm_eviction_lock() 98 static inline int amdgpu_vm_eviction_trylock(struct amdgpu_vm *vm) in amdgpu_vm_eviction_trylock() argument 100 if (mutex_trylock(&vm->eviction_lock)) { in amdgpu_vm_eviction_trylock() 101 vm->saved_flags = memalloc_nofs_save(); in amdgpu_vm_eviction_trylock() [all …]
|
| /kernel/linux/linux-6.6/drivers/gpu/drm/i915/selftests/ |
| D | mock_gtt.c | 27 static void mock_insert_page(struct i915_address_space *vm, in mock_insert_page() argument 35 static void mock_insert_entries(struct i915_address_space *vm, in mock_insert_entries() argument 41 static void mock_bind_ppgtt(struct i915_address_space *vm, in mock_bind_ppgtt() argument 51 static void mock_unbind_ppgtt(struct i915_address_space *vm, in mock_unbind_ppgtt() argument 56 static void mock_cleanup(struct i915_address_space *vm) in mock_cleanup() argument 60 static void mock_clear_range(struct i915_address_space *vm, in mock_clear_range() argument 73 ppgtt->vm.gt = to_gt(i915); in mock_ppgtt() 74 ppgtt->vm.i915 = i915; in mock_ppgtt() 75 ppgtt->vm.total = round_down(U64_MAX, PAGE_SIZE); in mock_ppgtt() 76 ppgtt->vm.dma = i915->drm.dev; in mock_ppgtt() [all …]
|
| /kernel/linux/linux-6.6/tools/testing/selftests/kvm/s390x/ |
| D | cmma_test.c | 102 static void create_main_memslot(struct kvm_vm *vm) in create_main_memslot() argument 106 vm_userspace_mem_region_add(vm, VM_MEM_SRC_ANONYMOUS, 0, 0, MAIN_PAGE_COUNT, 0); in create_main_memslot() 109 vm->memslots[i] = 0; in create_main_memslot() 112 static void create_test_memslot(struct kvm_vm *vm) in create_test_memslot() argument 114 vm_userspace_mem_region_add(vm, in create_test_memslot() 116 TEST_DATA_START_GFN << vm->page_shift, in create_test_memslot() 121 vm->memslots[MEM_REGION_TEST_DATA] = TEST_DATA_MEMSLOT; in create_test_memslot() 124 static void create_memslots(struct kvm_vm *vm) in create_memslots() argument 127 * Our VM has the following memory layout: in create_memslots() 140 create_main_memslot(vm); in create_memslots() [all …]
|
| /kernel/linux/linux-5.10/drivers/gpu/drm/i915/gt/ |
| D | intel_ggtt.c | 44 struct drm_i915_private *i915 = ggtt->vm.i915; in ggtt_init_hw() 46 i915_address_space_init(&ggtt->vm, VM_CLASS_GGTT); in ggtt_init_hw() 48 ggtt->vm.is_ggtt = true; in ggtt_init_hw() 51 ggtt->vm.has_read_only = IS_VALLEYVIEW(i915); in ggtt_init_hw() 54 ggtt->vm.mm.color_adjust = i915_ggtt_color_adjust; in ggtt_init_hw() 60 ggtt->vm.cleanup(&ggtt->vm); in ggtt_init_hw() 112 mutex_lock(&ggtt->vm.mutex); in i915_ggtt_suspend() 115 open = atomic_xchg(&ggtt->vm.open, 0); in i915_ggtt_suspend() 117 list_for_each_entry_safe(vma, vn, &ggtt->vm.bound_list, vm_link) { in i915_ggtt_suspend() 130 ggtt->vm.clear_range(&ggtt->vm, 0, ggtt->vm.total); in i915_ggtt_suspend() [all …]
|
| /kernel/linux/linux-6.6/sound/pci/ctxfi/ |
| D | ctvmem.c | 26 * Find or create vm block based on requested @size. 30 get_vm_block(struct ct_vm *vm, unsigned int size, struct ct_atc *atc) in get_vm_block() argument 36 if (size > vm->size) { in get_vm_block() 42 mutex_lock(&vm->lock); in get_vm_block() 43 list_for_each(pos, &vm->unused) { in get_vm_block() 48 if (pos == &vm->unused) in get_vm_block() 52 /* Move the vm node from unused list to used list directly */ in get_vm_block() 53 list_move(&entry->list, &vm->used); in get_vm_block() 54 vm->size -= size; in get_vm_block() 65 list_add(&block->list, &vm->used); in get_vm_block() [all …]
|