/kernel/linux/linux-5.10/drivers/virtio/ |
D | virtio_mem.c | 168 static int register_virtio_mem_device(struct virtio_mem *vm) in register_virtio_mem_device() argument 177 list_add_rcu(&vm->next, &virtio_mem_devices); in register_virtio_mem_device() 187 static void unregister_virtio_mem_device(struct virtio_mem *vm) in unregister_virtio_mem_device() argument 191 list_del_rcu(&vm->next); in unregister_virtio_mem_device() 218 static unsigned long virtio_mem_phys_to_sb_id(struct virtio_mem *vm, in virtio_mem_phys_to_sb_id() argument 224 return (addr - mb_addr) / vm->subblock_size; in virtio_mem_phys_to_sb_id() 230 static void virtio_mem_mb_set_state(struct virtio_mem *vm, unsigned long mb_id, in virtio_mem_mb_set_state() argument 233 const unsigned long idx = mb_id - vm->first_mb_id; in virtio_mem_mb_set_state() 236 old_state = vm->mb_state[idx]; in virtio_mem_mb_set_state() 237 vm->mb_state[idx] = state; in virtio_mem_mb_set_state() [all …]
|
/kernel/linux/linux-5.10/tools/testing/selftests/kvm/lib/aarch64/ |
D | processor.c | 19 static uint64_t page_align(struct kvm_vm *vm, uint64_t v) in page_align() argument 21 return (v + vm->page_size) & ~(vm->page_size - 1); in page_align() 24 static uint64_t pgd_index(struct kvm_vm *vm, vm_vaddr_t gva) in pgd_index() argument 26 unsigned int shift = (vm->pgtable_levels - 1) * (vm->page_shift - 3) + vm->page_shift; in pgd_index() 27 uint64_t mask = (1UL << (vm->va_bits - shift)) - 1; in pgd_index() 32 static uint64_t pud_index(struct kvm_vm *vm, vm_vaddr_t gva) in pud_index() argument 34 unsigned int shift = 2 * (vm->page_shift - 3) + vm->page_shift; in pud_index() 35 uint64_t mask = (1UL << (vm->page_shift - 3)) - 1; in pud_index() 37 TEST_ASSERT(vm->pgtable_levels == 4, in pud_index() 38 "Mode %d does not have 4 page table levels", vm->mode); in pud_index() [all …]
|
/kernel/linux/linux-5.10/tools/testing/selftests/kvm/lib/ |
D | kvm_util.c | 78 int vm_enable_cap(struct kvm_vm *vm, struct kvm_enable_cap *cap) in vm_enable_cap() argument 82 ret = ioctl(vm->fd, KVM_ENABLE_CAP, cap); in vm_enable_cap() 102 int vcpu_enable_cap(struct kvm_vm *vm, uint32_t vcpu_id, in vcpu_enable_cap() argument 105 struct vcpu *vcpu = vcpu_find(vm, vcpu_id); in vcpu_enable_cap() 117 static void vm_open(struct kvm_vm *vm, int perm) in vm_open() argument 119 vm->kvm_fd = open(KVM_DEV_PATH, perm); in vm_open() 120 if (vm->kvm_fd < 0) in vm_open() 128 vm->fd = ioctl(vm->kvm_fd, KVM_CREATE_VM, vm->type); in vm_open() 129 TEST_ASSERT(vm->fd >= 0, "KVM_CREATE_VM ioctl failed, " in vm_open() 130 "rc: %i errno: %i", vm->fd, errno); in vm_open() [all …]
|
/kernel/linux/linux-5.10/tools/testing/selftests/kvm/lib/s390x/ |
D | processor.c | 18 void virt_pgd_alloc(struct kvm_vm *vm, uint32_t memslot) in virt_pgd_alloc() argument 22 TEST_ASSERT(vm->page_size == 4096, "Unsupported page size: 0x%x", in virt_pgd_alloc() 23 vm->page_size); in virt_pgd_alloc() 25 if (vm->pgd_created) in virt_pgd_alloc() 28 paddr = vm_phy_pages_alloc(vm, PAGES_PER_REGION, in virt_pgd_alloc() 30 memset(addr_gpa2hva(vm, paddr), 0xff, PAGES_PER_REGION * vm->page_size); in virt_pgd_alloc() 32 vm->pgd = paddr; in virt_pgd_alloc() 33 vm->pgd_created = true; in virt_pgd_alloc() 41 static uint64_t virt_alloc_region(struct kvm_vm *vm, int ri, uint32_t memslot) in virt_alloc_region() argument 45 taddr = vm_phy_pages_alloc(vm, ri < 4 ? PAGES_PER_REGION : 1, in virt_alloc_region() [all …]
|
/kernel/linux/linux-5.10/drivers/gpu/drm/lima/ |
D | lima_vm.c | 18 struct lima_vm *vm; member 35 static void lima_vm_unmap_range(struct lima_vm *vm, u32 start, u32 end) in lima_vm_unmap_range() argument 43 vm->bts[pbe].cpu[bte] = 0; in lima_vm_unmap_range() 47 static int lima_vm_map_page(struct lima_vm *vm, dma_addr_t pa, u32 va) in lima_vm_map_page() argument 52 if (!vm->bts[pbe].cpu) { in lima_vm_map_page() 57 vm->bts[pbe].cpu = dma_alloc_wc( in lima_vm_map_page() 58 vm->dev->dev, LIMA_PAGE_SIZE << LIMA_VM_NUM_PT_PER_BT_SHIFT, in lima_vm_map_page() 59 &vm->bts[pbe].dma, GFP_KERNEL | __GFP_NOWARN | __GFP_ZERO); in lima_vm_map_page() 60 if (!vm->bts[pbe].cpu) in lima_vm_map_page() 63 pts = vm->bts[pbe].dma; in lima_vm_map_page() [all …]
|
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/selftests/ |
D | mock_gtt.c | 27 static void mock_insert_page(struct i915_address_space *vm, in mock_insert_page() argument 35 static void mock_insert_entries(struct i915_address_space *vm, in mock_insert_entries() argument 41 static void mock_bind_ppgtt(struct i915_address_space *vm, in mock_bind_ppgtt() argument 51 static void mock_unbind_ppgtt(struct i915_address_space *vm, in mock_unbind_ppgtt() argument 56 static void mock_cleanup(struct i915_address_space *vm) in mock_cleanup() argument 60 static void mock_clear_range(struct i915_address_space *vm, in mock_clear_range() argument 73 ppgtt->vm.gt = &i915->gt; in mock_ppgtt() 74 ppgtt->vm.i915 = i915; in mock_ppgtt() 75 ppgtt->vm.total = round_down(U64_MAX, PAGE_SIZE); in mock_ppgtt() 76 ppgtt->vm.file = ERR_PTR(-ENODEV); in mock_ppgtt() [all …]
|
D | i915_gem_gtt.c | 160 if (!ppgtt->vm.allocate_va_range) in igt_ppgtt_alloc() 171 limit = min(ppgtt->vm.total, limit); in igt_ppgtt_alloc() 177 err = i915_vm_alloc_pt_stash(&ppgtt->vm, &stash, size); in igt_ppgtt_alloc() 181 err = i915_vm_pin_pt_stash(&ppgtt->vm, &stash); in igt_ppgtt_alloc() 183 i915_vm_free_pt_stash(&ppgtt->vm, &stash); in igt_ppgtt_alloc() 187 ppgtt->vm.allocate_va_range(&ppgtt->vm, &stash, 0, size); in igt_ppgtt_alloc() 190 ppgtt->vm.clear_range(&ppgtt->vm, 0, size); in igt_ppgtt_alloc() 192 i915_vm_free_pt_stash(&ppgtt->vm, &stash); in igt_ppgtt_alloc() 199 err = i915_vm_alloc_pt_stash(&ppgtt->vm, &stash, size - last); in igt_ppgtt_alloc() 203 err = i915_vm_pin_pt_stash(&ppgtt->vm, &stash); in igt_ppgtt_alloc() [all …]
|
/kernel/linux/linux-5.10/tools/testing/selftests/kvm/include/ |
D | kvm_util.h | 65 int vm_enable_cap(struct kvm_vm *vm, struct kvm_enable_cap *cap); 66 int vcpu_enable_cap(struct kvm_vm *vm, uint32_t vcpu_id, 68 void vm_enable_dirty_ring(struct kvm_vm *vm, uint32_t ring_size); 74 void kvm_vm_get_dirty_log(struct kvm_vm *vm, int slot, void *log); 75 void kvm_vm_clear_dirty_log(struct kvm_vm *vm, int slot, void *log, 78 int kvm_memcmp_hva_gva(void *hva, struct kvm_vm *vm, const vm_vaddr_t gva, 81 void kvm_vm_elf_load(struct kvm_vm *vm, const char *filename, 84 void vm_dump(FILE *stream, struct kvm_vm *vm, uint8_t indent); 102 void vcpu_dump(FILE *stream, struct kvm_vm *vm, uint32_t vcpuid, 105 void vm_create_irqchip(struct kvm_vm *vm); [all …]
|
/kernel/linux/linux-5.10/sound/pci/ctxfi/ |
D | ctvmem.c | 30 get_vm_block(struct ct_vm *vm, unsigned int size, struct ct_atc *atc) in get_vm_block() argument 36 if (size > vm->size) { in get_vm_block() 42 mutex_lock(&vm->lock); in get_vm_block() 43 list_for_each(pos, &vm->unused) { in get_vm_block() 48 if (pos == &vm->unused) in get_vm_block() 53 list_move(&entry->list, &vm->used); in get_vm_block() 54 vm->size -= size; in get_vm_block() 65 list_add(&block->list, &vm->used); in get_vm_block() 68 vm->size -= size; in get_vm_block() 71 mutex_unlock(&vm->lock); in get_vm_block() [all …]
|
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gt/ |
D | intel_ggtt.c | 44 struct drm_i915_private *i915 = ggtt->vm.i915; in ggtt_init_hw() 46 i915_address_space_init(&ggtt->vm, VM_CLASS_GGTT); in ggtt_init_hw() 48 ggtt->vm.is_ggtt = true; in ggtt_init_hw() 51 ggtt->vm.has_read_only = IS_VALLEYVIEW(i915); in ggtt_init_hw() 54 ggtt->vm.mm.color_adjust = i915_ggtt_color_adjust; in ggtt_init_hw() 60 ggtt->vm.cleanup(&ggtt->vm); in ggtt_init_hw() 112 mutex_lock(&ggtt->vm.mutex); in i915_ggtt_suspend() 115 open = atomic_xchg(&ggtt->vm.open, 0); in i915_ggtt_suspend() 117 list_for_each_entry_safe(vma, vn, &ggtt->vm.bound_list, vm_link) { in i915_ggtt_suspend() 130 ggtt->vm.clear_range(&ggtt->vm, 0, ggtt->vm.total); in i915_ggtt_suspend() [all …]
|
D | gen8_ppgtt.c | 55 struct drm_i915_private *i915 = ppgtt->vm.i915; in gen8_ppgtt_notify_vgt() 56 struct intel_uncore *uncore = ppgtt->vm.gt->uncore; in gen8_ppgtt_notify_vgt() 67 if (i915_vm_is_4lvl(&ppgtt->vm)) { in gen8_ppgtt_notify_vgt() 146 gen8_pd_top_count(const struct i915_address_space *vm) in gen8_pd_top_count() argument 148 unsigned int shift = __gen8_pte_shift(vm->top); in gen8_pd_top_count() 149 return (vm->total + (1ull << shift) - 1) >> shift; in gen8_pd_top_count() 153 gen8_pdp_for_page_index(struct i915_address_space * const vm, const u64 idx) in gen8_pdp_for_page_index() argument 155 struct i915_ppgtt * const ppgtt = i915_vm_to_ppgtt(vm); in gen8_pdp_for_page_index() 157 if (vm->top == 2) in gen8_pdp_for_page_index() 160 return i915_pd_entry(ppgtt->pd, gen8_pd_index(idx, vm->top)); in gen8_pdp_for_page_index() [all …]
|
D | intel_gtt.h | 61 #define ggtt_total_entries(ggtt) ((ggtt)->vm.total >> PAGE_SHIFT) 192 void (*bind_vma)(struct i915_address_space *vm, 201 void (*unbind_vma)(struct i915_address_space *vm, 261 (*alloc_pt_dma)(struct i915_address_space *vm, int sz); 268 void (*allocate_va_range)(struct i915_address_space *vm, 271 void (*clear_range)(struct i915_address_space *vm, 273 void (*insert_page)(struct i915_address_space *vm, 278 void (*insert_entries)(struct i915_address_space *vm, 282 void (*cleanup)(struct i915_address_space *vm); 299 struct i915_address_space vm; member [all …]
|
D | intel_gtt.c | 14 struct drm_i915_gem_object *alloc_pt_dma(struct i915_address_space *vm, int sz) in alloc_pt_dma() argument 16 if (I915_SELFTEST_ONLY(should_fail(&vm->fault_attr, 1))) in alloc_pt_dma() 17 i915_gem_shrink_all(vm->i915); in alloc_pt_dma() 19 return i915_gem_object_create_internal(vm->i915, sz); in alloc_pt_dma() 22 int pin_pt_dma(struct i915_address_space *vm, struct drm_i915_gem_object *obj) in pin_pt_dma() argument 34 void __i915_vm_close(struct i915_address_space *vm) in __i915_vm_close() argument 38 if (!atomic_dec_and_mutex_lock(&vm->open, &vm->mutex)) in __i915_vm_close() 41 list_for_each_entry_safe(vma, vn, &vm->bound_list, vm_link) { in __i915_vm_close() 54 GEM_BUG_ON(!list_empty(&vm->bound_list)); in __i915_vm_close() 56 mutex_unlock(&vm->mutex); in __i915_vm_close() [all …]
|
D | gen6_ppgtt.c | 19 dma_addr_t addr = pt ? px_dma(pt) : px_dma(ppgtt->base.vm.scratch[1]); in gen6_write_pde() 79 static void gen6_ppgtt_clear_range(struct i915_address_space *vm, in gen6_ppgtt_clear_range() argument 82 struct gen6_ppgtt * const ppgtt = to_gen6_ppgtt(i915_vm_to_ppgtt(vm)); in gen6_ppgtt_clear_range() 84 const gen6_pte_t scratch_pte = vm->scratch[0]->encode; in gen6_ppgtt_clear_range() 116 static void gen6_ppgtt_insert_entries(struct i915_address_space *vm, in gen6_ppgtt_insert_entries() argument 121 struct i915_ppgtt *ppgtt = i915_vm_to_ppgtt(vm); in gen6_ppgtt_insert_entries() 126 const u32 pte_encode = vm->pte_encode(0, cache_level, flags); in gen6_ppgtt_insert_entries() 174 gen6_ggtt_invalidate(ppgtt->base.vm.gt->ggtt); in gen6_flush_pd() 180 static void gen6_alloc_va_range(struct i915_address_space *vm, in gen6_alloc_va_range() argument 184 struct gen6_ppgtt *ppgtt = to_gen6_ppgtt(i915_vm_to_ppgtt(vm)); in gen6_alloc_va_range() [all …]
|
/kernel/linux/linux-5.10/tools/testing/selftests/kvm/lib/x86_64/ |
D | processor.c | 212 void virt_pgd_alloc(struct kvm_vm *vm, uint32_t pgd_memslot) in virt_pgd_alloc() argument 214 TEST_ASSERT(vm->mode == VM_MODE_PXXV48_4K, "Attempt to use " in virt_pgd_alloc() 215 "unknown or unsupported guest mode, mode: 0x%x", vm->mode); in virt_pgd_alloc() 218 if (!vm->pgd_created) { in virt_pgd_alloc() 219 vm_paddr_t paddr = vm_phy_page_alloc(vm, in virt_pgd_alloc() 221 vm->pgd = paddr; in virt_pgd_alloc() 222 vm->pgd_created = true; in virt_pgd_alloc() 226 void virt_pg_map(struct kvm_vm *vm, uint64_t vaddr, uint64_t paddr, in virt_pg_map() argument 232 TEST_ASSERT(vm->mode == VM_MODE_PXXV48_4K, "Attempt to use " in virt_pg_map() 233 "unknown or unsupported guest mode, mode: 0x%x", vm->mode); in virt_pg_map() [all …]
|
D | vmx.c | 46 int vcpu_enable_evmcs(struct kvm_vm *vm, int vcpu_id) in vcpu_enable_evmcs() argument 55 vcpu_ioctl(vm, vcpu_id, KVM_ENABLE_CAP, &enable_evmcs_cap); in vcpu_enable_evmcs() 78 vcpu_alloc_vmx(struct kvm_vm *vm, vm_vaddr_t *p_vmx_gva) in vcpu_alloc_vmx() argument 80 vm_vaddr_t vmx_gva = vm_vaddr_alloc(vm, getpagesize(), 0x10000, 0, 0); in vcpu_alloc_vmx() 81 struct vmx_pages *vmx = addr_gva2hva(vm, vmx_gva); in vcpu_alloc_vmx() 84 vmx->vmxon = (void *)vm_vaddr_alloc(vm, getpagesize(), 0x10000, 0, 0); in vcpu_alloc_vmx() 85 vmx->vmxon_hva = addr_gva2hva(vm, (uintptr_t)vmx->vmxon); in vcpu_alloc_vmx() 86 vmx->vmxon_gpa = addr_gva2gpa(vm, (uintptr_t)vmx->vmxon); in vcpu_alloc_vmx() 89 vmx->vmcs = (void *)vm_vaddr_alloc(vm, getpagesize(), 0x10000, 0, 0); in vcpu_alloc_vmx() 90 vmx->vmcs_hva = addr_gva2hva(vm, (uintptr_t)vmx->vmcs); in vcpu_alloc_vmx() [all …]
|
/kernel/linux/linux-5.10/drivers/gpu/drm/radeon/ |
D | radeon_vm.c | 129 struct radeon_vm *vm, in radeon_vm_get_bos() argument 135 list = kvmalloc_array(vm->max_pde_used + 2, in radeon_vm_get_bos() 141 list[0].robj = vm->page_directory; in radeon_vm_get_bos() 144 list[0].tv.bo = &vm->page_directory->tbo; in radeon_vm_get_bos() 149 for (i = 0, idx = 1; i <= vm->max_pde_used; i++) { in radeon_vm_get_bos() 150 if (!vm->page_tables[i].bo) in radeon_vm_get_bos() 153 list[idx].robj = vm->page_tables[i].bo; in radeon_vm_get_bos() 178 struct radeon_vm *vm, int ring) in radeon_vm_grab_id() argument 181 struct radeon_vm_id *vm_id = &vm->ids[ring]; in radeon_vm_grab_id() 237 struct radeon_vm *vm, in radeon_vm_flush() argument [all …]
|
/kernel/linux/linux-5.10/tools/testing/selftests/kvm/x86_64/ |
D | vmx_set_nested_state_test.c | 30 void test_nested_state(struct kvm_vm *vm, struct kvm_nested_state *state) in test_nested_state() argument 32 vcpu_nested_state_set(vm, VCPU_ID, state, false); in test_nested_state() 35 void test_nested_state_expect_errno(struct kvm_vm *vm, in test_nested_state_expect_errno() argument 41 rv = vcpu_nested_state_set(vm, VCPU_ID, state, true); in test_nested_state_expect_errno() 48 void test_nested_state_expect_einval(struct kvm_vm *vm, in test_nested_state_expect_einval() argument 51 test_nested_state_expect_errno(vm, state, EINVAL); in test_nested_state_expect_einval() 54 void test_nested_state_expect_efault(struct kvm_vm *vm, in test_nested_state_expect_efault() argument 57 test_nested_state_expect_errno(vm, state, EFAULT); in test_nested_state_expect_efault() 89 void test_vmx_nested_state(struct kvm_vm *vm) in test_vmx_nested_state() argument 99 test_nested_state_expect_einval(vm, state); in test_vmx_nested_state() [all …]
|
D | platform_info_test.c | 38 static void set_msr_platform_info_enabled(struct kvm_vm *vm, bool enable) in set_msr_platform_info_enabled() argument 45 vm_enable_cap(vm, &cap); in set_msr_platform_info_enabled() 48 static void test_msr_platform_info_enabled(struct kvm_vm *vm) in test_msr_platform_info_enabled() argument 50 struct kvm_run *run = vcpu_state(vm, VCPU_ID); in test_msr_platform_info_enabled() 53 set_msr_platform_info_enabled(vm, true); in test_msr_platform_info_enabled() 54 vcpu_run(vm, VCPU_ID); in test_msr_platform_info_enabled() 59 get_ucall(vm, VCPU_ID, &uc); in test_msr_platform_info_enabled() 68 static void test_msr_platform_info_disabled(struct kvm_vm *vm) in test_msr_platform_info_disabled() argument 70 struct kvm_run *run = vcpu_state(vm, VCPU_ID); in test_msr_platform_info_disabled() 72 set_msr_platform_info_enabled(vm, false); in test_msr_platform_info_disabled() [all …]
|
D | smm_test.c | 97 struct kvm_vm *vm; in main() local 103 vm = vm_create_default(VCPU_ID, 0, guest_code); in main() 105 vcpu_set_cpuid(vm, VCPU_ID, kvm_get_supported_cpuid()); in main() 107 run = vcpu_state(vm, VCPU_ID); in main() 109 vm_userspace_mem_region_add(vm, VM_MEM_SRC_ANONYMOUS, SMRAM_GPA, in main() 111 TEST_ASSERT(vm_phy_pages_alloc(vm, SMRAM_PAGES, SMRAM_GPA, SMRAM_MEMSLOT) in main() 114 memset(addr_gpa2hva(vm, SMRAM_GPA), 0x0, SMRAM_SIZE); in main() 115 memcpy(addr_gpa2hva(vm, SMRAM_GPA) + 0x8000, smi_handler, in main() 118 vcpu_set_msr(vm, VCPU_ID, MSR_IA32_SMBASE, SMRAM_GPA); in main() 122 vcpu_alloc_svm(vm, &nested_gva); in main() [all …]
|
/kernel/linux/linux-5.10/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_vm.c | 92 static inline void amdgpu_vm_eviction_lock(struct amdgpu_vm *vm) in amdgpu_vm_eviction_lock() argument 94 mutex_lock(&vm->eviction_lock); in amdgpu_vm_eviction_lock() 95 vm->saved_flags = memalloc_nofs_save(); in amdgpu_vm_eviction_lock() 98 static inline int amdgpu_vm_eviction_trylock(struct amdgpu_vm *vm) in amdgpu_vm_eviction_trylock() argument 100 if (mutex_trylock(&vm->eviction_lock)) { in amdgpu_vm_eviction_trylock() 101 vm->saved_flags = memalloc_nofs_save(); in amdgpu_vm_eviction_trylock() 107 static inline void amdgpu_vm_eviction_unlock(struct amdgpu_vm *vm) in amdgpu_vm_eviction_unlock() argument 109 memalloc_nofs_restore(vm->saved_flags); in amdgpu_vm_eviction_unlock() 110 mutex_unlock(&vm->eviction_lock); in amdgpu_vm_eviction_unlock() 225 struct amdgpu_vm *vm = vm_bo->vm; in amdgpu_vm_bo_evicted() local [all …]
|
/kernel/linux/linux-5.10/tools/testing/selftests/kvm/ |
D | set_memory_region_test.c | 57 struct kvm_vm *vm = data; in vcpu_worker() local 67 run = vcpu_state(vm, VCPU_ID); in vcpu_worker() 70 vcpu_run(vm, VCPU_ID); in vcpu_worker() 73 cmd = get_ucall(vm, VCPU_ID, &uc); in vcpu_worker() 118 struct kvm_vm *vm; in spawn_vm() local 122 vm = vm_create_default(VCPU_ID, 0, guest_code); in spawn_vm() 124 vcpu_set_cpuid(vm, VCPU_ID, kvm_get_supported_cpuid()); in spawn_vm() 126 vm_userspace_mem_region_add(vm, VM_MEM_SRC_ANONYMOUS_THP, in spawn_vm() 134 gpa = vm_phy_pages_alloc(vm, 2, MEM_REGION_GPA, MEM_REGION_SLOT); in spawn_vm() 137 virt_map(vm, MEM_REGION_GPA, MEM_REGION_GPA, 2, 0); in spawn_vm() [all …]
|
/kernel/linux/linux-5.10/drivers/irqchip/ |
D | irq-gic-v4.c | 136 int its_alloc_vcpu_irqs(struct its_vm *vm) in its_alloc_vcpu_irqs() argument 140 vm->fwnode = irq_domain_alloc_named_id_fwnode("GICv4-vpe", in its_alloc_vcpu_irqs() 142 if (!vm->fwnode) in its_alloc_vcpu_irqs() 145 vm->domain = irq_domain_create_hierarchy(gic_domain, 0, vm->nr_vpes, in its_alloc_vcpu_irqs() 146 vm->fwnode, vpe_domain_ops, in its_alloc_vcpu_irqs() 147 vm); in its_alloc_vcpu_irqs() 148 if (!vm->domain) in its_alloc_vcpu_irqs() 151 for (i = 0; i < vm->nr_vpes; i++) { in its_alloc_vcpu_irqs() 152 vm->vpes[i]->its_vm = vm; in its_alloc_vcpu_irqs() 153 vm->vpes[i]->idai = true; in its_alloc_vcpu_irqs() [all …]
|
/kernel/linux/linux-5.10/drivers/gpu/drm/imx/dcss/ |
D | dcss-ss.c | 126 void dcss_ss_sync_set(struct dcss_ss *ss, struct videomode *vm, in dcss_ss_sync_set() argument 135 lrc_x = vm->hfront_porch + vm->hback_porch + vm->hsync_len + in dcss_ss_sync_set() 136 vm->hactive - 1; in dcss_ss_sync_set() 137 lrc_y = vm->vfront_porch + vm->vback_porch + vm->vsync_len + in dcss_ss_sync_set() 138 vm->vactive - 1; in dcss_ss_sync_set() 142 hsync_start = vm->hfront_porch + vm->hback_porch + vm->hsync_len + in dcss_ss_sync_set() 143 vm->hactive - 1; in dcss_ss_sync_set() 144 hsync_end = vm->hsync_len - 1; in dcss_ss_sync_set() 150 vsync_start = vm->vfront_porch - 1; in dcss_ss_sync_set() 151 vsync_end = vm->vfront_porch + vm->vsync_len - 1; in dcss_ss_sync_set() [all …]
|
/kernel/linux/linux-5.10/drivers/video/fbdev/omap2/omapfb/dss/ |
D | display.c | 259 void videomode_to_omap_video_timings(const struct videomode *vm, in videomode_to_omap_video_timings() argument 264 ovt->pixelclock = vm->pixelclock; in videomode_to_omap_video_timings() 265 ovt->x_res = vm->hactive; in videomode_to_omap_video_timings() 266 ovt->hbp = vm->hback_porch; in videomode_to_omap_video_timings() 267 ovt->hfp = vm->hfront_porch; in videomode_to_omap_video_timings() 268 ovt->hsw = vm->hsync_len; in videomode_to_omap_video_timings() 269 ovt->y_res = vm->vactive; in videomode_to_omap_video_timings() 270 ovt->vbp = vm->vback_porch; in videomode_to_omap_video_timings() 271 ovt->vfp = vm->vfront_porch; in videomode_to_omap_video_timings() 272 ovt->vsw = vm->vsync_len; in videomode_to_omap_video_timings() [all …]
|