1 // SPDX-License-Identifier: GPL-2.0
2 /*
3 * ACRN HSM irqfd: use eventfd objects to inject virtual interrupts
4 *
5 * Copyright (C) 2020 Intel Corporation. All rights reserved.
6 *
7 * Authors:
8 * Shuo Liu <shuo.a.liu@intel.com>
9 * Yakui Zhao <yakui.zhao@intel.com>
10 */
11
12 #include <linux/eventfd.h>
13 #include <linux/file.h>
14 #include <linux/poll.h>
15 #include <linux/slab.h>
16
17 #include "acrn_drv.h"
18
19 static LIST_HEAD(acrn_irqfd_clients);
20 static DEFINE_MUTEX(acrn_irqfds_mutex);
21
22 /**
23 * struct hsm_irqfd - Properties of HSM irqfd
24 * @vm: Associated VM pointer
25 * @wait: Entry of wait-queue
26 * @shutdown: Async shutdown work
27 * @eventfd: Associated eventfd
28 * @list: Entry within &acrn_vm.irqfds of irqfds of a VM
29 * @pt: Structure for select/poll on the associated eventfd
30 * @msi: MSI data
31 */
32 struct hsm_irqfd {
33 struct acrn_vm *vm;
34 wait_queue_entry_t wait;
35 struct work_struct shutdown;
36 struct eventfd_ctx *eventfd;
37 struct list_head list;
38 poll_table pt;
39 struct acrn_msi_entry msi;
40 };
41
acrn_irqfd_inject(struct hsm_irqfd * irqfd)42 static void acrn_irqfd_inject(struct hsm_irqfd *irqfd)
43 {
44 struct acrn_vm *vm = irqfd->vm;
45
46 acrn_msi_inject(vm, irqfd->msi.msi_addr,
47 irqfd->msi.msi_data);
48 }
49
hsm_irqfd_shutdown(struct hsm_irqfd * irqfd)50 static void hsm_irqfd_shutdown(struct hsm_irqfd *irqfd)
51 {
52 u64 cnt;
53
54 lockdep_assert_held(&irqfd->vm->irqfds_lock);
55
56 /* remove from wait queue */
57 list_del_init(&irqfd->list);
58 eventfd_ctx_remove_wait_queue(irqfd->eventfd, &irqfd->wait, &cnt);
59 eventfd_ctx_put(irqfd->eventfd);
60 kfree(irqfd);
61 }
62
hsm_irqfd_shutdown_work(struct work_struct * work)63 static void hsm_irqfd_shutdown_work(struct work_struct *work)
64 {
65 struct hsm_irqfd *irqfd;
66 struct acrn_vm *vm;
67
68 irqfd = container_of(work, struct hsm_irqfd, shutdown);
69 vm = irqfd->vm;
70 mutex_lock(&vm->irqfds_lock);
71 if (!list_empty(&irqfd->list))
72 hsm_irqfd_shutdown(irqfd);
73 mutex_unlock(&vm->irqfds_lock);
74 }
75
76 /* Called with wqh->lock held and interrupts disabled */
hsm_irqfd_wakeup(wait_queue_entry_t * wait,unsigned int mode,int sync,void * key)77 static int hsm_irqfd_wakeup(wait_queue_entry_t *wait, unsigned int mode,
78 int sync, void *key)
79 {
80 unsigned long poll_bits = (unsigned long)key;
81 struct hsm_irqfd *irqfd;
82 struct acrn_vm *vm;
83
84 irqfd = container_of(wait, struct hsm_irqfd, wait);
85 vm = irqfd->vm;
86 if (poll_bits & POLLIN)
87 /* An event has been signaled, inject an interrupt */
88 acrn_irqfd_inject(irqfd);
89
90 if (poll_bits & POLLHUP)
91 /* Do shutdown work in thread to hold wqh->lock */
92 queue_work(vm->irqfd_wq, &irqfd->shutdown);
93
94 return 0;
95 }
96
hsm_irqfd_poll_func(struct file * file,wait_queue_head_t * wqh,poll_table * pt)97 static void hsm_irqfd_poll_func(struct file *file, wait_queue_head_t *wqh,
98 poll_table *pt)
99 {
100 struct hsm_irqfd *irqfd;
101
102 irqfd = container_of(pt, struct hsm_irqfd, pt);
103 add_wait_queue(wqh, &irqfd->wait);
104 }
105
106 /*
107 * Assign an eventfd to a VM and create a HSM irqfd associated with the
108 * eventfd. The properties of the HSM irqfd are built from a &struct
109 * acrn_irqfd.
110 */
acrn_irqfd_assign(struct acrn_vm * vm,struct acrn_irqfd * args)111 static int acrn_irqfd_assign(struct acrn_vm *vm, struct acrn_irqfd *args)
112 {
113 struct eventfd_ctx *eventfd = NULL;
114 struct hsm_irqfd *irqfd, *tmp;
115 __poll_t events;
116 struct fd f;
117 int ret = 0;
118
119 irqfd = kzalloc(sizeof(*irqfd), GFP_KERNEL);
120 if (!irqfd)
121 return -ENOMEM;
122
123 irqfd->vm = vm;
124 memcpy(&irqfd->msi, &args->msi, sizeof(args->msi));
125 INIT_LIST_HEAD(&irqfd->list);
126 INIT_WORK(&irqfd->shutdown, hsm_irqfd_shutdown_work);
127
128 f = fdget(args->fd);
129 if (!f.file) {
130 ret = -EBADF;
131 goto out;
132 }
133
134 eventfd = eventfd_ctx_fileget(f.file);
135 if (IS_ERR(eventfd)) {
136 ret = PTR_ERR(eventfd);
137 goto fail;
138 }
139
140 irqfd->eventfd = eventfd;
141
142 /*
143 * Install custom wake-up handling to be notified whenever underlying
144 * eventfd is signaled.
145 */
146 init_waitqueue_func_entry(&irqfd->wait, hsm_irqfd_wakeup);
147 init_poll_funcptr(&irqfd->pt, hsm_irqfd_poll_func);
148
149 mutex_lock(&vm->irqfds_lock);
150 list_for_each_entry(tmp, &vm->irqfds, list) {
151 if (irqfd->eventfd != tmp->eventfd)
152 continue;
153 ret = -EBUSY;
154 mutex_unlock(&vm->irqfds_lock);
155 goto fail;
156 }
157 list_add_tail(&irqfd->list, &vm->irqfds);
158 mutex_unlock(&vm->irqfds_lock);
159
160 /* Check the pending event in this stage */
161 events = vfs_poll(f.file, &irqfd->pt);
162
163 if (events & EPOLLIN)
164 acrn_irqfd_inject(irqfd);
165
166 fdput(f);
167 return 0;
168 fail:
169 if (eventfd && !IS_ERR(eventfd))
170 eventfd_ctx_put(eventfd);
171
172 fdput(f);
173 out:
174 kfree(irqfd);
175 return ret;
176 }
177
acrn_irqfd_deassign(struct acrn_vm * vm,struct acrn_irqfd * args)178 static int acrn_irqfd_deassign(struct acrn_vm *vm,
179 struct acrn_irqfd *args)
180 {
181 struct hsm_irqfd *irqfd, *tmp;
182 struct eventfd_ctx *eventfd;
183
184 eventfd = eventfd_ctx_fdget(args->fd);
185 if (IS_ERR(eventfd))
186 return PTR_ERR(eventfd);
187
188 mutex_lock(&vm->irqfds_lock);
189 list_for_each_entry_safe(irqfd, tmp, &vm->irqfds, list) {
190 if (irqfd->eventfd == eventfd) {
191 hsm_irqfd_shutdown(irqfd);
192 break;
193 }
194 }
195 mutex_unlock(&vm->irqfds_lock);
196 eventfd_ctx_put(eventfd);
197
198 return 0;
199 }
200
acrn_irqfd_config(struct acrn_vm * vm,struct acrn_irqfd * args)201 int acrn_irqfd_config(struct acrn_vm *vm, struct acrn_irqfd *args)
202 {
203 int ret;
204
205 if (args->flags & ACRN_IRQFD_FLAG_DEASSIGN)
206 ret = acrn_irqfd_deassign(vm, args);
207 else
208 ret = acrn_irqfd_assign(vm, args);
209
210 return ret;
211 }
212
acrn_irqfd_init(struct acrn_vm * vm)213 int acrn_irqfd_init(struct acrn_vm *vm)
214 {
215 INIT_LIST_HEAD(&vm->irqfds);
216 mutex_init(&vm->irqfds_lock);
217 vm->irqfd_wq = alloc_workqueue("acrn_irqfd-%u", 0, 0, vm->vmid);
218 if (!vm->irqfd_wq)
219 return -ENOMEM;
220
221 dev_dbg(acrn_dev.this_device, "VM %u irqfd init.\n", vm->vmid);
222 return 0;
223 }
224
acrn_irqfd_deinit(struct acrn_vm * vm)225 void acrn_irqfd_deinit(struct acrn_vm *vm)
226 {
227 struct hsm_irqfd *irqfd, *next;
228
229 dev_dbg(acrn_dev.this_device, "VM %u irqfd deinit.\n", vm->vmid);
230 destroy_workqueue(vm->irqfd_wq);
231 mutex_lock(&vm->irqfds_lock);
232 list_for_each_entry_safe(irqfd, next, &vm->irqfds, list)
233 hsm_irqfd_shutdown(irqfd);
234 mutex_unlock(&vm->irqfds_lock);
235 }
236