• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 // SPDX-License-Identifier: GPL-2.0-only
2 /* Copyright (c) 2020 NVIDIA Corporation */
3 
4 #include <linux/host1x.h>
5 #include <linux/iommu.h>
6 #include <linux/list.h>
7 
8 #include <drm/drm_drv.h>
9 #include <drm/drm_file.h>
10 #include <drm/drm_utils.h>
11 
12 #include "drm.h"
13 #include "uapi.h"
14 
tegra_drm_mapping_release(struct kref * ref)15 static void tegra_drm_mapping_release(struct kref *ref)
16 {
17 	struct tegra_drm_mapping *mapping =
18 		container_of(ref, struct tegra_drm_mapping, ref);
19 
20 	host1x_bo_unpin(mapping->map);
21 	host1x_bo_put(mapping->bo);
22 
23 	kfree(mapping);
24 }
25 
tegra_drm_mapping_put(struct tegra_drm_mapping * mapping)26 void tegra_drm_mapping_put(struct tegra_drm_mapping *mapping)
27 {
28 	kref_put(&mapping->ref, tegra_drm_mapping_release);
29 }
30 
tegra_drm_channel_context_close(struct tegra_drm_context * context)31 static void tegra_drm_channel_context_close(struct tegra_drm_context *context)
32 {
33 	struct tegra_drm_mapping *mapping;
34 	unsigned long id;
35 
36 	if (context->memory_context)
37 		host1x_memory_context_put(context->memory_context);
38 
39 	xa_for_each(&context->mappings, id, mapping)
40 		tegra_drm_mapping_put(mapping);
41 
42 	xa_destroy(&context->mappings);
43 
44 	host1x_channel_put(context->channel);
45 
46 	kfree(context);
47 }
48 
tegra_drm_uapi_close_file(struct tegra_drm_file * file)49 void tegra_drm_uapi_close_file(struct tegra_drm_file *file)
50 {
51 	struct tegra_drm_context *context;
52 	struct host1x_syncpt *sp;
53 	unsigned long id;
54 
55 	xa_for_each(&file->contexts, id, context)
56 		tegra_drm_channel_context_close(context);
57 
58 	xa_for_each(&file->syncpoints, id, sp)
59 		host1x_syncpt_put(sp);
60 
61 	xa_destroy(&file->contexts);
62 	xa_destroy(&file->syncpoints);
63 }
64 
tegra_drm_find_client(struct tegra_drm * tegra,u32 class)65 static struct tegra_drm_client *tegra_drm_find_client(struct tegra_drm *tegra, u32 class)
66 {
67 	struct tegra_drm_client *client;
68 
69 	list_for_each_entry(client, &tegra->clients, list)
70 		if (client->base.class == class)
71 			return client;
72 
73 	return NULL;
74 }
75 
tegra_drm_ioctl_channel_open(struct drm_device * drm,void * data,struct drm_file * file)76 int tegra_drm_ioctl_channel_open(struct drm_device *drm, void *data, struct drm_file *file)
77 {
78 	struct host1x *host = tegra_drm_to_host1x(drm->dev_private);
79 	struct tegra_drm_file *fpriv = file->driver_priv;
80 	struct tegra_drm *tegra = drm->dev_private;
81 	struct drm_tegra_channel_open *args = data;
82 	struct tegra_drm_client *client = NULL;
83 	struct tegra_drm_context *context;
84 	int err;
85 
86 	if (args->flags)
87 		return -EINVAL;
88 
89 	context = kzalloc(sizeof(*context), GFP_KERNEL);
90 	if (!context)
91 		return -ENOMEM;
92 
93 	client = tegra_drm_find_client(tegra, args->host1x_class);
94 	if (!client) {
95 		err = -ENODEV;
96 		goto free;
97 	}
98 
99 	if (client->shared_channel) {
100 		context->channel = host1x_channel_get(client->shared_channel);
101 	} else {
102 		context->channel = host1x_channel_request(&client->base);
103 		if (!context->channel) {
104 			err = -EBUSY;
105 			goto free;
106 		}
107 	}
108 
109 	/* Only allocate context if the engine supports context isolation. */
110 	if (device_iommu_mapped(client->base.dev) && client->ops->can_use_memory_ctx) {
111 		bool supported;
112 
113 		err = client->ops->can_use_memory_ctx(client, &supported);
114 		if (err)
115 			goto put_channel;
116 
117 		if (supported)
118 			context->memory_context = host1x_memory_context_alloc(
119 				host, get_task_pid(current, PIDTYPE_TGID));
120 
121 		if (IS_ERR(context->memory_context)) {
122 			if (PTR_ERR(context->memory_context) != -EOPNOTSUPP) {
123 				err = PTR_ERR(context->memory_context);
124 				goto put_channel;
125 			} else {
126 				/*
127 				 * OK, HW does not support contexts or contexts
128 				 * are disabled.
129 				 */
130 				context->memory_context = NULL;
131 			}
132 		}
133 	}
134 
135 	err = xa_alloc(&fpriv->contexts, &args->context, context, XA_LIMIT(1, U32_MAX),
136 		       GFP_KERNEL);
137 	if (err < 0)
138 		goto put_memctx;
139 
140 	context->client = client;
141 	xa_init_flags(&context->mappings, XA_FLAGS_ALLOC1);
142 
143 	args->version = client->version;
144 	args->capabilities = 0;
145 
146 	if (device_get_dma_attr(client->base.dev) == DEV_DMA_COHERENT)
147 		args->capabilities |= DRM_TEGRA_CHANNEL_CAP_CACHE_COHERENT;
148 
149 	return 0;
150 
151 put_memctx:
152 	if (context->memory_context)
153 		host1x_memory_context_put(context->memory_context);
154 put_channel:
155 	host1x_channel_put(context->channel);
156 free:
157 	kfree(context);
158 
159 	return err;
160 }
161 
tegra_drm_ioctl_channel_close(struct drm_device * drm,void * data,struct drm_file * file)162 int tegra_drm_ioctl_channel_close(struct drm_device *drm, void *data, struct drm_file *file)
163 {
164 	struct tegra_drm_file *fpriv = file->driver_priv;
165 	struct drm_tegra_channel_close *args = data;
166 	struct tegra_drm_context *context;
167 
168 	mutex_lock(&fpriv->lock);
169 
170 	context = xa_load(&fpriv->contexts, args->context);
171 	if (!context) {
172 		mutex_unlock(&fpriv->lock);
173 		return -EINVAL;
174 	}
175 
176 	xa_erase(&fpriv->contexts, args->context);
177 
178 	mutex_unlock(&fpriv->lock);
179 
180 	tegra_drm_channel_context_close(context);
181 
182 	return 0;
183 }
184 
tegra_drm_ioctl_channel_map(struct drm_device * drm,void * data,struct drm_file * file)185 int tegra_drm_ioctl_channel_map(struct drm_device *drm, void *data, struct drm_file *file)
186 {
187 	struct tegra_drm_file *fpriv = file->driver_priv;
188 	struct drm_tegra_channel_map *args = data;
189 	struct tegra_drm_mapping *mapping;
190 	struct tegra_drm_context *context;
191 	enum dma_data_direction direction;
192 	struct device *mapping_dev;
193 	int err = 0;
194 
195 	if (args->flags & ~DRM_TEGRA_CHANNEL_MAP_READ_WRITE)
196 		return -EINVAL;
197 
198 	mutex_lock(&fpriv->lock);
199 
200 	context = xa_load(&fpriv->contexts, args->context);
201 	if (!context) {
202 		mutex_unlock(&fpriv->lock);
203 		return -EINVAL;
204 	}
205 
206 	mapping = kzalloc(sizeof(*mapping), GFP_KERNEL);
207 	if (!mapping) {
208 		err = -ENOMEM;
209 		goto unlock;
210 	}
211 
212 	kref_init(&mapping->ref);
213 
214 	if (context->memory_context)
215 		mapping_dev = &context->memory_context->dev;
216 	else
217 		mapping_dev = context->client->base.dev;
218 
219 	mapping->bo = tegra_gem_lookup(file, args->handle);
220 	if (!mapping->bo) {
221 		err = -EINVAL;
222 		goto free;
223 	}
224 
225 	switch (args->flags & DRM_TEGRA_CHANNEL_MAP_READ_WRITE) {
226 	case DRM_TEGRA_CHANNEL_MAP_READ_WRITE:
227 		direction = DMA_BIDIRECTIONAL;
228 		break;
229 
230 	case DRM_TEGRA_CHANNEL_MAP_WRITE:
231 		direction = DMA_FROM_DEVICE;
232 		break;
233 
234 	case DRM_TEGRA_CHANNEL_MAP_READ:
235 		direction = DMA_TO_DEVICE;
236 		break;
237 
238 	default:
239 		err = -EINVAL;
240 		goto put_gem;
241 	}
242 
243 	mapping->map = host1x_bo_pin(mapping_dev, mapping->bo, direction, NULL);
244 	if (IS_ERR(mapping->map)) {
245 		err = PTR_ERR(mapping->map);
246 		goto put_gem;
247 	}
248 
249 	mapping->iova = mapping->map->phys;
250 	mapping->iova_end = mapping->iova + host1x_to_tegra_bo(mapping->bo)->gem.size;
251 
252 	err = xa_alloc(&context->mappings, &args->mapping, mapping, XA_LIMIT(1, U32_MAX),
253 		       GFP_KERNEL);
254 	if (err < 0)
255 		goto unpin;
256 
257 	mutex_unlock(&fpriv->lock);
258 
259 	return 0;
260 
261 unpin:
262 	host1x_bo_unpin(mapping->map);
263 put_gem:
264 	host1x_bo_put(mapping->bo);
265 free:
266 	kfree(mapping);
267 unlock:
268 	mutex_unlock(&fpriv->lock);
269 	return err;
270 }
271 
tegra_drm_ioctl_channel_unmap(struct drm_device * drm,void * data,struct drm_file * file)272 int tegra_drm_ioctl_channel_unmap(struct drm_device *drm, void *data, struct drm_file *file)
273 {
274 	struct tegra_drm_file *fpriv = file->driver_priv;
275 	struct drm_tegra_channel_unmap *args = data;
276 	struct tegra_drm_mapping *mapping;
277 	struct tegra_drm_context *context;
278 
279 	mutex_lock(&fpriv->lock);
280 
281 	context = xa_load(&fpriv->contexts, args->context);
282 	if (!context) {
283 		mutex_unlock(&fpriv->lock);
284 		return -EINVAL;
285 	}
286 
287 	mapping = xa_erase(&context->mappings, args->mapping);
288 
289 	mutex_unlock(&fpriv->lock);
290 
291 	if (!mapping)
292 		return -EINVAL;
293 
294 	tegra_drm_mapping_put(mapping);
295 	return 0;
296 }
297 
tegra_drm_ioctl_syncpoint_allocate(struct drm_device * drm,void * data,struct drm_file * file)298 int tegra_drm_ioctl_syncpoint_allocate(struct drm_device *drm, void *data, struct drm_file *file)
299 {
300 	struct host1x *host1x = tegra_drm_to_host1x(drm->dev_private);
301 	struct tegra_drm_file *fpriv = file->driver_priv;
302 	struct drm_tegra_syncpoint_allocate *args = data;
303 	struct host1x_syncpt *sp;
304 	int err;
305 
306 	if (args->id)
307 		return -EINVAL;
308 
309 	sp = host1x_syncpt_alloc(host1x, HOST1X_SYNCPT_CLIENT_MANAGED, current->comm);
310 	if (!sp)
311 		return -EBUSY;
312 
313 	args->id = host1x_syncpt_id(sp);
314 
315 	err = xa_insert(&fpriv->syncpoints, args->id, sp, GFP_KERNEL);
316 	if (err) {
317 		host1x_syncpt_put(sp);
318 		return err;
319 	}
320 
321 	return 0;
322 }
323 
tegra_drm_ioctl_syncpoint_free(struct drm_device * drm,void * data,struct drm_file * file)324 int tegra_drm_ioctl_syncpoint_free(struct drm_device *drm, void *data, struct drm_file *file)
325 {
326 	struct tegra_drm_file *fpriv = file->driver_priv;
327 	struct drm_tegra_syncpoint_allocate *args = data;
328 	struct host1x_syncpt *sp;
329 
330 	mutex_lock(&fpriv->lock);
331 	sp = xa_erase(&fpriv->syncpoints, args->id);
332 	mutex_unlock(&fpriv->lock);
333 
334 	if (!sp)
335 		return -EINVAL;
336 
337 	host1x_syncpt_put(sp);
338 
339 	return 0;
340 }
341 
tegra_drm_ioctl_syncpoint_wait(struct drm_device * drm,void * data,struct drm_file * file)342 int tegra_drm_ioctl_syncpoint_wait(struct drm_device *drm, void *data, struct drm_file *file)
343 {
344 	struct host1x *host1x = tegra_drm_to_host1x(drm->dev_private);
345 	struct drm_tegra_syncpoint_wait *args = data;
346 	signed long timeout_jiffies;
347 	struct host1x_syncpt *sp;
348 
349 	if (args->padding != 0)
350 		return -EINVAL;
351 
352 	sp = host1x_syncpt_get_by_id_noref(host1x, args->id);
353 	if (!sp)
354 		return -EINVAL;
355 
356 	timeout_jiffies = drm_timeout_abs_to_jiffies(args->timeout_ns);
357 
358 	return host1x_syncpt_wait(sp, args->threshold, timeout_jiffies, &args->value);
359 }
360