• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /*
2  * Copyright 2006 Tungsten Graphics Inc., Bismarck, ND., USA.
3  * All rights reserved.
4  *
5  * Permission is hereby granted, free of charge, to any person obtaining a
6  * copy of this software and associated documentation files (the "Software"),
7  * to deal in the Software without restriction, including without limitation
8  * the rights to use, copy, modify, merge, publish, distribute, sub license,
9  * and/or sell copies of the Software, and to permit persons to whom the
10  * Software is furnished to do so, subject to the following conditions:
11  *
12  * The above copyright notice and this permission notice (including the
13  * next paragraph) shall be included in all copies or substantial portions
14  * of the Software.
15  *
16  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18  * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
19  * THE AUTHORS OR COPYRIGHT HOLDERS AND/OR THEIR SUPPLIERS BE LIABLE FOR ANY CLAIM,
20  * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
21  * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
22  * DEALINGS IN THE SOFTWARE.
23  */
24 /*
25  * Authors: Thomas Hellström <thomas-at-tungstengraphics-dot-com>
26  */
27 
28 #include <linux/slab.h>
29 
30 #include <drm/drm_device.h>
31 #include <drm/drm_file.h>
32 #include <drm/via_drm.h>
33 
34 #include "via_drv.h"
35 
36 #define VIA_MM_ALIGN_SHIFT 4
37 #define VIA_MM_ALIGN_MASK ((1 << VIA_MM_ALIGN_SHIFT) - 1)
38 
39 struct via_memblock {
40 	struct drm_mm_node mm_node;
41 	struct list_head owner_list;
42 };
43 
via_agp_init(struct drm_device * dev,void * data,struct drm_file * file_priv)44 int via_agp_init(struct drm_device *dev, void *data, struct drm_file *file_priv)
45 {
46 	drm_via_agp_t *agp = data;
47 	drm_via_private_t *dev_priv = (drm_via_private_t *) dev->dev_private;
48 
49 	mutex_lock(&dev->struct_mutex);
50 	drm_mm_init(&dev_priv->agp_mm, 0, agp->size >> VIA_MM_ALIGN_SHIFT);
51 
52 	dev_priv->agp_initialized = 1;
53 	dev_priv->agp_offset = agp->offset;
54 	mutex_unlock(&dev->struct_mutex);
55 
56 	DRM_DEBUG("offset = %u, size = %u\n", agp->offset, agp->size);
57 	return 0;
58 }
59 
via_fb_init(struct drm_device * dev,void * data,struct drm_file * file_priv)60 int via_fb_init(struct drm_device *dev, void *data, struct drm_file *file_priv)
61 {
62 	drm_via_fb_t *fb = data;
63 	drm_via_private_t *dev_priv = (drm_via_private_t *) dev->dev_private;
64 
65 	mutex_lock(&dev->struct_mutex);
66 	drm_mm_init(&dev_priv->vram_mm, 0, fb->size >> VIA_MM_ALIGN_SHIFT);
67 
68 	dev_priv->vram_initialized = 1;
69 	dev_priv->vram_offset = fb->offset;
70 
71 	mutex_unlock(&dev->struct_mutex);
72 	DRM_DEBUG("offset = %u, size = %u\n", fb->offset, fb->size);
73 
74 	return 0;
75 
76 }
77 
via_final_context(struct drm_device * dev,int context)78 int via_final_context(struct drm_device *dev, int context)
79 {
80 	drm_via_private_t *dev_priv = (drm_via_private_t *) dev->dev_private;
81 
82 	via_release_futex(dev_priv, context);
83 
84 	/* Linux specific until context tracking code gets ported to BSD */
85 	/* Last context, perform cleanup */
86 	if (list_is_singular(&dev->ctxlist)) {
87 		DRM_DEBUG("Last Context\n");
88 		drm_legacy_irq_uninstall(dev);
89 		via_cleanup_futex(dev_priv);
90 		via_do_cleanup_map(dev);
91 	}
92 	return 1;
93 }
94 
via_lastclose(struct drm_device * dev)95 void via_lastclose(struct drm_device *dev)
96 {
97 	drm_via_private_t *dev_priv = (drm_via_private_t *) dev->dev_private;
98 
99 	if (!dev_priv)
100 		return;
101 
102 	mutex_lock(&dev->struct_mutex);
103 	if (dev_priv->vram_initialized) {
104 		drm_mm_takedown(&dev_priv->vram_mm);
105 		dev_priv->vram_initialized = 0;
106 	}
107 	if (dev_priv->agp_initialized) {
108 		drm_mm_takedown(&dev_priv->agp_mm);
109 		dev_priv->agp_initialized = 0;
110 	}
111 	mutex_unlock(&dev->struct_mutex);
112 }
113 
via_mem_alloc(struct drm_device * dev,void * data,struct drm_file * file)114 int via_mem_alloc(struct drm_device *dev, void *data,
115 		  struct drm_file *file)
116 {
117 	drm_via_mem_t *mem = data;
118 	int retval = 0, user_key;
119 	struct via_memblock *item;
120 	drm_via_private_t *dev_priv = (drm_via_private_t *) dev->dev_private;
121 	struct via_file_private *file_priv = file->driver_priv;
122 	unsigned long tmpSize;
123 
124 	if (mem->type > VIA_MEM_AGP) {
125 		DRM_ERROR("Unknown memory type allocation\n");
126 		return -EINVAL;
127 	}
128 	mutex_lock(&dev->struct_mutex);
129 	if (0 == ((mem->type == VIA_MEM_VIDEO) ? dev_priv->vram_initialized :
130 		      dev_priv->agp_initialized)) {
131 		mutex_unlock(&dev->struct_mutex);
132 		DRM_ERROR
133 		    ("Attempt to allocate from uninitialized memory manager.\n");
134 		return -EINVAL;
135 	}
136 
137 	item = kzalloc(sizeof(*item), GFP_KERNEL);
138 	if (!item) {
139 		retval = -ENOMEM;
140 		goto fail_alloc;
141 	}
142 
143 	tmpSize = (mem->size + VIA_MM_ALIGN_MASK) >> VIA_MM_ALIGN_SHIFT;
144 	if (mem->type == VIA_MEM_AGP)
145 		retval = drm_mm_insert_node(&dev_priv->agp_mm,
146 					    &item->mm_node,
147 					    tmpSize);
148 	else
149 		retval = drm_mm_insert_node(&dev_priv->vram_mm,
150 					    &item->mm_node,
151 					    tmpSize);
152 	if (retval)
153 		goto fail_alloc;
154 
155 	retval = idr_alloc(&dev_priv->object_idr, item, 1, 0, GFP_KERNEL);
156 	if (retval < 0)
157 		goto fail_idr;
158 	user_key = retval;
159 
160 	list_add(&item->owner_list, &file_priv->obj_list);
161 	mutex_unlock(&dev->struct_mutex);
162 
163 	mem->offset = ((mem->type == VIA_MEM_VIDEO) ?
164 		      dev_priv->vram_offset : dev_priv->agp_offset) +
165 	    ((item->mm_node.start) << VIA_MM_ALIGN_SHIFT);
166 	mem->index = user_key;
167 
168 	return 0;
169 
170 fail_idr:
171 	drm_mm_remove_node(&item->mm_node);
172 fail_alloc:
173 	kfree(item);
174 	mutex_unlock(&dev->struct_mutex);
175 
176 	mem->offset = 0;
177 	mem->size = 0;
178 	mem->index = 0;
179 	DRM_DEBUG("Video memory allocation failed\n");
180 
181 	return retval;
182 }
183 
via_mem_free(struct drm_device * dev,void * data,struct drm_file * file_priv)184 int via_mem_free(struct drm_device *dev, void *data, struct drm_file *file_priv)
185 {
186 	drm_via_private_t *dev_priv = dev->dev_private;
187 	drm_via_mem_t *mem = data;
188 	struct via_memblock *obj;
189 
190 	mutex_lock(&dev->struct_mutex);
191 	obj = idr_find(&dev_priv->object_idr, mem->index);
192 	if (obj == NULL) {
193 		mutex_unlock(&dev->struct_mutex);
194 		return -EINVAL;
195 	}
196 
197 	idr_remove(&dev_priv->object_idr, mem->index);
198 	list_del(&obj->owner_list);
199 	drm_mm_remove_node(&obj->mm_node);
200 	kfree(obj);
201 	mutex_unlock(&dev->struct_mutex);
202 
203 	DRM_DEBUG("free = 0x%lx\n", mem->index);
204 
205 	return 0;
206 }
207 
208 
via_reclaim_buffers_locked(struct drm_device * dev,struct drm_file * file)209 void via_reclaim_buffers_locked(struct drm_device *dev,
210 				struct drm_file *file)
211 {
212 	struct via_file_private *file_priv = file->driver_priv;
213 	struct via_memblock *entry, *next;
214 
215 	if (!(dev->master && file->master->lock.hw_lock))
216 		return;
217 
218 	drm_legacy_idlelock_take(&file->master->lock);
219 
220 	mutex_lock(&dev->struct_mutex);
221 	if (list_empty(&file_priv->obj_list)) {
222 		mutex_unlock(&dev->struct_mutex);
223 		drm_legacy_idlelock_release(&file->master->lock);
224 
225 		return;
226 	}
227 
228 	via_driver_dma_quiescent(dev);
229 
230 	list_for_each_entry_safe(entry, next, &file_priv->obj_list,
231 				 owner_list) {
232 		list_del(&entry->owner_list);
233 		drm_mm_remove_node(&entry->mm_node);
234 		kfree(entry);
235 	}
236 	mutex_unlock(&dev->struct_mutex);
237 
238 	drm_legacy_idlelock_release(&file->master->lock);
239 
240 	return;
241 }
242