3 * Memory management wrappers for DRM.
5 * \author Rickard E. (Rik) Faith <faith@valinux.com>
6 * \author Gareth Hughes <gareth@valinux.com>
10 * Copyright 1999 Precision Insight, Inc., Cedar Park, Texas.
11 * Copyright 2000 VA Linux Systems, Inc., Sunnyvale, California.
12 * All Rights Reserved.
14 * Permission is hereby granted, free of charge, to any person obtaining a
15 * copy of this software and associated documentation files (the "Software"),
16 * to deal in the Software without restriction, including without limitation
17 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
18 * and/or sell copies of the Software, and to permit persons to whom the
19 * Software is furnished to do so, subject to the following conditions:
21 * The above copyright notice and this permission notice (including the next
22 * paragraph) shall be included in all copies or substantial portions of the
25 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
26 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
27 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
28 * VA LINUX SYSTEMS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, DAMAGES OR
29 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
30 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
31 * OTHER DEALINGS IN THE SOFTWARE.
34 #include <linux/config.h>
37 typedef struct drm_mem_stats {
42 unsigned long bytes_allocated;
43 unsigned long bytes_freed;
46 static spinlock_t DRM(mem_lock) = SPIN_LOCK_UNLOCKED;
47 static unsigned long DRM(ram_available) = 0; /* In pages */
48 static unsigned long DRM(ram_used) = 0;
49 static drm_mem_stats_t DRM(mem_stats)[] = {
50 [DRM_MEM_DMA] = { "dmabufs" },
51 [DRM_MEM_SAREA] = { "sareas" },
52 [DRM_MEM_DRIVER] = { "driver" },
53 [DRM_MEM_MAGIC] = { "magic" },
54 [DRM_MEM_IOCTLS] = { "ioctltab" },
55 [DRM_MEM_MAPS] = { "maplist" },
56 [DRM_MEM_VMAS] = { "vmalist" },
57 [DRM_MEM_BUFS] = { "buflist" },
58 [DRM_MEM_SEGS] = { "seglist" },
59 [DRM_MEM_PAGES] = { "pagelist" },
60 [DRM_MEM_FILES] = { "files" },
61 [DRM_MEM_QUEUES] = { "queues" },
62 [DRM_MEM_CMDS] = { "commands" },
63 [DRM_MEM_MAPPINGS] = { "mappings" },
64 [DRM_MEM_BUFLISTS] = { "buflists" },
65 [DRM_MEM_AGPLISTS] = { "agplist" },
66 [DRM_MEM_SGLISTS] = { "sglist" },
67 [DRM_MEM_TOTALAGP] = { "totalagp" },
68 [DRM_MEM_BOUNDAGP] = { "boundagp" },
69 [DRM_MEM_CTXBITMAP] = { "ctxbitmap"},
70 [DRM_MEM_STUB] = { "stub" },
71 { NULL, 0, } /* Last entry must be null */
74 void DRM(mem_init)(void)
79 for (mem = DRM(mem_stats); mem->name; ++mem) {
80 mem->succeed_count = 0;
83 mem->bytes_allocated = 0;
88 DRM(ram_available) = si.totalram;
92 /* drm_mem_info is called whenever a process reads /dev/drm/mem. */
94 static int DRM(_mem_info)(char *buf, char **start, off_t offset,
95 int request, int *eof, void *data)
100 if (offset > DRM_PROC_LIMIT) {
106 *start = &buf[offset];
108 DRM_PROC_PRINT(" total counts "
109 " | outstanding \n");
110 DRM_PROC_PRINT("type alloc freed fail bytes freed"
111 " | allocs bytes\n\n");
112 DRM_PROC_PRINT("%-9.9s %5d %5d %4d %10lu kB |\n",
114 DRM(ram_available) << (PAGE_SHIFT - 10));
115 DRM_PROC_PRINT("%-9.9s %5d %5d %4d %10lu kB |\n",
116 "locked", 0, 0, 0, DRM(ram_used) >> 10);
117 DRM_PROC_PRINT("\n");
118 for (pt = DRM(mem_stats); pt->name; pt++) {
119 DRM_PROC_PRINT("%-9.9s %5d %5d %4d %10lu %10lu | %6d %10ld\n",
126 pt->succeed_count - pt->free_count,
127 (long)pt->bytes_allocated
128 - (long)pt->bytes_freed);
131 if (len > request + offset) return request;
136 int DRM(mem_info)(char *buf, char **start, off_t offset,
137 int len, int *eof, void *data)
141 spin_lock(&DRM(mem_lock));
142 ret = DRM(_mem_info)(buf, start, offset, len, eof, data);
143 spin_unlock(&DRM(mem_lock));
147 void *DRM(alloc)(size_t size, int area)
152 DRM_MEM_ERROR(area, "Allocating 0 bytes\n");
156 if (!(pt = kmalloc(size, GFP_KERNEL))) {
157 spin_lock(&DRM(mem_lock));
158 ++DRM(mem_stats)[area].fail_count;
159 spin_unlock(&DRM(mem_lock));
162 spin_lock(&DRM(mem_lock));
163 ++DRM(mem_stats)[area].succeed_count;
164 DRM(mem_stats)[area].bytes_allocated += size;
165 spin_unlock(&DRM(mem_lock));
169 void *DRM(calloc)(size_t size, size_t nmemb, int area)
173 addr = DRM(alloc)(nmemb * size, area);
175 memset((void *)addr, 0, size * nmemb);
180 void *DRM(realloc)(void *oldpt, size_t oldsize, size_t size, int area)
184 if (!(pt = DRM(alloc)(size, area))) return NULL;
185 if (oldpt && oldsize) {
186 memcpy(pt, oldpt, oldsize);
187 DRM(free)(oldpt, oldsize, area);
192 void DRM(free)(void *pt, size_t size, int area)
197 if (!pt) DRM_MEM_ERROR(area, "Attempt to free NULL pointer\n");
199 spin_lock(&DRM(mem_lock));
200 DRM(mem_stats)[area].bytes_freed += size;
201 free_count = ++DRM(mem_stats)[area].free_count;
202 alloc_count = DRM(mem_stats)[area].succeed_count;
203 spin_unlock(&DRM(mem_lock));
204 if (free_count > alloc_count) {
205 DRM_MEM_ERROR(area, "Excess frees: %d frees, %d allocs\n",
206 free_count, alloc_count);
210 unsigned long DRM(alloc_pages)(int order, int area)
212 unsigned long address;
213 unsigned long bytes = PAGE_SIZE << order;
217 spin_lock(&DRM(mem_lock));
218 if ((DRM(ram_used) >> PAGE_SHIFT)
219 > (DRM_RAM_PERCENT * DRM(ram_available)) / 100) {
220 spin_unlock(&DRM(mem_lock));
223 spin_unlock(&DRM(mem_lock));
225 address = __get_free_pages(GFP_KERNEL, order);
227 spin_lock(&DRM(mem_lock));
228 ++DRM(mem_stats)[area].fail_count;
229 spin_unlock(&DRM(mem_lock));
232 spin_lock(&DRM(mem_lock));
233 ++DRM(mem_stats)[area].succeed_count;
234 DRM(mem_stats)[area].bytes_allocated += bytes;
235 DRM(ram_used) += bytes;
236 spin_unlock(&DRM(mem_lock));
239 /* Zero outside the lock */
240 memset((void *)address, 0, bytes);
243 for (addr = address, sz = bytes;
245 addr += PAGE_SIZE, sz -= PAGE_SIZE) {
246 SetPageReserved(virt_to_page(addr));
252 void DRM(free_pages)(unsigned long address, int order, int area)
254 unsigned long bytes = PAGE_SIZE << order;
261 DRM_MEM_ERROR(area, "Attempt to free address 0\n");
264 for (addr = address, sz = bytes;
266 addr += PAGE_SIZE, sz -= PAGE_SIZE) {
267 ClearPageReserved(virt_to_page(addr));
269 free_pages(address, order);
272 spin_lock(&DRM(mem_lock));
273 free_count = ++DRM(mem_stats)[area].free_count;
274 alloc_count = DRM(mem_stats)[area].succeed_count;
275 DRM(mem_stats)[area].bytes_freed += bytes;
276 DRM(ram_used) -= bytes;
277 spin_unlock(&DRM(mem_lock));
278 if (free_count > alloc_count) {
280 "Excess frees: %d frees, %d allocs\n",
281 free_count, alloc_count);
285 void *DRM(ioremap)(unsigned long offset, unsigned long size, drm_device_t *dev)
290 DRM_MEM_ERROR(DRM_MEM_MAPPINGS,
291 "Mapping 0 bytes at 0x%08lx\n", offset);
295 if (!(pt = drm_ioremap(offset, size, dev))) {
296 spin_lock(&DRM(mem_lock));
297 ++DRM(mem_stats)[DRM_MEM_MAPPINGS].fail_count;
298 spin_unlock(&DRM(mem_lock));
301 spin_lock(&DRM(mem_lock));
302 ++DRM(mem_stats)[DRM_MEM_MAPPINGS].succeed_count;
303 DRM(mem_stats)[DRM_MEM_MAPPINGS].bytes_allocated += size;
304 spin_unlock(&DRM(mem_lock));
308 void *DRM(ioremap_nocache)(unsigned long offset, unsigned long size, drm_device_t *dev)
313 DRM_MEM_ERROR(DRM_MEM_MAPPINGS,
314 "Mapping 0 bytes at 0x%08lx\n", offset);
318 if (!(pt = drm_ioremap_nocache(offset, size, dev))) {
319 spin_lock(&DRM(mem_lock));
320 ++DRM(mem_stats)[DRM_MEM_MAPPINGS].fail_count;
321 spin_unlock(&DRM(mem_lock));
324 spin_lock(&DRM(mem_lock));
325 ++DRM(mem_stats)[DRM_MEM_MAPPINGS].succeed_count;
326 DRM(mem_stats)[DRM_MEM_MAPPINGS].bytes_allocated += size;
327 spin_unlock(&DRM(mem_lock));
331 void DRM(ioremapfree)(void *pt, unsigned long size, drm_device_t *dev)
337 DRM_MEM_ERROR(DRM_MEM_MAPPINGS,
338 "Attempt to free NULL pointer\n");
340 drm_ioremapfree(pt, size, dev);
342 spin_lock(&DRM(mem_lock));
343 DRM(mem_stats)[DRM_MEM_MAPPINGS].bytes_freed += size;
344 free_count = ++DRM(mem_stats)[DRM_MEM_MAPPINGS].free_count;
345 alloc_count = DRM(mem_stats)[DRM_MEM_MAPPINGS].succeed_count;
346 spin_unlock(&DRM(mem_lock));
347 if (free_count > alloc_count) {
348 DRM_MEM_ERROR(DRM_MEM_MAPPINGS,
349 "Excess frees: %d frees, %d allocs\n",
350 free_count, alloc_count);
354 #if __REALLY_HAVE_AGP
356 DRM_AGP_MEM *DRM(alloc_agp)(int pages, u32 type)
361 DRM_MEM_ERROR(DRM_MEM_TOTALAGP, "Allocating 0 pages\n");
365 if ((handle = DRM(agp_allocate_memory)(pages, type))) {
366 spin_lock(&DRM(mem_lock));
367 ++DRM(mem_stats)[DRM_MEM_TOTALAGP].succeed_count;
368 DRM(mem_stats)[DRM_MEM_TOTALAGP].bytes_allocated
369 += pages << PAGE_SHIFT;
370 spin_unlock(&DRM(mem_lock));
373 spin_lock(&DRM(mem_lock));
374 ++DRM(mem_stats)[DRM_MEM_TOTALAGP].fail_count;
375 spin_unlock(&DRM(mem_lock));
379 int DRM(free_agp)(DRM_AGP_MEM *handle, int pages)
383 int retval = -EINVAL;
386 DRM_MEM_ERROR(DRM_MEM_TOTALAGP,
387 "Attempt to free NULL AGP handle\n");
391 if (DRM(agp_free_memory)(handle)) {
392 spin_lock(&DRM(mem_lock));
393 free_count = ++DRM(mem_stats)[DRM_MEM_TOTALAGP].free_count;
394 alloc_count = DRM(mem_stats)[DRM_MEM_TOTALAGP].succeed_count;
395 DRM(mem_stats)[DRM_MEM_TOTALAGP].bytes_freed
396 += pages << PAGE_SHIFT;
397 spin_unlock(&DRM(mem_lock));
398 if (free_count > alloc_count) {
399 DRM_MEM_ERROR(DRM_MEM_TOTALAGP,
400 "Excess frees: %d frees, %d allocs\n",
401 free_count, alloc_count);
408 int DRM(bind_agp)(DRM_AGP_MEM *handle, unsigned int start)
410 int retcode = -EINVAL;
413 DRM_MEM_ERROR(DRM_MEM_BOUNDAGP,
414 "Attempt to bind NULL AGP handle\n");
418 if (!(retcode = DRM(agp_bind_memory)(handle, start))) {
419 spin_lock(&DRM(mem_lock));
420 ++DRM(mem_stats)[DRM_MEM_BOUNDAGP].succeed_count;
421 DRM(mem_stats)[DRM_MEM_BOUNDAGP].bytes_allocated
422 += handle->page_count << PAGE_SHIFT;
423 spin_unlock(&DRM(mem_lock));
426 spin_lock(&DRM(mem_lock));
427 ++DRM(mem_stats)[DRM_MEM_BOUNDAGP].fail_count;
428 spin_unlock(&DRM(mem_lock));
432 int DRM(unbind_agp)(DRM_AGP_MEM *handle)
436 int retcode = -EINVAL;
439 DRM_MEM_ERROR(DRM_MEM_BOUNDAGP,
440 "Attempt to unbind NULL AGP handle\n");
444 if ((retcode = DRM(agp_unbind_memory)(handle))) return retcode;
445 spin_lock(&DRM(mem_lock));
446 free_count = ++DRM(mem_stats)[DRM_MEM_BOUNDAGP].free_count;
447 alloc_count = DRM(mem_stats)[DRM_MEM_BOUNDAGP].succeed_count;
448 DRM(mem_stats)[DRM_MEM_BOUNDAGP].bytes_freed
449 += handle->page_count << PAGE_SHIFT;
450 spin_unlock(&DRM(mem_lock));
451 if (free_count > alloc_count) {
452 DRM_MEM_ERROR(DRM_MEM_BOUNDAGP,
453 "Excess frees: %d frees, %d allocs\n",
454 free_count, alloc_count);