1 // SPDX-License-Identifier: MIT
2 /*
3 * Copyright © 2021 Intel Corporation
4 */
5
6 #include <linux/kmemleak.h>
7 #include <linux/module.h>
8 #include <linux/sizes.h>
9
10 #include <drm/drm_buddy.h>
11
12 static struct kmem_cache *slab_blocks;
13
drm_block_alloc(struct drm_buddy * mm,struct drm_buddy_block * parent,unsigned int order,u64 offset)14 static struct drm_buddy_block *drm_block_alloc(struct drm_buddy *mm,
15 struct drm_buddy_block *parent,
16 unsigned int order,
17 u64 offset)
18 {
19 struct drm_buddy_block *block;
20
21 BUG_ON(order > DRM_BUDDY_MAX_ORDER);
22
23 block = kmem_cache_zalloc(slab_blocks, GFP_KERNEL);
24 if (!block)
25 return NULL;
26
27 block->header = offset;
28 block->header |= order;
29 block->parent = parent;
30
31 BUG_ON(block->header & DRM_BUDDY_HEADER_UNUSED);
32 return block;
33 }
34
drm_block_free(struct drm_buddy * mm,struct drm_buddy_block * block)35 static void drm_block_free(struct drm_buddy *mm,
36 struct drm_buddy_block *block)
37 {
38 kmem_cache_free(slab_blocks, block);
39 }
40
mark_allocated(struct drm_buddy_block * block)41 static void mark_allocated(struct drm_buddy_block *block)
42 {
43 block->header &= ~DRM_BUDDY_HEADER_STATE;
44 block->header |= DRM_BUDDY_ALLOCATED;
45
46 list_del(&block->link);
47 }
48
mark_free(struct drm_buddy * mm,struct drm_buddy_block * block)49 static void mark_free(struct drm_buddy *mm,
50 struct drm_buddy_block *block)
51 {
52 block->header &= ~DRM_BUDDY_HEADER_STATE;
53 block->header |= DRM_BUDDY_FREE;
54
55 list_add(&block->link,
56 &mm->free_list[drm_buddy_block_order(block)]);
57 }
58
mark_split(struct drm_buddy_block * block)59 static void mark_split(struct drm_buddy_block *block)
60 {
61 block->header &= ~DRM_BUDDY_HEADER_STATE;
62 block->header |= DRM_BUDDY_SPLIT;
63
64 list_del(&block->link);
65 }
66
67 /**
68 * drm_buddy_init - init memory manager
69 *
70 * @mm: DRM buddy manager to initialize
71 * @size: size in bytes to manage
72 * @chunk_size: minimum page size in bytes for our allocations
73 *
74 * Initializes the memory manager and its resources.
75 *
76 * Returns:
77 * 0 on success, error code on failure.
78 */
drm_buddy_init(struct drm_buddy * mm,u64 size,u64 chunk_size)79 int drm_buddy_init(struct drm_buddy *mm, u64 size, u64 chunk_size)
80 {
81 unsigned int i;
82 u64 offset;
83
84 if (size < chunk_size)
85 return -EINVAL;
86
87 if (chunk_size < PAGE_SIZE)
88 return -EINVAL;
89
90 if (!is_power_of_2(chunk_size))
91 return -EINVAL;
92
93 size = round_down(size, chunk_size);
94
95 mm->size = size;
96 mm->avail = size;
97 mm->chunk_size = chunk_size;
98 mm->max_order = ilog2(size) - ilog2(chunk_size);
99
100 BUG_ON(mm->max_order > DRM_BUDDY_MAX_ORDER);
101
102 mm->free_list = kmalloc_array(mm->max_order + 1,
103 sizeof(struct list_head),
104 GFP_KERNEL);
105 if (!mm->free_list)
106 return -ENOMEM;
107
108 for (i = 0; i <= mm->max_order; ++i)
109 INIT_LIST_HEAD(&mm->free_list[i]);
110
111 mm->n_roots = hweight64(size);
112
113 mm->roots = kmalloc_array(mm->n_roots,
114 sizeof(struct drm_buddy_block *),
115 GFP_KERNEL);
116 if (!mm->roots)
117 goto out_free_list;
118
119 offset = 0;
120 i = 0;
121
122 /*
123 * Split into power-of-two blocks, in case we are given a size that is
124 * not itself a power-of-two.
125 */
126 do {
127 struct drm_buddy_block *root;
128 unsigned int order;
129 u64 root_size;
130
131 root_size = rounddown_pow_of_two(size);
132 order = ilog2(root_size) - ilog2(chunk_size);
133
134 root = drm_block_alloc(mm, NULL, order, offset);
135 if (!root)
136 goto out_free_roots;
137
138 mark_free(mm, root);
139
140 BUG_ON(i > mm->max_order);
141 BUG_ON(drm_buddy_block_size(mm, root) < chunk_size);
142
143 mm->roots[i] = root;
144
145 offset += root_size;
146 size -= root_size;
147 i++;
148 } while (size);
149
150 return 0;
151
152 out_free_roots:
153 while (i--)
154 drm_block_free(mm, mm->roots[i]);
155 kfree(mm->roots);
156 out_free_list:
157 kfree(mm->free_list);
158 return -ENOMEM;
159 }
160 EXPORT_SYMBOL(drm_buddy_init);
161
162 /**
163 * drm_buddy_fini - tear down the memory manager
164 *
165 * @mm: DRM buddy manager to free
166 *
167 * Cleanup memory manager resources and the freelist
168 */
drm_buddy_fini(struct drm_buddy * mm)169 void drm_buddy_fini(struct drm_buddy *mm)
170 {
171 int i;
172
173 for (i = 0; i < mm->n_roots; ++i) {
174 WARN_ON(!drm_buddy_block_is_free(mm->roots[i]));
175 drm_block_free(mm, mm->roots[i]);
176 }
177
178 WARN_ON(mm->avail != mm->size);
179
180 kfree(mm->roots);
181 kfree(mm->free_list);
182 }
183 EXPORT_SYMBOL(drm_buddy_fini);
184
split_block(struct drm_buddy * mm,struct drm_buddy_block * block)185 static int split_block(struct drm_buddy *mm,
186 struct drm_buddy_block *block)
187 {
188 unsigned int block_order = drm_buddy_block_order(block) - 1;
189 u64 offset = drm_buddy_block_offset(block);
190
191 BUG_ON(!drm_buddy_block_is_free(block));
192 BUG_ON(!drm_buddy_block_order(block));
193
194 block->left = drm_block_alloc(mm, block, block_order, offset);
195 if (!block->left)
196 return -ENOMEM;
197
198 block->right = drm_block_alloc(mm, block, block_order,
199 offset + (mm->chunk_size << block_order));
200 if (!block->right) {
201 drm_block_free(mm, block->left);
202 return -ENOMEM;
203 }
204
205 mark_free(mm, block->left);
206 mark_free(mm, block->right);
207
208 mark_split(block);
209
210 return 0;
211 }
212
213 static struct drm_buddy_block *
__get_buddy(struct drm_buddy_block * block)214 __get_buddy(struct drm_buddy_block *block)
215 {
216 struct drm_buddy_block *parent;
217
218 parent = block->parent;
219 if (!parent)
220 return NULL;
221
222 if (parent->left == block)
223 return parent->right;
224
225 return parent->left;
226 }
227
228 /**
229 * drm_get_buddy - get buddy address
230 *
231 * @block: DRM buddy block
232 *
233 * Returns the corresponding buddy block for @block, or NULL
234 * if this is a root block and can't be merged further.
235 * Requires some kind of locking to protect against
236 * any concurrent allocate and free operations.
237 */
238 struct drm_buddy_block *
drm_get_buddy(struct drm_buddy_block * block)239 drm_get_buddy(struct drm_buddy_block *block)
240 {
241 return __get_buddy(block);
242 }
243 EXPORT_SYMBOL(drm_get_buddy);
244
__drm_buddy_free(struct drm_buddy * mm,struct drm_buddy_block * block)245 static void __drm_buddy_free(struct drm_buddy *mm,
246 struct drm_buddy_block *block)
247 {
248 struct drm_buddy_block *parent;
249
250 while ((parent = block->parent)) {
251 struct drm_buddy_block *buddy;
252
253 buddy = __get_buddy(block);
254
255 if (!drm_buddy_block_is_free(buddy))
256 break;
257
258 list_del(&buddy->link);
259
260 drm_block_free(mm, block);
261 drm_block_free(mm, buddy);
262
263 block = parent;
264 }
265
266 mark_free(mm, block);
267 }
268
269 /**
270 * drm_buddy_free_block - free a block
271 *
272 * @mm: DRM buddy manager
273 * @block: block to be freed
274 */
drm_buddy_free_block(struct drm_buddy * mm,struct drm_buddy_block * block)275 void drm_buddy_free_block(struct drm_buddy *mm,
276 struct drm_buddy_block *block)
277 {
278 BUG_ON(!drm_buddy_block_is_allocated(block));
279 mm->avail += drm_buddy_block_size(mm, block);
280 __drm_buddy_free(mm, block);
281 }
282 EXPORT_SYMBOL(drm_buddy_free_block);
283
284 /**
285 * drm_buddy_free_list - free blocks
286 *
287 * @mm: DRM buddy manager
288 * @objects: input list head to free blocks
289 */
drm_buddy_free_list(struct drm_buddy * mm,struct list_head * objects)290 void drm_buddy_free_list(struct drm_buddy *mm, struct list_head *objects)
291 {
292 struct drm_buddy_block *block, *on;
293
294 list_for_each_entry_safe(block, on, objects, link) {
295 drm_buddy_free_block(mm, block);
296 cond_resched();
297 }
298 INIT_LIST_HEAD(objects);
299 }
300 EXPORT_SYMBOL(drm_buddy_free_list);
301
overlaps(u64 s1,u64 e1,u64 s2,u64 e2)302 static inline bool overlaps(u64 s1, u64 e1, u64 s2, u64 e2)
303 {
304 return s1 <= e2 && e1 >= s2;
305 }
306
contains(u64 s1,u64 e1,u64 s2,u64 e2)307 static inline bool contains(u64 s1, u64 e1, u64 s2, u64 e2)
308 {
309 return s1 <= s2 && e1 >= e2;
310 }
311
312 static struct drm_buddy_block *
alloc_range_bias(struct drm_buddy * mm,u64 start,u64 end,unsigned int order)313 alloc_range_bias(struct drm_buddy *mm,
314 u64 start, u64 end,
315 unsigned int order)
316 {
317 struct drm_buddy_block *block;
318 struct drm_buddy_block *buddy;
319 LIST_HEAD(dfs);
320 int err;
321 int i;
322
323 end = end - 1;
324
325 for (i = 0; i < mm->n_roots; ++i)
326 list_add_tail(&mm->roots[i]->tmp_link, &dfs);
327
328 do {
329 u64 block_start;
330 u64 block_end;
331
332 block = list_first_entry_or_null(&dfs,
333 struct drm_buddy_block,
334 tmp_link);
335 if (!block)
336 break;
337
338 list_del(&block->tmp_link);
339
340 if (drm_buddy_block_order(block) < order)
341 continue;
342
343 block_start = drm_buddy_block_offset(block);
344 block_end = block_start + drm_buddy_block_size(mm, block) - 1;
345
346 if (!overlaps(start, end, block_start, block_end))
347 continue;
348
349 if (drm_buddy_block_is_allocated(block))
350 continue;
351
352 if (contains(start, end, block_start, block_end) &&
353 order == drm_buddy_block_order(block)) {
354 /*
355 * Find the free block within the range.
356 */
357 if (drm_buddy_block_is_free(block))
358 return block;
359
360 continue;
361 }
362
363 if (!drm_buddy_block_is_split(block)) {
364 err = split_block(mm, block);
365 if (unlikely(err))
366 goto err_undo;
367 }
368
369 list_add(&block->right->tmp_link, &dfs);
370 list_add(&block->left->tmp_link, &dfs);
371 } while (1);
372
373 return ERR_PTR(-ENOSPC);
374
375 err_undo:
376 /*
377 * We really don't want to leave around a bunch of split blocks, since
378 * bigger is better, so make sure we merge everything back before we
379 * free the allocated blocks.
380 */
381 buddy = __get_buddy(block);
382 if (buddy &&
383 (drm_buddy_block_is_free(block) &&
384 drm_buddy_block_is_free(buddy)))
385 __drm_buddy_free(mm, block);
386 return ERR_PTR(err);
387 }
388
389 static struct drm_buddy_block *
get_maxblock(struct list_head * head)390 get_maxblock(struct list_head *head)
391 {
392 struct drm_buddy_block *max_block = NULL, *node;
393
394 max_block = list_first_entry_or_null(head,
395 struct drm_buddy_block,
396 link);
397 if (!max_block)
398 return NULL;
399
400 list_for_each_entry(node, head, link) {
401 if (drm_buddy_block_offset(node) >
402 drm_buddy_block_offset(max_block))
403 max_block = node;
404 }
405
406 return max_block;
407 }
408
409 static struct drm_buddy_block *
alloc_from_freelist(struct drm_buddy * mm,unsigned int order,unsigned long flags)410 alloc_from_freelist(struct drm_buddy *mm,
411 unsigned int order,
412 unsigned long flags)
413 {
414 struct drm_buddy_block *block = NULL;
415 unsigned int i;
416 int err;
417
418 for (i = order; i <= mm->max_order; ++i) {
419 if (flags & DRM_BUDDY_TOPDOWN_ALLOCATION) {
420 block = get_maxblock(&mm->free_list[i]);
421 if (block)
422 break;
423 } else {
424 block = list_first_entry_or_null(&mm->free_list[i],
425 struct drm_buddy_block,
426 link);
427 if (block)
428 break;
429 }
430 }
431
432 if (!block)
433 return ERR_PTR(-ENOSPC);
434
435 BUG_ON(!drm_buddy_block_is_free(block));
436
437 while (i != order) {
438 err = split_block(mm, block);
439 if (unlikely(err))
440 goto err_undo;
441
442 block = block->right;
443 i--;
444 }
445 return block;
446
447 err_undo:
448 if (i != order)
449 __drm_buddy_free(mm, block);
450 return ERR_PTR(err);
451 }
452
__alloc_range(struct drm_buddy * mm,struct list_head * dfs,u64 start,u64 size,struct list_head * blocks)453 static int __alloc_range(struct drm_buddy *mm,
454 struct list_head *dfs,
455 u64 start, u64 size,
456 struct list_head *blocks)
457 {
458 struct drm_buddy_block *block;
459 struct drm_buddy_block *buddy;
460 LIST_HEAD(allocated);
461 u64 end;
462 int err;
463
464 end = start + size - 1;
465
466 do {
467 u64 block_start;
468 u64 block_end;
469
470 block = list_first_entry_or_null(dfs,
471 struct drm_buddy_block,
472 tmp_link);
473 if (!block)
474 break;
475
476 list_del(&block->tmp_link);
477
478 block_start = drm_buddy_block_offset(block);
479 block_end = block_start + drm_buddy_block_size(mm, block) - 1;
480
481 if (!overlaps(start, end, block_start, block_end))
482 continue;
483
484 if (drm_buddy_block_is_allocated(block)) {
485 err = -ENOSPC;
486 goto err_free;
487 }
488
489 if (contains(start, end, block_start, block_end)) {
490 if (!drm_buddy_block_is_free(block)) {
491 err = -ENOSPC;
492 goto err_free;
493 }
494
495 mark_allocated(block);
496 mm->avail -= drm_buddy_block_size(mm, block);
497 list_add_tail(&block->link, &allocated);
498 continue;
499 }
500
501 if (!drm_buddy_block_is_split(block)) {
502 err = split_block(mm, block);
503 if (unlikely(err))
504 goto err_undo;
505 }
506
507 list_add(&block->right->tmp_link, dfs);
508 list_add(&block->left->tmp_link, dfs);
509 } while (1);
510
511 list_splice_tail(&allocated, blocks);
512 return 0;
513
514 err_undo:
515 /*
516 * We really don't want to leave around a bunch of split blocks, since
517 * bigger is better, so make sure we merge everything back before we
518 * free the allocated blocks.
519 */
520 buddy = __get_buddy(block);
521 if (buddy &&
522 (drm_buddy_block_is_free(block) &&
523 drm_buddy_block_is_free(buddy)))
524 __drm_buddy_free(mm, block);
525
526 err_free:
527 drm_buddy_free_list(mm, &allocated);
528 return err;
529 }
530
__drm_buddy_alloc_range(struct drm_buddy * mm,u64 start,u64 size,struct list_head * blocks)531 static int __drm_buddy_alloc_range(struct drm_buddy *mm,
532 u64 start,
533 u64 size,
534 struct list_head *blocks)
535 {
536 LIST_HEAD(dfs);
537 int i;
538
539 for (i = 0; i < mm->n_roots; ++i)
540 list_add_tail(&mm->roots[i]->tmp_link, &dfs);
541
542 return __alloc_range(mm, &dfs, start, size, blocks);
543 }
544
545 /**
546 * drm_buddy_block_trim - free unused pages
547 *
548 * @mm: DRM buddy manager
549 * @new_size: original size requested
550 * @blocks: Input and output list of allocated blocks.
551 * MUST contain single block as input to be trimmed.
552 * On success will contain the newly allocated blocks
553 * making up the @new_size. Blocks always appear in
554 * ascending order
555 *
556 * For contiguous allocation, we round up the size to the nearest
557 * power of two value, drivers consume *actual* size, so remaining
558 * portions are unused and can be optionally freed with this function
559 *
560 * Returns:
561 * 0 on success, error code on failure.
562 */
drm_buddy_block_trim(struct drm_buddy * mm,u64 new_size,struct list_head * blocks)563 int drm_buddy_block_trim(struct drm_buddy *mm,
564 u64 new_size,
565 struct list_head *blocks)
566 {
567 struct drm_buddy_block *parent;
568 struct drm_buddy_block *block;
569 LIST_HEAD(dfs);
570 u64 new_start;
571 int err;
572
573 if (!list_is_singular(blocks))
574 return -EINVAL;
575
576 block = list_first_entry(blocks,
577 struct drm_buddy_block,
578 link);
579
580 if (WARN_ON(!drm_buddy_block_is_allocated(block)))
581 return -EINVAL;
582
583 if (new_size > drm_buddy_block_size(mm, block))
584 return -EINVAL;
585
586 if (!new_size || !IS_ALIGNED(new_size, mm->chunk_size))
587 return -EINVAL;
588
589 if (new_size == drm_buddy_block_size(mm, block))
590 return 0;
591
592 list_del(&block->link);
593 mark_free(mm, block);
594 mm->avail += drm_buddy_block_size(mm, block);
595
596 /* Prevent recursively freeing this node */
597 parent = block->parent;
598 block->parent = NULL;
599
600 new_start = drm_buddy_block_offset(block);
601 list_add(&block->tmp_link, &dfs);
602 err = __alloc_range(mm, &dfs, new_start, new_size, blocks);
603 if (err) {
604 mark_allocated(block);
605 mm->avail -= drm_buddy_block_size(mm, block);
606 list_add(&block->link, blocks);
607 }
608
609 block->parent = parent;
610 return err;
611 }
612 EXPORT_SYMBOL(drm_buddy_block_trim);
613
614 /**
615 * drm_buddy_alloc_blocks - allocate power-of-two blocks
616 *
617 * @mm: DRM buddy manager to allocate from
618 * @start: start of the allowed range for this block
619 * @end: end of the allowed range for this block
620 * @size: size of the allocation
621 * @min_page_size: alignment of the allocation
622 * @blocks: output list head to add allocated blocks
623 * @flags: DRM_BUDDY_*_ALLOCATION flags
624 *
625 * alloc_range_bias() called on range limitations, which traverses
626 * the tree and returns the desired block.
627 *
628 * alloc_from_freelist() called when *no* range restrictions
629 * are enforced, which picks the block from the freelist.
630 *
631 * Returns:
632 * 0 on success, error code on failure.
633 */
drm_buddy_alloc_blocks(struct drm_buddy * mm,u64 start,u64 end,u64 size,u64 min_page_size,struct list_head * blocks,unsigned long flags)634 int drm_buddy_alloc_blocks(struct drm_buddy *mm,
635 u64 start, u64 end, u64 size,
636 u64 min_page_size,
637 struct list_head *blocks,
638 unsigned long flags)
639 {
640 struct drm_buddy_block *block = NULL;
641 unsigned int min_order, order;
642 unsigned long pages;
643 LIST_HEAD(allocated);
644 int err;
645
646 if (size < mm->chunk_size)
647 return -EINVAL;
648
649 if (min_page_size < mm->chunk_size)
650 return -EINVAL;
651
652 if (!is_power_of_2(min_page_size))
653 return -EINVAL;
654
655 if (!IS_ALIGNED(start | end | size, mm->chunk_size))
656 return -EINVAL;
657
658 if (end > mm->size)
659 return -EINVAL;
660
661 if (range_overflows(start, size, mm->size))
662 return -EINVAL;
663
664 /* Actual range allocation */
665 if (start + size == end)
666 return __drm_buddy_alloc_range(mm, start, size, blocks);
667
668 if (!IS_ALIGNED(size, min_page_size))
669 return -EINVAL;
670
671 pages = size >> ilog2(mm->chunk_size);
672 order = fls(pages) - 1;
673 min_order = ilog2(min_page_size) - ilog2(mm->chunk_size);
674
675 do {
676 order = min(order, (unsigned int)fls(pages) - 1);
677 BUG_ON(order > mm->max_order);
678 BUG_ON(order < min_order);
679
680 do {
681 if (flags & DRM_BUDDY_RANGE_ALLOCATION)
682 /* Allocate traversing within the range */
683 block = alloc_range_bias(mm, start, end, order);
684 else
685 /* Allocate from freelist */
686 block = alloc_from_freelist(mm, order, flags);
687
688 if (!IS_ERR(block))
689 break;
690
691 if (order-- == min_order) {
692 err = -ENOSPC;
693 goto err_free;
694 }
695 } while (1);
696
697 mark_allocated(block);
698 mm->avail -= drm_buddy_block_size(mm, block);
699 kmemleak_update_trace(block);
700 list_add_tail(&block->link, &allocated);
701
702 pages -= BIT(order);
703
704 if (!pages)
705 break;
706 } while (1);
707
708 list_splice_tail(&allocated, blocks);
709 return 0;
710
711 err_free:
712 drm_buddy_free_list(mm, &allocated);
713 return err;
714 }
715 EXPORT_SYMBOL(drm_buddy_alloc_blocks);
716
717 /**
718 * drm_buddy_block_print - print block information
719 *
720 * @mm: DRM buddy manager
721 * @block: DRM buddy block
722 * @p: DRM printer to use
723 */
drm_buddy_block_print(struct drm_buddy * mm,struct drm_buddy_block * block,struct drm_printer * p)724 void drm_buddy_block_print(struct drm_buddy *mm,
725 struct drm_buddy_block *block,
726 struct drm_printer *p)
727 {
728 u64 start = drm_buddy_block_offset(block);
729 u64 size = drm_buddy_block_size(mm, block);
730
731 drm_printf(p, "%#018llx-%#018llx: %llu\n", start, start + size, size);
732 }
733 EXPORT_SYMBOL(drm_buddy_block_print);
734
735 /**
736 * drm_buddy_print - print allocator state
737 *
738 * @mm: DRM buddy manager
739 * @p: DRM printer to use
740 */
drm_buddy_print(struct drm_buddy * mm,struct drm_printer * p)741 void drm_buddy_print(struct drm_buddy *mm, struct drm_printer *p)
742 {
743 int order;
744
745 drm_printf(p, "chunk_size: %lluKiB, total: %lluMiB, free: %lluMiB\n",
746 mm->chunk_size >> 10, mm->size >> 20, mm->avail >> 20);
747
748 for (order = mm->max_order; order >= 0; order--) {
749 struct drm_buddy_block *block;
750 u64 count = 0, free;
751
752 list_for_each_entry(block, &mm->free_list[order], link) {
753 BUG_ON(!drm_buddy_block_is_free(block));
754 count++;
755 }
756
757 drm_printf(p, "order-%d ", order);
758
759 free = count * (mm->chunk_size << order);
760 if (free < SZ_1M)
761 drm_printf(p, "free: %lluKiB", free >> 10);
762 else
763 drm_printf(p, "free: %lluMiB", free >> 20);
764
765 drm_printf(p, ", pages: %llu\n", count);
766 }
767 }
768 EXPORT_SYMBOL(drm_buddy_print);
769
drm_buddy_module_exit(void)770 static void drm_buddy_module_exit(void)
771 {
772 kmem_cache_destroy(slab_blocks);
773 }
774
drm_buddy_module_init(void)775 static int __init drm_buddy_module_init(void)
776 {
777 slab_blocks = KMEM_CACHE(drm_buddy_block, 0);
778 if (!slab_blocks)
779 return -ENOMEM;
780
781 return 0;
782 }
783
784 module_init(drm_buddy_module_init);
785 module_exit(drm_buddy_module_exit);
786
787 MODULE_DESCRIPTION("DRM Buddy Allocator");
788 MODULE_LICENSE("Dual MIT/GPL");
789