1 // SPDX-License-Identifier: MIT
3 * Copyright © 2021 Intel Corporation
6 #include <linux/kmemleak.h>
7 #include <linux/module.h>
8 #include <linux/sizes.h>
10 #include <drm/drm_buddy.h>
12 static struct kmem_cache *slab_blocks;
14 static struct drm_buddy_block *drm_block_alloc(struct drm_buddy *mm,
15 struct drm_buddy_block *parent,
19 struct drm_buddy_block *block;
21 BUG_ON(order > DRM_BUDDY_MAX_ORDER);
23 block = kmem_cache_zalloc(slab_blocks, GFP_KERNEL);
27 block->header = offset;
28 block->header |= order;
29 block->parent = parent;
31 BUG_ON(block->header & DRM_BUDDY_HEADER_UNUSED);
35 static void drm_block_free(struct drm_buddy *mm,
36 struct drm_buddy_block *block)
38 kmem_cache_free(slab_blocks, block);
41 static void list_insert_sorted(struct drm_buddy *mm,
42 struct drm_buddy_block *block)
44 struct drm_buddy_block *node;
45 struct list_head *head;
47 head = &mm->free_list[drm_buddy_block_order(block)];
48 if (list_empty(head)) {
49 list_add(&block->link, head);
53 list_for_each_entry(node, head, link)
54 if (drm_buddy_block_offset(block) < drm_buddy_block_offset(node))
57 __list_add(&block->link, node->link.prev, &node->link);
60 static void clear_reset(struct drm_buddy_block *block)
62 block->header &= ~DRM_BUDDY_HEADER_CLEAR;
65 static void mark_cleared(struct drm_buddy_block *block)
67 block->header |= DRM_BUDDY_HEADER_CLEAR;
70 static void mark_allocated(struct drm_buddy_block *block)
72 block->header &= ~DRM_BUDDY_HEADER_STATE;
73 block->header |= DRM_BUDDY_ALLOCATED;
75 list_del(&block->link);
78 static void mark_free(struct drm_buddy *mm,
79 struct drm_buddy_block *block)
81 block->header &= ~DRM_BUDDY_HEADER_STATE;
82 block->header |= DRM_BUDDY_FREE;
84 list_insert_sorted(mm, block);
87 static void mark_split(struct drm_buddy_block *block)
89 block->header &= ~DRM_BUDDY_HEADER_STATE;
90 block->header |= DRM_BUDDY_SPLIT;
92 list_del(&block->link);
95 static inline bool overlaps(u64 s1, u64 e1, u64 s2, u64 e2)
97 return s1 <= e2 && e1 >= s2;
100 static inline bool contains(u64 s1, u64 e1, u64 s2, u64 e2)
102 return s1 <= s2 && e1 >= e2;
105 static struct drm_buddy_block *
106 __get_buddy(struct drm_buddy_block *block)
108 struct drm_buddy_block *parent;
110 parent = block->parent;
114 if (parent->left == block)
115 return parent->right;
120 static unsigned int __drm_buddy_free(struct drm_buddy *mm,
121 struct drm_buddy_block *block,
124 struct drm_buddy_block *parent;
127 while ((parent = block->parent)) {
128 struct drm_buddy_block *buddy;
130 buddy = __get_buddy(block);
132 if (!drm_buddy_block_is_free(buddy))
137 * Check the block and its buddy clear state and exit
138 * the loop if they both have the dissimilar state.
140 if (drm_buddy_block_is_clear(block) !=
141 drm_buddy_block_is_clear(buddy))
144 if (drm_buddy_block_is_clear(block))
145 mark_cleared(parent);
148 list_del(&buddy->link);
149 if (force_merge && drm_buddy_block_is_clear(buddy))
150 mm->clear_avail -= drm_buddy_block_size(mm, buddy);
152 drm_block_free(mm, block);
153 drm_block_free(mm, buddy);
158 order = drm_buddy_block_order(block);
159 mark_free(mm, block);
164 static int __force_merge(struct drm_buddy *mm,
167 unsigned int min_order)
175 if (min_order > mm->max_order)
178 for (i = min_order - 1; i >= 0; i--) {
179 struct drm_buddy_block *block, *prev;
181 list_for_each_entry_safe_reverse(block, prev, &mm->free_list[i], link) {
182 struct drm_buddy_block *buddy;
183 u64 block_start, block_end;
188 block_start = drm_buddy_block_offset(block);
189 block_end = block_start + drm_buddy_block_size(mm, block) - 1;
191 if (!contains(start, end, block_start, block_end))
194 buddy = __get_buddy(block);
195 if (!drm_buddy_block_is_free(buddy))
198 WARN_ON(drm_buddy_block_is_clear(block) ==
199 drm_buddy_block_is_clear(buddy));
202 * If the prev block is same as buddy, don't access the
203 * block in the next iteration as we would free the
204 * buddy block as part of the free function.
207 prev = list_prev_entry(prev, link);
209 list_del(&block->link);
210 if (drm_buddy_block_is_clear(block))
211 mm->clear_avail -= drm_buddy_block_size(mm, block);
213 order = __drm_buddy_free(mm, block, true);
214 if (order >= min_order)
223 * drm_buddy_init - init memory manager
225 * @mm: DRM buddy manager to initialize
226 * @size: size in bytes to manage
227 * @chunk_size: minimum page size in bytes for our allocations
229 * Initializes the memory manager and its resources.
232 * 0 on success, error code on failure.
234 int drm_buddy_init(struct drm_buddy *mm, u64 size, u64 chunk_size)
239 if (size < chunk_size)
242 if (chunk_size < PAGE_SIZE)
245 if (!is_power_of_2(chunk_size))
248 size = round_down(size, chunk_size);
252 mm->chunk_size = chunk_size;
253 mm->max_order = ilog2(size) - ilog2(chunk_size);
255 BUG_ON(mm->max_order > DRM_BUDDY_MAX_ORDER);
257 mm->free_list = kmalloc_array(mm->max_order + 1,
258 sizeof(struct list_head),
263 for (i = 0; i <= mm->max_order; ++i)
264 INIT_LIST_HEAD(&mm->free_list[i]);
266 mm->n_roots = hweight64(size);
268 mm->roots = kmalloc_array(mm->n_roots,
269 sizeof(struct drm_buddy_block *),
278 * Split into power-of-two blocks, in case we are given a size that is
279 * not itself a power-of-two.
282 struct drm_buddy_block *root;
286 order = ilog2(size) - ilog2(chunk_size);
287 root_size = chunk_size << order;
289 root = drm_block_alloc(mm, NULL, order, offset);
295 BUG_ON(i > mm->max_order);
296 BUG_ON(drm_buddy_block_size(mm, root) < chunk_size);
309 drm_block_free(mm, mm->roots[i]);
312 kfree(mm->free_list);
315 EXPORT_SYMBOL(drm_buddy_init);
318 * drm_buddy_fini - tear down the memory manager
320 * @mm: DRM buddy manager to free
322 * Cleanup memory manager resources and the freelist
324 void drm_buddy_fini(struct drm_buddy *mm)
332 for (i = 0; i < mm->n_roots; ++i) {
333 order = ilog2(size) - ilog2(mm->chunk_size);
334 __force_merge(mm, 0, size, order);
336 WARN_ON(!drm_buddy_block_is_free(mm->roots[i]));
337 drm_block_free(mm, mm->roots[i]);
339 root_size = mm->chunk_size << order;
343 WARN_ON(mm->avail != mm->size);
346 kfree(mm->free_list);
348 EXPORT_SYMBOL(drm_buddy_fini);
350 static int split_block(struct drm_buddy *mm,
351 struct drm_buddy_block *block)
353 unsigned int block_order = drm_buddy_block_order(block) - 1;
354 u64 offset = drm_buddy_block_offset(block);
356 BUG_ON(!drm_buddy_block_is_free(block));
357 BUG_ON(!drm_buddy_block_order(block));
359 block->left = drm_block_alloc(mm, block, block_order, offset);
363 block->right = drm_block_alloc(mm, block, block_order,
364 offset + (mm->chunk_size << block_order));
366 drm_block_free(mm, block->left);
370 mark_free(mm, block->left);
371 mark_free(mm, block->right);
373 if (drm_buddy_block_is_clear(block)) {
374 mark_cleared(block->left);
375 mark_cleared(block->right);
385 * drm_get_buddy - get buddy address
387 * @block: DRM buddy block
389 * Returns the corresponding buddy block for @block, or NULL
390 * if this is a root block and can't be merged further.
391 * Requires some kind of locking to protect against
392 * any concurrent allocate and free operations.
394 struct drm_buddy_block *
395 drm_get_buddy(struct drm_buddy_block *block)
397 return __get_buddy(block);
399 EXPORT_SYMBOL(drm_get_buddy);
402 * drm_buddy_free_block - free a block
404 * @mm: DRM buddy manager
405 * @block: block to be freed
407 void drm_buddy_free_block(struct drm_buddy *mm,
408 struct drm_buddy_block *block)
410 BUG_ON(!drm_buddy_block_is_allocated(block));
411 mm->avail += drm_buddy_block_size(mm, block);
412 if (drm_buddy_block_is_clear(block))
413 mm->clear_avail += drm_buddy_block_size(mm, block);
415 __drm_buddy_free(mm, block, false);
417 EXPORT_SYMBOL(drm_buddy_free_block);
419 static void __drm_buddy_free_list(struct drm_buddy *mm,
420 struct list_head *objects,
424 struct drm_buddy_block *block, *on;
426 WARN_ON(mark_dirty && mark_clear);
428 list_for_each_entry_safe(block, on, objects, link) {
433 drm_buddy_free_block(mm, block);
436 INIT_LIST_HEAD(objects);
439 static void drm_buddy_free_list_internal(struct drm_buddy *mm,
440 struct list_head *objects)
443 * Don't touch the clear/dirty bit, since allocation is still internal
444 * at this point. For example we might have just failed part of the
447 __drm_buddy_free_list(mm, objects, false, false);
451 * drm_buddy_free_list - free blocks
453 * @mm: DRM buddy manager
454 * @objects: input list head to free blocks
455 * @flags: optional flags like DRM_BUDDY_CLEARED
457 void drm_buddy_free_list(struct drm_buddy *mm,
458 struct list_head *objects,
461 bool mark_clear = flags & DRM_BUDDY_CLEARED;
463 __drm_buddy_free_list(mm, objects, mark_clear, !mark_clear);
465 EXPORT_SYMBOL(drm_buddy_free_list);
467 static bool block_incompatible(struct drm_buddy_block *block, unsigned int flags)
469 bool needs_clear = flags & DRM_BUDDY_CLEAR_ALLOCATION;
471 return needs_clear != drm_buddy_block_is_clear(block);
474 static struct drm_buddy_block *
475 __alloc_range_bias(struct drm_buddy *mm,
481 u64 req_size = mm->chunk_size << order;
482 struct drm_buddy_block *block;
483 struct drm_buddy_block *buddy;
490 for (i = 0; i < mm->n_roots; ++i)
491 list_add_tail(&mm->roots[i]->tmp_link, &dfs);
497 block = list_first_entry_or_null(&dfs,
498 struct drm_buddy_block,
503 list_del(&block->tmp_link);
505 if (drm_buddy_block_order(block) < order)
508 block_start = drm_buddy_block_offset(block);
509 block_end = block_start + drm_buddy_block_size(mm, block) - 1;
511 if (!overlaps(start, end, block_start, block_end))
514 if (drm_buddy_block_is_allocated(block))
517 if (block_start < start || block_end > end) {
518 u64 adjusted_start = max(block_start, start);
519 u64 adjusted_end = min(block_end, end);
521 if (round_down(adjusted_end + 1, req_size) <=
522 round_up(adjusted_start, req_size))
526 if (contains(start, end, block_start, block_end) &&
527 order == drm_buddy_block_order(block)) {
528 if (!fallback && block_incompatible(block, flags))
532 * Find the free block within the range.
534 if (drm_buddy_block_is_free(block))
540 if (!drm_buddy_block_is_split(block)) {
541 err = split_block(mm, block);
546 list_add(&block->right->tmp_link, &dfs);
547 list_add(&block->left->tmp_link, &dfs);
550 return ERR_PTR(-ENOSPC);
554 * We really don't want to leave around a bunch of split blocks, since
555 * bigger is better, so make sure we merge everything back before we
556 * free the allocated blocks.
558 buddy = __get_buddy(block);
560 (drm_buddy_block_is_free(block) &&
561 drm_buddy_block_is_free(buddy)))
562 __drm_buddy_free(mm, block, false);
566 static struct drm_buddy_block *
567 __drm_buddy_alloc_range_bias(struct drm_buddy *mm,
572 struct drm_buddy_block *block;
573 bool fallback = false;
575 block = __alloc_range_bias(mm, start, end, order,
577 if (IS_ERR(block) && mm->clear_avail)
578 return __alloc_range_bias(mm, start, end, order,
584 static struct drm_buddy_block *
585 get_maxblock(struct drm_buddy *mm, unsigned int order,
588 struct drm_buddy_block *max_block = NULL, *block = NULL;
591 for (i = order; i <= mm->max_order; ++i) {
592 struct drm_buddy_block *tmp_block;
594 list_for_each_entry_reverse(tmp_block, &mm->free_list[i], link) {
595 if (block_incompatible(tmp_block, flags))
610 if (drm_buddy_block_offset(block) >
611 drm_buddy_block_offset(max_block)) {
619 static struct drm_buddy_block *
620 alloc_from_freelist(struct drm_buddy *mm,
624 struct drm_buddy_block *block = NULL;
628 if (flags & DRM_BUDDY_TOPDOWN_ALLOCATION) {
629 block = get_maxblock(mm, order, flags);
631 /* Store the obtained block order */
632 tmp = drm_buddy_block_order(block);
634 for (tmp = order; tmp <= mm->max_order; ++tmp) {
635 struct drm_buddy_block *tmp_block;
637 list_for_each_entry_reverse(tmp_block, &mm->free_list[tmp], link) {
638 if (block_incompatible(tmp_block, flags))
651 /* Fallback method */
652 for (tmp = order; tmp <= mm->max_order; ++tmp) {
653 if (!list_empty(&mm->free_list[tmp])) {
654 block = list_last_entry(&mm->free_list[tmp],
655 struct drm_buddy_block,
663 return ERR_PTR(-ENOSPC);
666 BUG_ON(!drm_buddy_block_is_free(block));
668 while (tmp != order) {
669 err = split_block(mm, block);
673 block = block->right;
680 __drm_buddy_free(mm, block, false);
684 static int __alloc_range(struct drm_buddy *mm,
685 struct list_head *dfs,
687 struct list_head *blocks,
688 u64 *total_allocated_on_err)
690 struct drm_buddy_block *block;
691 struct drm_buddy_block *buddy;
692 u64 total_allocated = 0;
693 LIST_HEAD(allocated);
697 end = start + size - 1;
703 block = list_first_entry_or_null(dfs,
704 struct drm_buddy_block,
709 list_del(&block->tmp_link);
711 block_start = drm_buddy_block_offset(block);
712 block_end = block_start + drm_buddy_block_size(mm, block) - 1;
714 if (!overlaps(start, end, block_start, block_end))
717 if (drm_buddy_block_is_allocated(block)) {
722 if (contains(start, end, block_start, block_end)) {
723 if (drm_buddy_block_is_free(block)) {
724 mark_allocated(block);
725 total_allocated += drm_buddy_block_size(mm, block);
726 mm->avail -= drm_buddy_block_size(mm, block);
727 if (drm_buddy_block_is_clear(block))
728 mm->clear_avail -= drm_buddy_block_size(mm, block);
729 list_add_tail(&block->link, &allocated);
731 } else if (!mm->clear_avail) {
737 if (!drm_buddy_block_is_split(block)) {
738 err = split_block(mm, block);
743 list_add(&block->right->tmp_link, dfs);
744 list_add(&block->left->tmp_link, dfs);
747 if (total_allocated < size) {
752 list_splice_tail(&allocated, blocks);
758 * We really don't want to leave around a bunch of split blocks, since
759 * bigger is better, so make sure we merge everything back before we
760 * free the allocated blocks.
762 buddy = __get_buddy(block);
764 (drm_buddy_block_is_free(block) &&
765 drm_buddy_block_is_free(buddy)))
766 __drm_buddy_free(mm, block, false);
769 if (err == -ENOSPC && total_allocated_on_err) {
770 list_splice_tail(&allocated, blocks);
771 *total_allocated_on_err = total_allocated;
773 drm_buddy_free_list_internal(mm, &allocated);
779 static int __drm_buddy_alloc_range(struct drm_buddy *mm,
782 u64 *total_allocated_on_err,
783 struct list_head *blocks)
788 for (i = 0; i < mm->n_roots; ++i)
789 list_add_tail(&mm->roots[i]->tmp_link, &dfs);
791 return __alloc_range(mm, &dfs, start, size,
792 blocks, total_allocated_on_err);
795 static int __alloc_contig_try_harder(struct drm_buddy *mm,
798 struct list_head *blocks)
800 u64 rhs_offset, lhs_offset, lhs_size, filled;
801 struct drm_buddy_block *block;
802 struct list_head *list;
803 LIST_HEAD(blocks_lhs);
809 modify_size = rounddown_pow_of_two(size);
810 pages = modify_size >> ilog2(mm->chunk_size);
811 order = fls(pages) - 1;
815 list = &mm->free_list[order];
816 if (list_empty(list))
819 list_for_each_entry_reverse(block, list, link) {
820 /* Allocate blocks traversing RHS */
821 rhs_offset = drm_buddy_block_offset(block);
822 err = __drm_buddy_alloc_range(mm, rhs_offset, size,
824 if (!err || err != -ENOSPC)
827 lhs_size = max((size - filled), min_block_size);
828 if (!IS_ALIGNED(lhs_size, min_block_size))
829 lhs_size = round_up(lhs_size, min_block_size);
831 /* Allocate blocks traversing LHS */
832 lhs_offset = drm_buddy_block_offset(block) - lhs_size;
833 err = __drm_buddy_alloc_range(mm, lhs_offset, lhs_size,
836 list_splice(&blocks_lhs, blocks);
838 } else if (err != -ENOSPC) {
839 drm_buddy_free_list_internal(mm, blocks);
842 /* Free blocks for the next iteration */
843 drm_buddy_free_list_internal(mm, blocks);
850 * drm_buddy_block_trim - free unused pages
852 * @mm: DRM buddy manager
853 * @new_size: original size requested
854 * @blocks: Input and output list of allocated blocks.
855 * MUST contain single block as input to be trimmed.
856 * On success will contain the newly allocated blocks
857 * making up the @new_size. Blocks always appear in
860 * For contiguous allocation, we round up the size to the nearest
861 * power of two value, drivers consume *actual* size, so remaining
862 * portions are unused and can be optionally freed with this function
865 * 0 on success, error code on failure.
867 int drm_buddy_block_trim(struct drm_buddy *mm,
869 struct list_head *blocks)
871 struct drm_buddy_block *parent;
872 struct drm_buddy_block *block;
877 if (!list_is_singular(blocks))
880 block = list_first_entry(blocks,
881 struct drm_buddy_block,
884 if (WARN_ON(!drm_buddy_block_is_allocated(block)))
887 if (new_size > drm_buddy_block_size(mm, block))
890 if (!new_size || !IS_ALIGNED(new_size, mm->chunk_size))
893 if (new_size == drm_buddy_block_size(mm, block))
896 list_del(&block->link);
897 mark_free(mm, block);
898 mm->avail += drm_buddy_block_size(mm, block);
899 if (drm_buddy_block_is_clear(block))
900 mm->clear_avail += drm_buddy_block_size(mm, block);
902 /* Prevent recursively freeing this node */
903 parent = block->parent;
904 block->parent = NULL;
906 new_start = drm_buddy_block_offset(block);
907 list_add(&block->tmp_link, &dfs);
908 err = __alloc_range(mm, &dfs, new_start, new_size, blocks, NULL);
910 mark_allocated(block);
911 mm->avail -= drm_buddy_block_size(mm, block);
912 if (drm_buddy_block_is_clear(block))
913 mm->clear_avail -= drm_buddy_block_size(mm, block);
914 list_add(&block->link, blocks);
917 block->parent = parent;
920 EXPORT_SYMBOL(drm_buddy_block_trim);
922 static struct drm_buddy_block *
923 __drm_buddy_alloc_blocks(struct drm_buddy *mm,
928 if (flags & DRM_BUDDY_RANGE_ALLOCATION)
929 /* Allocate traversing within the range */
930 return __drm_buddy_alloc_range_bias(mm, start, end,
933 /* Allocate from freelist */
934 return alloc_from_freelist(mm, order, flags);
938 * drm_buddy_alloc_blocks - allocate power-of-two blocks
940 * @mm: DRM buddy manager to allocate from
941 * @start: start of the allowed range for this block
942 * @end: end of the allowed range for this block
943 * @size: size of the allocation in bytes
944 * @min_block_size: alignment of the allocation
945 * @blocks: output list head to add allocated blocks
946 * @flags: DRM_BUDDY_*_ALLOCATION flags
948 * alloc_range_bias() called on range limitations, which traverses
949 * the tree and returns the desired block.
951 * alloc_from_freelist() called when *no* range restrictions
952 * are enforced, which picks the block from the freelist.
955 * 0 on success, error code on failure.
957 int drm_buddy_alloc_blocks(struct drm_buddy *mm,
958 u64 start, u64 end, u64 size,
960 struct list_head *blocks,
963 struct drm_buddy_block *block = NULL;
964 u64 original_size, original_min_size;
965 unsigned int min_order, order;
966 LIST_HEAD(allocated);
970 if (size < mm->chunk_size)
973 if (min_block_size < mm->chunk_size)
976 if (!is_power_of_2(min_block_size))
979 if (!IS_ALIGNED(start | end | size, mm->chunk_size))
985 if (range_overflows(start, size, mm->size))
988 /* Actual range allocation */
989 if (start + size == end) {
990 if (!IS_ALIGNED(start | end, min_block_size))
993 return __drm_buddy_alloc_range(mm, start, size, NULL, blocks);
996 original_size = size;
997 original_min_size = min_block_size;
999 /* Roundup the size to power of 2 */
1000 if (flags & DRM_BUDDY_CONTIGUOUS_ALLOCATION) {
1001 size = roundup_pow_of_two(size);
1002 min_block_size = size;
1003 /* Align size value to min_block_size */
1004 } else if (!IS_ALIGNED(size, min_block_size)) {
1005 size = round_up(size, min_block_size);
1008 pages = size >> ilog2(mm->chunk_size);
1009 order = fls(pages) - 1;
1010 min_order = ilog2(min_block_size) - ilog2(mm->chunk_size);
1013 order = min(order, (unsigned int)fls(pages) - 1);
1014 BUG_ON(order > mm->max_order);
1015 BUG_ON(order < min_order);
1018 block = __drm_buddy_alloc_blocks(mm, start,
1025 if (order-- == min_order) {
1026 /* Try allocation through force merge method */
1027 if (mm->clear_avail &&
1028 !__force_merge(mm, start, end, min_order)) {
1029 block = __drm_buddy_alloc_blocks(mm, start,
1033 if (!IS_ERR(block)) {
1040 * Try contiguous block allocation through
1041 * try harder method.
1043 if (flags & DRM_BUDDY_CONTIGUOUS_ALLOCATION &&
1044 !(flags & DRM_BUDDY_RANGE_ALLOCATION))
1045 return __alloc_contig_try_harder(mm,
1054 mark_allocated(block);
1055 mm->avail -= drm_buddy_block_size(mm, block);
1056 if (drm_buddy_block_is_clear(block))
1057 mm->clear_avail -= drm_buddy_block_size(mm, block);
1058 kmemleak_update_trace(block);
1059 list_add_tail(&block->link, &allocated);
1061 pages -= BIT(order);
1067 /* Trim the allocated block to the required size */
1068 if (original_size != size) {
1069 struct list_head *trim_list;
1073 trim_list = &allocated;
1074 trim_size = original_size;
1076 if (!list_is_singular(&allocated)) {
1077 block = list_last_entry(&allocated, typeof(*block), link);
1078 list_move(&block->link, &temp);
1080 trim_size = drm_buddy_block_size(mm, block) -
1081 (size - original_size);
1084 drm_buddy_block_trim(mm,
1088 if (!list_empty(&temp))
1089 list_splice_tail(trim_list, &allocated);
1092 list_splice_tail(&allocated, blocks);
1096 drm_buddy_free_list_internal(mm, &allocated);
1099 EXPORT_SYMBOL(drm_buddy_alloc_blocks);
1102 * drm_buddy_block_print - print block information
1104 * @mm: DRM buddy manager
1105 * @block: DRM buddy block
1106 * @p: DRM printer to use
1108 void drm_buddy_block_print(struct drm_buddy *mm,
1109 struct drm_buddy_block *block,
1110 struct drm_printer *p)
1112 u64 start = drm_buddy_block_offset(block);
1113 u64 size = drm_buddy_block_size(mm, block);
1115 drm_printf(p, "%#018llx-%#018llx: %llu\n", start, start + size, size);
1117 EXPORT_SYMBOL(drm_buddy_block_print);
1120 * drm_buddy_print - print allocator state
1122 * @mm: DRM buddy manager
1123 * @p: DRM printer to use
1125 void drm_buddy_print(struct drm_buddy *mm, struct drm_printer *p)
1129 drm_printf(p, "chunk_size: %lluKiB, total: %lluMiB, free: %lluMiB, clear_free: %lluMiB\n",
1130 mm->chunk_size >> 10, mm->size >> 20, mm->avail >> 20, mm->clear_avail >> 20);
1132 for (order = mm->max_order; order >= 0; order--) {
1133 struct drm_buddy_block *block;
1134 u64 count = 0, free;
1136 list_for_each_entry(block, &mm->free_list[order], link) {
1137 BUG_ON(!drm_buddy_block_is_free(block));
1141 drm_printf(p, "order-%2d ", order);
1143 free = count * (mm->chunk_size << order);
1145 drm_printf(p, "free: %8llu KiB", free >> 10);
1147 drm_printf(p, "free: %8llu MiB", free >> 20);
1149 drm_printf(p, ", blocks: %llu\n", count);
1152 EXPORT_SYMBOL(drm_buddy_print);
1154 static void drm_buddy_module_exit(void)
1156 kmem_cache_destroy(slab_blocks);
1159 static int __init drm_buddy_module_init(void)
1161 slab_blocks = KMEM_CACHE(drm_buddy_block, 0);
1168 module_init(drm_buddy_module_init);
1169 module_exit(drm_buddy_module_exit);
1171 MODULE_DESCRIPTION("DRM Buddy Allocator");
1172 MODULE_LICENSE("Dual MIT/GPL");