]> Git Repo - linux.git/blob - drivers/gpu/drm/i915/i915_buddy.c
platform/x86: amd-pmc: Move to later in the suspend process
[linux.git] / drivers / gpu / drm / i915 / i915_buddy.c
1 // SPDX-License-Identifier: MIT
2 /*
3  * Copyright © 2021 Intel Corporation
4  */
5
6 #include <linux/kmemleak.h>
7 #include <linux/sizes.h>
8
9 #include "i915_buddy.h"
10
11 #include "i915_gem.h"
12 #include "i915_utils.h"
13
14 static struct kmem_cache *slab_blocks;
15
16 static struct i915_buddy_block *i915_block_alloc(struct i915_buddy_mm *mm,
17                                                  struct i915_buddy_block *parent,
18                                                  unsigned int order,
19                                                  u64 offset)
20 {
21         struct i915_buddy_block *block;
22
23         GEM_BUG_ON(order > I915_BUDDY_MAX_ORDER);
24
25         block = kmem_cache_zalloc(slab_blocks, GFP_KERNEL);
26         if (!block)
27                 return NULL;
28
29         block->header = offset;
30         block->header |= order;
31         block->parent = parent;
32
33         GEM_BUG_ON(block->header & I915_BUDDY_HEADER_UNUSED);
34         return block;
35 }
36
37 static void i915_block_free(struct i915_buddy_mm *mm,
38                             struct i915_buddy_block *block)
39 {
40         kmem_cache_free(slab_blocks, block);
41 }
42
43 static void mark_allocated(struct i915_buddy_block *block)
44 {
45         block->header &= ~I915_BUDDY_HEADER_STATE;
46         block->header |= I915_BUDDY_ALLOCATED;
47
48         list_del(&block->link);
49 }
50
51 static void mark_free(struct i915_buddy_mm *mm,
52                       struct i915_buddy_block *block)
53 {
54         block->header &= ~I915_BUDDY_HEADER_STATE;
55         block->header |= I915_BUDDY_FREE;
56
57         list_add(&block->link,
58                  &mm->free_list[i915_buddy_block_order(block)]);
59 }
60
61 static void mark_split(struct i915_buddy_block *block)
62 {
63         block->header &= ~I915_BUDDY_HEADER_STATE;
64         block->header |= I915_BUDDY_SPLIT;
65
66         list_del(&block->link);
67 }
68
69 int i915_buddy_init(struct i915_buddy_mm *mm, u64 size, u64 chunk_size)
70 {
71         unsigned int i;
72         u64 offset;
73
74         if (size < chunk_size)
75                 return -EINVAL;
76
77         if (chunk_size < PAGE_SIZE)
78                 return -EINVAL;
79
80         if (!is_power_of_2(chunk_size))
81                 return -EINVAL;
82
83         size = round_down(size, chunk_size);
84
85         mm->size = size;
86         mm->avail = size;
87         mm->chunk_size = chunk_size;
88         mm->max_order = ilog2(size) - ilog2(chunk_size);
89
90         GEM_BUG_ON(mm->max_order > I915_BUDDY_MAX_ORDER);
91
92         mm->free_list = kmalloc_array(mm->max_order + 1,
93                                       sizeof(struct list_head),
94                                       GFP_KERNEL);
95         if (!mm->free_list)
96                 return -ENOMEM;
97
98         for (i = 0; i <= mm->max_order; ++i)
99                 INIT_LIST_HEAD(&mm->free_list[i]);
100
101         mm->n_roots = hweight64(size);
102
103         mm->roots = kmalloc_array(mm->n_roots,
104                                   sizeof(struct i915_buddy_block *),
105                                   GFP_KERNEL);
106         if (!mm->roots)
107                 goto out_free_list;
108
109         offset = 0;
110         i = 0;
111
112         /*
113          * Split into power-of-two blocks, in case we are given a size that is
114          * not itself a power-of-two.
115          */
116         do {
117                 struct i915_buddy_block *root;
118                 unsigned int order;
119                 u64 root_size;
120
121                 root_size = rounddown_pow_of_two(size);
122                 order = ilog2(root_size) - ilog2(chunk_size);
123
124                 root = i915_block_alloc(mm, NULL, order, offset);
125                 if (!root)
126                         goto out_free_roots;
127
128                 mark_free(mm, root);
129
130                 GEM_BUG_ON(i > mm->max_order);
131                 GEM_BUG_ON(i915_buddy_block_size(mm, root) < chunk_size);
132
133                 mm->roots[i] = root;
134
135                 offset += root_size;
136                 size -= root_size;
137                 i++;
138         } while (size);
139
140         return 0;
141
142 out_free_roots:
143         while (i--)
144                 i915_block_free(mm, mm->roots[i]);
145         kfree(mm->roots);
146 out_free_list:
147         kfree(mm->free_list);
148         return -ENOMEM;
149 }
150
151 void i915_buddy_fini(struct i915_buddy_mm *mm)
152 {
153         int i;
154
155         for (i = 0; i < mm->n_roots; ++i) {
156                 GEM_WARN_ON(!i915_buddy_block_is_free(mm->roots[i]));
157                 i915_block_free(mm, mm->roots[i]);
158         }
159
160         GEM_WARN_ON(mm->avail != mm->size);
161
162         kfree(mm->roots);
163         kfree(mm->free_list);
164 }
165
166 static int split_block(struct i915_buddy_mm *mm,
167                        struct i915_buddy_block *block)
168 {
169         unsigned int block_order = i915_buddy_block_order(block) - 1;
170         u64 offset = i915_buddy_block_offset(block);
171
172         GEM_BUG_ON(!i915_buddy_block_is_free(block));
173         GEM_BUG_ON(!i915_buddy_block_order(block));
174
175         block->left = i915_block_alloc(mm, block, block_order, offset);
176         if (!block->left)
177                 return -ENOMEM;
178
179         block->right = i915_block_alloc(mm, block, block_order,
180                                         offset + (mm->chunk_size << block_order));
181         if (!block->right) {
182                 i915_block_free(mm, block->left);
183                 return -ENOMEM;
184         }
185
186         mark_free(mm, block->left);
187         mark_free(mm, block->right);
188
189         mark_split(block);
190
191         return 0;
192 }
193
194 static struct i915_buddy_block *
195 get_buddy(struct i915_buddy_block *block)
196 {
197         struct i915_buddy_block *parent;
198
199         parent = block->parent;
200         if (!parent)
201                 return NULL;
202
203         if (parent->left == block)
204                 return parent->right;
205
206         return parent->left;
207 }
208
209 static void __i915_buddy_free(struct i915_buddy_mm *mm,
210                               struct i915_buddy_block *block)
211 {
212         struct i915_buddy_block *parent;
213
214         while ((parent = block->parent)) {
215                 struct i915_buddy_block *buddy;
216
217                 buddy = get_buddy(block);
218
219                 if (!i915_buddy_block_is_free(buddy))
220                         break;
221
222                 list_del(&buddy->link);
223
224                 i915_block_free(mm, block);
225                 i915_block_free(mm, buddy);
226
227                 block = parent;
228         }
229
230         mark_free(mm, block);
231 }
232
233 void i915_buddy_free(struct i915_buddy_mm *mm,
234                      struct i915_buddy_block *block)
235 {
236         GEM_BUG_ON(!i915_buddy_block_is_allocated(block));
237         mm->avail += i915_buddy_block_size(mm, block);
238         __i915_buddy_free(mm, block);
239 }
240
241 void i915_buddy_free_list(struct i915_buddy_mm *mm, struct list_head *objects)
242 {
243         struct i915_buddy_block *block, *on;
244
245         list_for_each_entry_safe(block, on, objects, link) {
246                 i915_buddy_free(mm, block);
247                 cond_resched();
248         }
249         INIT_LIST_HEAD(objects);
250 }
251
252 /*
253  * Allocate power-of-two block. The order value here translates to:
254  *
255  *   0 = 2^0 * mm->chunk_size
256  *   1 = 2^1 * mm->chunk_size
257  *   2 = 2^2 * mm->chunk_size
258  *   ...
259  */
260 struct i915_buddy_block *
261 i915_buddy_alloc(struct i915_buddy_mm *mm, unsigned int order)
262 {
263         struct i915_buddy_block *block = NULL;
264         unsigned int i;
265         int err;
266
267         for (i = order; i <= mm->max_order; ++i) {
268                 block = list_first_entry_or_null(&mm->free_list[i],
269                                                  struct i915_buddy_block,
270                                                  link);
271                 if (block)
272                         break;
273         }
274
275         if (!block)
276                 return ERR_PTR(-ENOSPC);
277
278         GEM_BUG_ON(!i915_buddy_block_is_free(block));
279
280         while (i != order) {
281                 err = split_block(mm, block);
282                 if (unlikely(err))
283                         goto out_free;
284
285                 /* Go low */
286                 block = block->left;
287                 i--;
288         }
289
290         mark_allocated(block);
291         mm->avail -= i915_buddy_block_size(mm, block);
292         kmemleak_update_trace(block);
293         return block;
294
295 out_free:
296         if (i != order)
297                 __i915_buddy_free(mm, block);
298         return ERR_PTR(err);
299 }
300
301 static inline bool overlaps(u64 s1, u64 e1, u64 s2, u64 e2)
302 {
303         return s1 <= e2 && e1 >= s2;
304 }
305
306 static inline bool contains(u64 s1, u64 e1, u64 s2, u64 e2)
307 {
308         return s1 <= s2 && e1 >= e2;
309 }
310
311 /*
312  * Allocate range. Note that it's safe to chain together multiple alloc_ranges
313  * with the same blocks list.
314  *
315  * Intended for pre-allocating portions of the address space, for example to
316  * reserve a block for the initial framebuffer or similar, hence the expectation
317  * here is that i915_buddy_alloc() is still the main vehicle for
318  * allocations, so if that's not the case then the drm_mm range allocator is
319  * probably a much better fit, and so you should probably go use that instead.
320  */
321 int i915_buddy_alloc_range(struct i915_buddy_mm *mm,
322                            struct list_head *blocks,
323                            u64 start, u64 size)
324 {
325         struct i915_buddy_block *block;
326         struct i915_buddy_block *buddy;
327         LIST_HEAD(allocated);
328         LIST_HEAD(dfs);
329         u64 end;
330         int err;
331         int i;
332
333         if (size < mm->chunk_size)
334                 return -EINVAL;
335
336         if (!IS_ALIGNED(size | start, mm->chunk_size))
337                 return -EINVAL;
338
339         if (range_overflows(start, size, mm->size))
340                 return -EINVAL;
341
342         for (i = 0; i < mm->n_roots; ++i)
343                 list_add_tail(&mm->roots[i]->tmp_link, &dfs);
344
345         end = start + size - 1;
346
347         do {
348                 u64 block_start;
349                 u64 block_end;
350
351                 block = list_first_entry_or_null(&dfs,
352                                                  struct i915_buddy_block,
353                                                  tmp_link);
354                 if (!block)
355                         break;
356
357                 list_del(&block->tmp_link);
358
359                 block_start = i915_buddy_block_offset(block);
360                 block_end = block_start + i915_buddy_block_size(mm, block) - 1;
361
362                 if (!overlaps(start, end, block_start, block_end))
363                         continue;
364
365                 if (i915_buddy_block_is_allocated(block)) {
366                         err = -ENOSPC;
367                         goto err_free;
368                 }
369
370                 if (contains(start, end, block_start, block_end)) {
371                         if (!i915_buddy_block_is_free(block)) {
372                                 err = -ENOSPC;
373                                 goto err_free;
374                         }
375
376                         mark_allocated(block);
377                         mm->avail -= i915_buddy_block_size(mm, block);
378                         list_add_tail(&block->link, &allocated);
379                         continue;
380                 }
381
382                 if (!i915_buddy_block_is_split(block)) {
383                         err = split_block(mm, block);
384                         if (unlikely(err))
385                                 goto err_undo;
386                 }
387
388                 list_add(&block->right->tmp_link, &dfs);
389                 list_add(&block->left->tmp_link, &dfs);
390         } while (1);
391
392         list_splice_tail(&allocated, blocks);
393         return 0;
394
395 err_undo:
396         /*
397          * We really don't want to leave around a bunch of split blocks, since
398          * bigger is better, so make sure we merge everything back before we
399          * free the allocated blocks.
400          */
401         buddy = get_buddy(block);
402         if (buddy &&
403             (i915_buddy_block_is_free(block) &&
404              i915_buddy_block_is_free(buddy)))
405                 __i915_buddy_free(mm, block);
406
407 err_free:
408         i915_buddy_free_list(mm, &allocated);
409         return err;
410 }
411
412 void i915_buddy_block_print(struct i915_buddy_mm *mm,
413                             struct i915_buddy_block *block,
414                             struct drm_printer *p)
415 {
416         u64 start = i915_buddy_block_offset(block);
417         u64 size = i915_buddy_block_size(mm, block);
418
419         drm_printf(p, "%#018llx-%#018llx: %llu\n", start, start + size, size);
420 }
421
422 void i915_buddy_print(struct i915_buddy_mm *mm, struct drm_printer *p)
423 {
424         int order;
425
426         drm_printf(p, "chunk_size: %lluKiB, total: %lluMiB, free: %lluMiB\n",
427                    mm->chunk_size >> 10, mm->size >> 20, mm->avail >> 20);
428
429         for (order = mm->max_order; order >= 0; order--) {
430                 struct i915_buddy_block *block;
431                 u64 count = 0, free;
432
433                 list_for_each_entry(block, &mm->free_list[order], link) {
434                         GEM_BUG_ON(!i915_buddy_block_is_free(block));
435                         count++;
436                 }
437
438                 drm_printf(p, "order-%d ", order);
439
440                 free = count * (mm->chunk_size << order);
441                 if (free < SZ_1M)
442                         drm_printf(p, "free: %lluKiB", free >> 10);
443                 else
444                         drm_printf(p, "free: %lluMiB", free >> 20);
445
446                 drm_printf(p, ", pages: %llu\n", count);
447         }
448 }
449
450 #if IS_ENABLED(CONFIG_DRM_I915_SELFTEST)
451 #include "selftests/i915_buddy.c"
452 #endif
453
454 void i915_buddy_module_exit(void)
455 {
456         kmem_cache_destroy(slab_blocks);
457 }
458
459 int __init i915_buddy_module_init(void)
460 {
461         slab_blocks = KMEM_CACHE(i915_buddy_block, 0);
462         if (!slab_blocks)
463                 return -ENOMEM;
464
465         return 0;
466 }
This page took 0.05876 seconds and 4 git commands to generate.