1 // SPDX-License-Identifier: MIT
3 * Copyright © 2019 Intel Corporation
6 #include <linux/prime_numbers.h>
7 #include <linux/sort.h>
9 #include "../i915_selftest.h"
12 #include "mock_gem_device.h"
13 #include "mock_region.h"
15 #include "gem/i915_gem_context.h"
16 #include "gem/i915_gem_lmem.h"
17 #include "gem/i915_gem_region.h"
18 #include "gem/i915_gem_object_blt.h"
19 #include "gem/selftests/igt_gem_utils.h"
20 #include "gem/selftests/mock_context.h"
21 #include "gt/intel_engine_user.h"
22 #include "gt/intel_gt.h"
23 #include "i915_memcpy.h"
24 #include "selftests/igt_flush_test.h"
25 #include "selftests/i915_random.h"
27 static void close_objects(struct intel_memory_region *mem,
28 struct list_head *objects)
30 struct drm_i915_private *i915 = mem->i915;
31 struct drm_i915_gem_object *obj, *on;
33 list_for_each_entry_safe(obj, on, objects, st_link) {
34 i915_gem_object_lock(obj, NULL);
35 if (i915_gem_object_has_pinned_pages(obj))
36 i915_gem_object_unpin_pages(obj);
37 /* No polluting the memory region between tests */
38 __i915_gem_object_put_pages(obj);
39 i915_gem_object_unlock(obj);
40 list_del(&obj->st_link);
41 i915_gem_object_put(obj);
46 i915_gem_drain_freed_objects(i915);
49 static int igt_mock_fill(void *arg)
51 struct intel_memory_region *mem = arg;
52 resource_size_t total = resource_size(&mem->region);
53 resource_size_t page_size;
55 unsigned long max_pages;
56 unsigned long page_num;
60 page_size = mem->chunk_size;
63 max_pages = div64_u64(rem, page_size);
65 for_each_prime_number_from(page_num, 1, max_pages) {
66 resource_size_t size = page_num * page_size;
67 struct drm_i915_gem_object *obj;
69 obj = i915_gem_object_create_region(mem, size, 0);
75 err = i915_gem_object_pin_pages_unlocked(obj);
77 i915_gem_object_put(obj);
81 list_add(&obj->st_link, &objects);
88 if (page_num * page_size <= rem) {
89 if (mem->is_range_manager && max_pages > 1) {
94 pr_err("%s failed, space still left in region\n",
102 close_objects(mem, &objects);
107 static struct drm_i915_gem_object *
108 igt_object_create(struct intel_memory_region *mem,
109 struct list_head *objects,
113 struct drm_i915_gem_object *obj;
116 obj = i915_gem_object_create_region(mem, size, flags);
120 err = i915_gem_object_pin_pages_unlocked(obj);
124 list_add(&obj->st_link, objects);
128 i915_gem_object_put(obj);
132 static void igt_object_release(struct drm_i915_gem_object *obj)
134 i915_gem_object_lock(obj, NULL);
135 i915_gem_object_unpin_pages(obj);
136 __i915_gem_object_put_pages(obj);
137 i915_gem_object_unlock(obj);
138 list_del(&obj->st_link);
139 i915_gem_object_put(obj);
142 static bool is_contiguous(struct drm_i915_gem_object *obj)
144 struct scatterlist *sg;
145 dma_addr_t addr = -1;
147 for (sg = obj->mm.pages->sgl; sg; sg = sg_next(sg)) {
148 if (addr != -1 && sg_dma_address(sg) != addr)
151 addr = sg_dma_address(sg) + sg_dma_len(sg);
157 static int igt_mock_reserve(void *arg)
159 struct intel_memory_region *mem = arg;
160 resource_size_t avail = resource_size(&mem->region);
161 struct drm_i915_gem_object *obj;
162 const u32 chunk_size = SZ_32M;
163 u32 i, offset, count, *order;
164 u64 allocated, cur_avail;
165 I915_RND_STATE(prng);
169 if (!list_empty(&mem->reserved)) {
170 pr_err("%s region reserved list is not empty\n", __func__);
174 count = avail / chunk_size;
175 order = i915_random_order(count, &prng);
179 /* Reserve a bunch of ranges within the region */
180 for (i = 0; i < count; ++i) {
181 u64 start = order[i] * chunk_size;
182 u64 size = i915_prandom_u32_max_state(chunk_size, &prng);
184 /* Allow for some really big holes */
188 size = round_up(size, PAGE_SIZE);
189 offset = igt_random_offset(&prng, 0, chunk_size, size,
192 err = intel_memory_region_reserve(mem, start + offset, size);
194 pr_err("%s failed to reserve range", __func__);
198 /* XXX: maybe sanity check the block range here? */
202 /* Try to see if we can allocate from the remaining space */
206 u32 size = i915_prandom_u32_max_state(cur_avail, &prng);
209 size = max_t(u32, round_up(size, PAGE_SIZE), PAGE_SIZE);
210 obj = igt_object_create(mem, &objects, size, 0);
212 if (PTR_ERR(obj) == -ENXIO) {
213 if (mem->is_range_manager &&
214 size > mem->chunk_size) {
227 if (allocated != avail) {
228 pr_err("%s mismatch between allocation and free space", __func__);
234 close_objects(mem, &objects);
235 intel_memory_region_unreserve(mem);
239 static int igt_mock_contiguous(void *arg)
241 struct intel_memory_region *mem = arg;
242 struct drm_i915_gem_object *obj;
243 unsigned long n_objects;
246 I915_RND_STATE(prng);
247 resource_size_t total;
252 total = resource_size(&mem->region);
255 obj = igt_object_create(mem, &objects, mem->chunk_size,
256 I915_BO_ALLOC_CONTIGUOUS);
260 if (!is_contiguous(obj)) {
261 pr_err("%s min object spans disjoint sg entries\n", __func__);
263 goto err_close_objects;
266 igt_object_release(obj);
269 obj = igt_object_create(mem, &objects, total, I915_BO_ALLOC_CONTIGUOUS);
273 if (!is_contiguous(obj)) {
274 pr_err("%s max object spans disjoint sg entries\n", __func__);
276 goto err_close_objects;
279 igt_object_release(obj);
281 /* Internal fragmentation should not bleed into the object size */
282 target = i915_prandom_u64_state(&prng);
283 div64_u64_rem(target, total, &target);
284 target = round_up(target, PAGE_SIZE);
285 target = max_t(u64, PAGE_SIZE, target);
287 obj = igt_object_create(mem, &objects, target,
288 I915_BO_ALLOC_CONTIGUOUS);
292 if (obj->base.size != target) {
293 pr_err("%s obj->base.size(%zx) != target(%llx)\n", __func__,
294 obj->base.size, target);
296 goto err_close_objects;
299 if (!is_contiguous(obj)) {
300 pr_err("%s object spans disjoint sg entries\n", __func__);
302 goto err_close_objects;
305 igt_object_release(obj);
308 * Try to fragment the address space, such that half of it is free, but
309 * the max contiguous block size is SZ_64K.
313 n_objects = div64_u64(total, target);
315 while (n_objects--) {
316 struct list_head *list;
323 obj = igt_object_create(mem, list, target,
324 I915_BO_ALLOC_CONTIGUOUS);
327 goto err_close_objects;
331 close_objects(mem, &holes);
336 if (!mem->is_range_manager) {
337 /* Make sure we can still allocate all the fragmented space */
338 obj = igt_object_create(mem, &objects, target, 0);
341 goto err_close_objects;
344 igt_object_release(obj);
348 * Even though we have enough free space, we don't have a big enough
349 * contiguous block. Make sure that holds true.
353 bool should_fail = target > min;
355 obj = igt_object_create(mem, &objects, target,
356 I915_BO_ALLOC_CONTIGUOUS);
357 if (should_fail != IS_ERR(obj)) {
358 pr_err("%s target allocation(%llx) mismatch\n",
361 goto err_close_objects;
365 } while (target >= mem->chunk_size);
368 list_splice_tail(&holes, &objects);
369 close_objects(mem, &objects);
373 static int igt_mock_splintered_region(void *arg)
375 struct intel_memory_region *mem = arg;
376 struct drm_i915_private *i915 = mem->i915;
377 struct drm_i915_gem_object *obj;
378 unsigned int expected_order;
384 * Sanity check we can still allocate everything even if the
385 * max_order != mm.size. i.e our starting address space size is not a
389 size = (SZ_4G - 1) & PAGE_MASK;
390 mem = mock_region_create(i915, 0, size, PAGE_SIZE, 0);
394 expected_order = get_order(rounddown_pow_of_two(size));
395 if (mem->max_order != expected_order) {
396 pr_err("%s order mismatch(%u != %u)\n",
397 __func__, mem->max_order, expected_order);
402 obj = igt_object_create(mem, &objects, size, 0);
408 close_objects(mem, &objects);
411 * While we should be able allocate everything without any flag
412 * restrictions, if we consider I915_BO_ALLOC_CONTIGUOUS then we are
413 * actually limited to the largest power-of-two for the region size i.e
414 * max_order, due to the inner workings of the buddy allocator. So make
415 * sure that does indeed hold true.
418 if (!mem->is_range_manager) {
419 obj = igt_object_create(mem, &objects, size,
420 I915_BO_ALLOC_CONTIGUOUS);
422 pr_err("%s too large contiguous allocation was not rejected\n",
429 obj = igt_object_create(mem, &objects, rounddown_pow_of_two(size),
430 I915_BO_ALLOC_CONTIGUOUS);
432 pr_err("%s largest possible contiguous allocation failed\n",
439 close_objects(mem, &objects);
441 intel_memory_region_put(mem);
445 static int igt_gpu_write_dw(struct intel_context *ce,
446 struct i915_vma *vma,
450 return igt_gpu_fill_dw(ce, vma, dword * sizeof(u32),
451 vma->size >> PAGE_SHIFT, value);
454 static int igt_cpu_check(struct drm_i915_gem_object *obj, u32 dword, u32 val)
456 unsigned long n = obj->base.size >> PAGE_SHIFT;
460 err = i915_gem_object_wait(obj, 0, MAX_SCHEDULE_TIMEOUT);
464 ptr = i915_gem_object_pin_map(obj, I915_MAP_WC);
471 pr_err("base[%u]=%08x, val=%08x\n",
477 ptr += PAGE_SIZE / sizeof(*ptr);
480 i915_gem_object_unpin_map(obj);
484 static int igt_gpu_write(struct i915_gem_context *ctx,
485 struct drm_i915_gem_object *obj)
487 struct i915_gem_engines *engines;
488 struct i915_gem_engines_iter it;
489 struct i915_address_space *vm;
490 struct intel_context *ce;
491 I915_RND_STATE(prng);
492 IGT_TIMEOUT(end_time);
494 struct i915_vma *vma;
499 GEM_BUG_ON(!i915_gem_object_has_pinned_pages(obj));
503 for_each_gem_engine(ce, i915_gem_context_lock_engines(ctx), it) {
505 if (!intel_engine_can_store_dword(ce->engine))
511 i915_gem_context_unlock_engines(ctx);
515 order = i915_random_order(count * count, &prng);
519 vma = i915_vma_instance(obj, vm, NULL);
525 err = i915_vma_pin(vma, 0, 0, PIN_USER);
530 engines = i915_gem_context_lock_engines(ctx);
532 u32 rng = prandom_u32_state(&prng);
533 u32 dword = offset_in_page(rng) / 4;
535 ce = engines->engines[order[i] % engines->num_engines];
536 i = (i + 1) % (count * count);
537 if (!ce || !intel_engine_can_store_dword(ce->engine))
540 err = igt_gpu_write_dw(ce, vma, dword, rng);
544 i915_gem_object_lock(obj, NULL);
545 err = igt_cpu_check(obj, dword, rng);
546 i915_gem_object_unlock(obj);
549 } while (!__igt_timeout(end_time, NULL));
550 i915_gem_context_unlock_engines(ctx);
561 static int igt_lmem_create(void *arg)
563 struct drm_i915_private *i915 = arg;
564 struct drm_i915_gem_object *obj;
567 obj = i915_gem_object_create_lmem(i915, PAGE_SIZE, 0);
571 err = i915_gem_object_pin_pages_unlocked(obj);
575 i915_gem_object_unpin_pages(obj);
577 i915_gem_object_put(obj);
582 static int igt_lmem_create_cleared_cpu(void *arg)
584 struct drm_i915_private *i915 = arg;
585 I915_RND_STATE(prng);
586 IGT_TIMEOUT(end_time);
590 i915_gem_drain_freed_objects(i915);
592 size = max_t(u32, PAGE_SIZE, i915_prandom_u32_max_state(SZ_32M, &prng));
593 size = round_up(size, PAGE_SIZE);
597 struct drm_i915_gem_object *obj;
603 * Alternate between cleared and uncleared allocations, while
604 * also dirtying the pages each time to check that the pages are
605 * always cleared if requested, since we should get some overlap
606 * of the underlying pages, if not all, since we are the only
610 flags = I915_BO_ALLOC_CPU_CLEAR;
614 obj = i915_gem_object_create_lmem(i915, size, flags);
618 i915_gem_object_lock(obj, NULL);
619 err = i915_gem_object_pin_pages(obj);
623 dword = i915_prandom_u32_max_state(PAGE_SIZE / sizeof(u32),
626 if (flags & I915_BO_ALLOC_CPU_CLEAR) {
627 err = igt_cpu_check(obj, dword, 0);
629 pr_err("%s failed with size=%u, flags=%u\n",
630 __func__, size, flags);
635 vaddr = i915_gem_object_pin_map(obj, I915_MAP_WC);
637 err = PTR_ERR(vaddr);
641 val = prandom_u32_state(&prng);
643 memset32(vaddr, val, obj->base.size / sizeof(u32));
645 i915_gem_object_flush_map(obj);
646 i915_gem_object_unpin_map(obj);
648 i915_gem_object_unpin_pages(obj);
649 __i915_gem_object_put_pages(obj);
651 i915_gem_object_unlock(obj);
652 i915_gem_object_put(obj);
657 } while (!__igt_timeout(end_time, NULL));
659 pr_info("%s completed (%u) iterations\n", __func__, i);
664 static int igt_lmem_write_gpu(void *arg)
666 struct drm_i915_private *i915 = arg;
667 struct drm_i915_gem_object *obj;
668 struct i915_gem_context *ctx;
670 I915_RND_STATE(prng);
674 file = mock_file(i915);
676 return PTR_ERR(file);
678 ctx = live_context(i915, file);
684 sz = round_up(prandom_u32_state(&prng) % SZ_32M, PAGE_SIZE);
686 obj = i915_gem_object_create_lmem(i915, sz, 0);
692 err = i915_gem_object_pin_pages_unlocked(obj);
696 err = igt_gpu_write(ctx, obj);
698 pr_err("igt_gpu_write failed(%d)\n", err);
700 i915_gem_object_unpin_pages(obj);
702 i915_gem_object_put(obj);
708 static struct intel_engine_cs *
709 random_engine_class(struct drm_i915_private *i915,
711 struct rnd_state *prng)
713 struct intel_engine_cs *engine;
717 for (engine = intel_engine_lookup_user(i915, class, 0);
718 engine && engine->uabi_class == class;
719 engine = rb_entry_safe(rb_next(&engine->uabi_node),
720 typeof(*engine), uabi_node))
723 count = i915_prandom_u32_max_state(count, prng);
724 return intel_engine_lookup_user(i915, class, count);
727 static int igt_lmem_write_cpu(void *arg)
729 struct drm_i915_private *i915 = arg;
730 struct drm_i915_gem_object *obj;
731 I915_RND_STATE(prng);
732 IGT_TIMEOUT(end_time);
734 0, /* rng placeholder */
739 PAGE_SIZE - sizeof(u32),
740 PAGE_SIZE - sizeof(u64),
743 struct intel_engine_cs *engine;
751 engine = random_engine_class(i915, I915_ENGINE_CLASS_COPY, &prng);
755 pr_info("%s: using %s\n", __func__, engine->name);
757 sz = round_up(prandom_u32_state(&prng) % SZ_32M, PAGE_SIZE);
758 sz = max_t(u32, 2 * PAGE_SIZE, sz);
760 obj = i915_gem_object_create_lmem(i915, sz, I915_BO_ALLOC_CONTIGUOUS);
764 vaddr = i915_gem_object_pin_map_unlocked(obj, I915_MAP_WC);
766 err = PTR_ERR(vaddr);
770 /* Put the pages into a known state -- from the gpu for added fun */
771 intel_engine_pm_get(engine);
772 err = i915_gem_object_fill_blt(obj, engine->kernel_context, 0xdeadbeaf);
773 intel_engine_pm_put(engine);
777 i915_gem_object_lock(obj, NULL);
778 err = i915_gem_object_set_to_wc_domain(obj, true);
779 i915_gem_object_unlock(obj);
783 count = ARRAY_SIZE(bytes);
784 order = i915_random_order(count * count, &prng);
790 /* A random multiple of u32, picked between [64, PAGE_SIZE - 64] */
791 bytes[0] = igt_random_offset(&prng, 64, PAGE_SIZE - 64, 0, sizeof(u32));
792 GEM_BUG_ON(!IS_ALIGNED(bytes[0], sizeof(u32)));
802 size = bytes[order[i] % count];
803 i = (i + 1) % (count * count);
805 align = bytes[order[i] % count];
806 i = (i + 1) % (count * count);
808 align = max_t(u32, sizeof(u32), rounddown_pow_of_two(align));
810 offset = igt_random_offset(&prng, 0, obj->base.size,
813 val = prandom_u32_state(&prng);
814 memset32(vaddr + offset / sizeof(u32), val ^ 0xdeadbeaf,
818 * Sample random dw -- don't waste precious time reading every
821 dword = igt_random_offset(&prng, offset,
823 sizeof(u32), sizeof(u32));
824 dword /= sizeof(u32);
825 if (vaddr[dword] != (val ^ 0xdeadbeaf)) {
826 pr_err("%s vaddr[%u]=%u, val=%u, size=%u, align=%u, offset=%u\n",
827 __func__, dword, vaddr[dword], val ^ 0xdeadbeaf,
828 size, align, offset);
832 } while (!__igt_timeout(end_time, NULL));
835 i915_gem_object_unpin_map(obj);
837 i915_gem_object_put(obj);
842 static const char *repr_type(u32 type)
854 static struct drm_i915_gem_object *
855 create_region_for_mapping(struct intel_memory_region *mr, u64 size, u32 type,
858 struct drm_i915_gem_object *obj;
861 obj = i915_gem_object_create_region(mr, size, 0);
863 if (PTR_ERR(obj) == -ENOSPC) /* Stolen memory */
864 return ERR_PTR(-ENODEV);
868 addr = i915_gem_object_pin_map_unlocked(obj, type);
870 i915_gem_object_put(obj);
871 if (PTR_ERR(addr) == -ENXIO)
872 return ERR_PTR(-ENODEV);
880 static int wrap_ktime_compare(const void *A, const void *B)
882 const ktime_t *a = A, *b = B;
884 return ktime_compare(*a, *b);
887 static void igt_memcpy_long(void *dst, const void *src, size_t size)
889 unsigned long *tmp = dst;
890 const unsigned long *s = src;
892 size = size / sizeof(unsigned long);
897 static inline void igt_memcpy(void *dst, const void *src, size_t size)
899 memcpy(dst, src, size);
902 static inline void igt_memcpy_from_wc(void *dst, const void *src, size_t size)
904 i915_memcpy_from_wc(dst, src, size);
907 static int _perf_memcpy(struct intel_memory_region *src_mr,
908 struct intel_memory_region *dst_mr,
909 u64 size, u32 src_type, u32 dst_type)
911 struct drm_i915_private *i915 = src_mr->i915;
914 void (*copy)(void *dst, const void *src, size_t size);
928 !i915_has_memcpy_from_wc(),
931 struct drm_i915_gem_object *src, *dst;
932 void *src_addr, *dst_addr;
936 src = create_region_for_mapping(src_mr, size, src_type, &src_addr);
942 dst = create_region_for_mapping(dst_mr, size, dst_type, &dst_addr);
948 for (i = 0; i < ARRAY_SIZE(tests); ++i) {
955 for (pass = 0; pass < ARRAY_SIZE(t); pass++) {
960 tests[i].copy(dst_addr, src_addr, size);
963 t[pass] = ktime_sub(t1, t0);
966 sort(t, ARRAY_SIZE(t), sizeof(*t), wrap_ktime_compare, NULL);
968 /* ignore the impossible to protect our sanity */
969 pr_debug("Skipping %s src(%s, %s) -> dst(%s, %s) %14s %4lluKiB copy, unstable measurement [%lld, %lld]\n",
971 src_mr->name, repr_type(src_type),
972 dst_mr->name, repr_type(dst_type),
973 tests[i].name, size >> 10,
978 pr_info("%s src(%s, %s) -> dst(%s, %s) %14s %4llu KiB copy: %5lld MiB/s\n",
980 src_mr->name, repr_type(src_type),
981 dst_mr->name, repr_type(dst_type),
982 tests[i].name, size >> 10,
983 div64_u64(mul_u32_u32(4 * size,
985 t[1] + 2 * t[2] + t[3]) >> 20);
990 i915_gem_object_unpin_map(dst);
991 i915_gem_object_put(dst);
993 i915_gem_object_unpin_map(src);
994 i915_gem_object_put(src);
996 i915_gem_drain_freed_objects(i915);
1004 static int perf_memcpy(void *arg)
1006 struct drm_i915_private *i915 = arg;
1007 static const u32 types[] = {
1011 static const u32 sizes[] = {
1016 struct intel_memory_region *src_mr, *dst_mr;
1021 for_each_memory_region(src_mr, i915, src_id) {
1022 for_each_memory_region(dst_mr, i915, dst_id) {
1023 for (i = 0; i < ARRAY_SIZE(sizes); ++i) {
1024 for (j = 0; j < ARRAY_SIZE(types); ++j) {
1025 for (k = 0; k < ARRAY_SIZE(types); ++k) {
1026 ret = _perf_memcpy(src_mr,
1042 int intel_memory_region_mock_selftests(void)
1044 static const struct i915_subtest tests[] = {
1045 SUBTEST(igt_mock_reserve),
1046 SUBTEST(igt_mock_fill),
1047 SUBTEST(igt_mock_contiguous),
1048 SUBTEST(igt_mock_splintered_region),
1050 struct intel_memory_region *mem;
1051 struct drm_i915_private *i915;
1054 i915 = mock_gem_device();
1058 mem = mock_region_create(i915, 0, SZ_2G, I915_GTT_PAGE_SIZE_4K, 0);
1060 pr_err("failed to create memory region\n");
1065 err = i915_subtests(tests, mem);
1067 intel_memory_region_put(mem);
1069 mock_destroy_device(i915);
1073 int intel_memory_region_live_selftests(struct drm_i915_private *i915)
1075 static const struct i915_subtest tests[] = {
1076 SUBTEST(igt_lmem_create),
1077 SUBTEST(igt_lmem_create_cleared_cpu),
1078 SUBTEST(igt_lmem_write_cpu),
1079 SUBTEST(igt_lmem_write_gpu),
1082 if (!HAS_LMEM(i915)) {
1083 pr_info("device lacks LMEM support, skipping\n");
1087 if (intel_gt_is_wedged(&i915->gt))
1090 return i915_live_subtests(tests, i915);
1093 int intel_memory_region_perf_selftests(struct drm_i915_private *i915)
1095 static const struct i915_subtest tests[] = {
1096 SUBTEST(perf_memcpy),
1099 if (intel_gt_is_wedged(&i915->gt))
1102 return i915_live_subtests(tests, i915);