2 * Test cases for the drm_mm range manager
5 #define pr_fmt(fmt) "drm_mm: " fmt
7 #include <linux/module.h>
8 #include <linux/prime_numbers.h>
9 #include <linux/slab.h>
10 #include <linux/random.h>
11 #include <linux/vmalloc.h>
13 #include <drm/drm_mm.h>
15 #include "../lib/drm_random.h"
17 #define TESTS "drm_mm_selftests.h"
18 #include "drm_selftest.h"
20 static unsigned int random_seed;
21 static unsigned int max_iterations = 8192;
22 static unsigned int max_prime = 128;
31 static const struct insert_mode {
33 enum drm_mm_insert_mode mode;
35 [BEST] = { "best", DRM_MM_INSERT_BEST },
36 [BOTTOMUP] = { "bottom-up", DRM_MM_INSERT_LOW },
37 [TOPDOWN] = { "top-down", DRM_MM_INSERT_HIGH },
38 [EVICT] = { "evict", DRM_MM_INSERT_EVICT },
41 { "bottom-up", DRM_MM_INSERT_LOW },
42 { "top-down", DRM_MM_INSERT_HIGH },
46 static int igt_sanitycheck(void *ignored)
48 pr_info("%s - ok!\n", __func__);
52 static bool assert_no_holes(const struct drm_mm *mm)
54 struct drm_mm_node *hole;
55 u64 hole_start, hole_end;
59 drm_mm_for_each_hole(hole, mm, hole_start, hole_end)
62 pr_err("Expected to find no holes (after reserve), found %lu instead\n", count);
66 drm_mm_for_each_node(hole, mm) {
67 if (drm_mm_hole_follows(hole)) {
68 pr_err("Hole follows node, expected none!\n");
76 static bool assert_one_hole(const struct drm_mm *mm, u64 start, u64 end)
78 struct drm_mm_node *hole;
79 u64 hole_start, hole_end;
87 drm_mm_for_each_hole(hole, mm, hole_start, hole_end) {
88 if (start != hole_start || end != hole_end) {
90 pr_err("empty mm has incorrect hole, found (%llx, %llx), expect (%llx, %llx)\n",
98 pr_err("Expected to find one hole, found %lu instead\n", count);
105 static bool assert_continuous(const struct drm_mm *mm, u64 size)
107 struct drm_mm_node *node, *check, *found;
111 if (!assert_no_holes(mm))
116 drm_mm_for_each_node(node, mm) {
117 if (node->start != addr) {
118 pr_err("node[%ld] list out of order, expected %llx found %llx\n",
119 n, addr, node->start);
123 if (node->size != size) {
124 pr_err("node[%ld].size incorrect, expected %llx, found %llx\n",
125 n, size, node->size);
129 if (drm_mm_hole_follows(node)) {
130 pr_err("node[%ld] is followed by a hole!\n", n);
135 drm_mm_for_each_node_in_range(check, mm, addr, addr + size) {
137 pr_err("lookup return wrong node, expected start %llx, found %llx\n",
138 node->start, check->start);
144 pr_err("lookup failed for node %llx + %llx\n",
156 static u64 misalignment(struct drm_mm_node *node, u64 alignment)
163 div64_u64_rem(node->start, alignment, &rem);
167 static bool assert_node(struct drm_mm_node *node, struct drm_mm *mm,
168 u64 size, u64 alignment, unsigned long color)
172 if (!drm_mm_node_allocated(node) || node->mm != mm) {
173 pr_err("node not allocated\n");
177 if (node->size != size) {
178 pr_err("node has wrong size, found %llu, expected %llu\n",
183 if (misalignment(node, alignment)) {
184 pr_err("node is misaligned, start %llx rem %llu, expected alignment %llu\n",
185 node->start, misalignment(node, alignment), alignment);
189 if (node->color != color) {
190 pr_err("node has wrong color, found %lu, expected %lu\n",
198 #define show_mm(mm) do { \
199 struct drm_printer __p = drm_debug_printer(__func__); \
200 drm_mm_print((mm), &__p); } while (0)
202 static int igt_init(void *ignored)
204 const unsigned int size = 4096;
206 struct drm_mm_node tmp;
209 /* Start with some simple checks on initialising the struct drm_mm */
210 memset(&mm, 0, sizeof(mm));
211 if (drm_mm_initialized(&mm)) {
212 pr_err("zeroed mm claims to be initialized\n");
216 memset(&mm, 0xff, sizeof(mm));
217 drm_mm_init(&mm, 0, size);
218 if (!drm_mm_initialized(&mm)) {
219 pr_err("mm claims not to be initialized\n");
223 if (!drm_mm_clean(&mm)) {
224 pr_err("mm not empty on creation\n");
228 /* After creation, it should all be one massive hole */
229 if (!assert_one_hole(&mm, 0, size)) {
234 memset(&tmp, 0, sizeof(tmp));
237 ret = drm_mm_reserve_node(&mm, &tmp);
239 pr_err("failed to reserve whole drm_mm\n");
243 /* After filling the range entirely, there should be no holes */
244 if (!assert_no_holes(&mm)) {
249 /* And then after emptying it again, the massive hole should be back */
250 drm_mm_remove_node(&tmp);
251 if (!assert_one_hole(&mm, 0, size)) {
259 drm_mm_takedown(&mm);
263 static int igt_debug(void *ignored)
266 struct drm_mm_node nodes[2];
269 /* Create a small drm_mm with a couple of nodes and a few holes, and
270 * check that the debug iterator doesn't explode over a trivial drm_mm.
273 drm_mm_init(&mm, 0, 4096);
275 memset(nodes, 0, sizeof(nodes));
276 nodes[0].start = 512;
277 nodes[0].size = 1024;
278 ret = drm_mm_reserve_node(&mm, &nodes[0]);
280 pr_err("failed to reserve node[0] {start=%lld, size=%lld)\n",
281 nodes[0].start, nodes[0].size);
285 nodes[1].size = 1024;
286 nodes[1].start = 4096 - 512 - nodes[1].size;
287 ret = drm_mm_reserve_node(&mm, &nodes[1]);
289 pr_err("failed to reserve node[1] {start=%lld, size=%lld)\n",
290 nodes[1].start, nodes[1].size);
298 static struct drm_mm_node *set_node(struct drm_mm_node *node,
306 static bool expect_reserve_fail(struct drm_mm *mm, struct drm_mm_node *node)
310 err = drm_mm_reserve_node(mm, node);
311 if (likely(err == -ENOSPC))
315 pr_err("impossible reserve succeeded, node %llu + %llu\n",
316 node->start, node->size);
317 drm_mm_remove_node(node);
319 pr_err("impossible reserve failed with wrong error %d [expected %d], node %llu + %llu\n",
320 err, -ENOSPC, node->start, node->size);
325 static bool check_reserve_boundaries(struct drm_mm *mm,
329 const struct boundary {
333 #define B(st, sz) { (st), (sz), "{ " #st ", " #sz "}" }
344 B(count*size, -size),
345 B(count*size, count*size),
346 B(count*size, -count*size),
347 B(count*size, -(count+1)*size),
348 B((count+1)*size, size),
349 B((count+1)*size, -size),
350 B((count+1)*size, -2*size),
353 struct drm_mm_node tmp = {};
356 for (n = 0; n < ARRAY_SIZE(boundaries); n++) {
357 if (!expect_reserve_fail(mm,
360 boundaries[n].size))) {
361 pr_err("boundary[%d:%s] failed, count=%u, size=%lld\n",
362 n, boundaries[n].name, count, size);
370 static int __igt_reserve(unsigned int count, u64 size)
372 DRM_RND_STATE(prng, random_seed);
374 struct drm_mm_node tmp, *nodes, *node, *next;
375 unsigned int *order, n, m, o = 0;
378 /* For exercising drm_mm_reserve_node(), we want to check that
379 * reservations outside of the drm_mm range are rejected, and to
380 * overlapping and otherwise already occupied ranges. Afterwards,
381 * the tree and nodes should be intact.
384 DRM_MM_BUG_ON(!count);
385 DRM_MM_BUG_ON(!size);
388 order = drm_random_order(count, &prng);
392 nodes = vzalloc(sizeof(*nodes) * count);
397 drm_mm_init(&mm, 0, count * size);
399 if (!check_reserve_boundaries(&mm, count, size))
402 for (n = 0; n < count; n++) {
403 nodes[n].start = order[n] * size;
404 nodes[n].size = size;
406 err = drm_mm_reserve_node(&mm, &nodes[n]);
408 pr_err("reserve failed, step %d, start %llu\n",
414 if (!drm_mm_node_allocated(&nodes[n])) {
415 pr_err("reserved node not allocated! step %d, start %llu\n",
420 if (!expect_reserve_fail(&mm, &nodes[n]))
424 /* After random insertion the nodes should be in order */
425 if (!assert_continuous(&mm, size))
428 /* Repeated use should then fail */
429 drm_random_reorder(order, count, &prng);
430 for (n = 0; n < count; n++) {
431 if (!expect_reserve_fail(&mm,
432 set_node(&tmp, order[n] * size, 1)))
435 /* Remove and reinsert should work */
436 drm_mm_remove_node(&nodes[order[n]]);
437 err = drm_mm_reserve_node(&mm, &nodes[order[n]]);
439 pr_err("reserve failed, step %d, start %llu\n",
446 if (!assert_continuous(&mm, size))
449 /* Overlapping use should then fail */
450 for (n = 0; n < count; n++) {
451 if (!expect_reserve_fail(&mm, set_node(&tmp, 0, size*count)))
454 for (n = 0; n < count; n++) {
455 if (!expect_reserve_fail(&mm,
458 size * (count - n))))
462 /* Remove several, reinsert, check full */
463 for_each_prime_number(n, min(max_prime, count)) {
464 for (m = 0; m < n; m++) {
465 node = &nodes[order[(o + m) % count]];
466 drm_mm_remove_node(node);
469 for (m = 0; m < n; m++) {
470 node = &nodes[order[(o + m) % count]];
471 err = drm_mm_reserve_node(&mm, node);
473 pr_err("reserve failed, step %d/%d, start %llu\n",
482 if (!assert_continuous(&mm, size))
488 drm_mm_for_each_node_safe(node, next, &mm)
489 drm_mm_remove_node(node);
490 drm_mm_takedown(&mm);
498 static int igt_reserve(void *ignored)
500 const unsigned int count = min_t(unsigned int, BIT(10), max_iterations);
503 for_each_prime_number_from(n, 1, 54) {
504 u64 size = BIT_ULL(n);
506 ret = __igt_reserve(count, size - 1);
510 ret = __igt_reserve(count, size);
514 ret = __igt_reserve(count, size + 1);
522 static bool expect_insert(struct drm_mm *mm, struct drm_mm_node *node,
523 u64 size, u64 alignment, unsigned long color,
524 const struct insert_mode *mode)
528 err = drm_mm_insert_node_generic(mm, node,
529 size, alignment, color,
532 pr_err("insert (size=%llu, alignment=%llu, color=%lu, mode=%s) failed with err=%d\n",
533 size, alignment, color, mode->name, err);
537 if (!assert_node(node, mm, size, alignment, color)) {
538 drm_mm_remove_node(node);
545 static bool expect_insert_fail(struct drm_mm *mm, u64 size)
547 struct drm_mm_node tmp = {};
550 err = drm_mm_insert_node(mm, &tmp, size);
551 if (likely(err == -ENOSPC))
555 pr_err("impossible insert succeeded, node %llu + %llu\n",
556 tmp.start, tmp.size);
557 drm_mm_remove_node(&tmp);
559 pr_err("impossible insert failed with wrong error %d [expected %d], size %llu\n",
565 static int __igt_insert(unsigned int count, u64 size, bool replace)
567 DRM_RND_STATE(prng, random_seed);
568 const struct insert_mode *mode;
570 struct drm_mm_node *nodes, *node, *next;
571 unsigned int *order, n, m, o = 0;
574 /* Fill a range with lots of nodes, check it doesn't fail too early */
576 DRM_MM_BUG_ON(!count);
577 DRM_MM_BUG_ON(!size);
580 nodes = vmalloc(count * sizeof(*nodes));
584 order = drm_random_order(count, &prng);
589 drm_mm_init(&mm, 0, count * size);
591 for (mode = insert_modes; mode->name; mode++) {
592 for (n = 0; n < count; n++) {
593 struct drm_mm_node tmp;
595 node = replace ? &tmp : &nodes[n];
596 memset(node, 0, sizeof(*node));
597 if (!expect_insert(&mm, node, size, 0, n, mode)) {
598 pr_err("%s insert failed, size %llu step %d\n",
599 mode->name, size, n);
604 drm_mm_replace_node(&tmp, &nodes[n]);
605 if (drm_mm_node_allocated(&tmp)) {
606 pr_err("replaced old-node still allocated! step %d\n",
611 if (!assert_node(&nodes[n], &mm, size, 0, n)) {
612 pr_err("replaced node did not inherit parameters, size %llu step %d\n",
617 if (tmp.start != nodes[n].start) {
618 pr_err("replaced node mismatch location expected [%llx + %llx], found [%llx + %llx]\n",
620 nodes[n].start, nodes[n].size);
626 /* After random insertion the nodes should be in order */
627 if (!assert_continuous(&mm, size))
630 /* Repeated use should then fail */
631 if (!expect_insert_fail(&mm, size))
634 /* Remove one and reinsert, as the only hole it should refill itself */
635 for (n = 0; n < count; n++) {
636 u64 addr = nodes[n].start;
638 drm_mm_remove_node(&nodes[n]);
639 if (!expect_insert(&mm, &nodes[n], size, 0, n, mode)) {
640 pr_err("%s reinsert failed, size %llu step %d\n",
641 mode->name, size, n);
645 if (nodes[n].start != addr) {
646 pr_err("%s reinsert node moved, step %d, expected %llx, found %llx\n",
647 mode->name, n, addr, nodes[n].start);
651 if (!assert_continuous(&mm, size))
655 /* Remove several, reinsert, check full */
656 for_each_prime_number(n, min(max_prime, count)) {
657 for (m = 0; m < n; m++) {
658 node = &nodes[order[(o + m) % count]];
659 drm_mm_remove_node(node);
662 for (m = 0; m < n; m++) {
663 node = &nodes[order[(o + m) % count]];
664 if (!expect_insert(&mm, node, size, 0, n, mode)) {
665 pr_err("%s multiple reinsert failed, size %llu step %d\n",
666 mode->name, size, n);
673 if (!assert_continuous(&mm, size))
676 if (!expect_insert_fail(&mm, size))
680 drm_mm_for_each_node_safe(node, next, &mm)
681 drm_mm_remove_node(node);
682 DRM_MM_BUG_ON(!drm_mm_clean(&mm));
687 drm_mm_for_each_node_safe(node, next, &mm)
688 drm_mm_remove_node(node);
689 drm_mm_takedown(&mm);
697 static int igt_insert(void *ignored)
699 const unsigned int count = min_t(unsigned int, BIT(10), max_iterations);
703 for_each_prime_number_from(n, 1, 54) {
704 u64 size = BIT_ULL(n);
706 ret = __igt_insert(count, size - 1, false);
710 ret = __igt_insert(count, size, false);
714 ret = __igt_insert(count, size + 1, false);
720 static int igt_replace(void *ignored)
722 const unsigned int count = min_t(unsigned int, BIT(10), max_iterations);
726 /* Reuse igt_insert to exercise replacement by inserting a dummy node,
727 * then replacing it with the intended node. We want to check that
728 * the tree is intact and all the information we need is carried
729 * across to the target node.
732 for_each_prime_number_from(n, 1, 54) {
733 u64 size = BIT_ULL(n);
735 ret = __igt_insert(count, size - 1, true);
739 ret = __igt_insert(count, size, true);
743 ret = __igt_insert(count, size + 1, true);
749 static bool expect_insert_in_range(struct drm_mm *mm, struct drm_mm_node *node,
750 u64 size, u64 alignment, unsigned long color,
751 u64 range_start, u64 range_end,
752 const struct insert_mode *mode)
756 err = drm_mm_insert_node_in_range(mm, node,
757 size, alignment, color,
758 range_start, range_end,
761 pr_err("insert (size=%llu, alignment=%llu, color=%lu, mode=%s) nto range [%llx, %llx] failed with err=%d\n",
762 size, alignment, color, mode->name,
763 range_start, range_end, err);
767 if (!assert_node(node, mm, size, alignment, color)) {
768 drm_mm_remove_node(node);
775 static bool expect_insert_in_range_fail(struct drm_mm *mm,
780 struct drm_mm_node tmp = {};
783 err = drm_mm_insert_node_in_range(mm, &tmp,
785 range_start, range_end,
787 if (likely(err == -ENOSPC))
791 pr_err("impossible insert succeeded, node %llx + %llu, range [%llx, %llx]\n",
792 tmp.start, tmp.size, range_start, range_end);
793 drm_mm_remove_node(&tmp);
795 pr_err("impossible insert failed with wrong error %d [expected %d], size %llu, range [%llx, %llx]\n",
796 err, -ENOSPC, size, range_start, range_end);
802 static bool assert_contiguous_in_range(struct drm_mm *mm,
807 struct drm_mm_node *node;
810 if (!expect_insert_in_range_fail(mm, size, start, end))
813 n = div64_u64(start + size - 1, size);
814 drm_mm_for_each_node(node, mm) {
815 if (node->start < start || node->start + node->size > end) {
816 pr_err("node %d out of range, address [%llx + %llu], range [%llx, %llx]\n",
817 n, node->start, node->start + node->size, start, end);
821 if (node->start != n * size) {
822 pr_err("node %d out of order, expected start %llx, found %llx\n",
823 n, n * size, node->start);
827 if (node->size != size) {
828 pr_err("node %d has wrong size, expected size %llx, found %llx\n",
829 n, size, node->size);
833 if (drm_mm_hole_follows(node) &&
834 drm_mm_hole_node_end(node) < end) {
835 pr_err("node %d is followed by a hole!\n", n);
843 node = __drm_mm_interval_first(mm, 0, start - 1);
844 if (node->allocated) {
845 pr_err("node before start: node=%llx+%llu, start=%llx\n",
846 node->start, node->size, start);
852 node = __drm_mm_interval_first(mm, end, U64_MAX);
853 if (node->allocated) {
854 pr_err("node after end: node=%llx+%llu, end=%llx\n",
855 node->start, node->size, end);
863 static int __igt_insert_range(unsigned int count, u64 size, u64 start, u64 end)
865 const struct insert_mode *mode;
867 struct drm_mm_node *nodes, *node, *next;
868 unsigned int n, start_n, end_n;
871 DRM_MM_BUG_ON(!count);
872 DRM_MM_BUG_ON(!size);
873 DRM_MM_BUG_ON(end <= start);
875 /* Very similar to __igt_insert(), but now instead of populating the
876 * full range of the drm_mm, we try to fill a small portion of it.
880 nodes = vzalloc(count * sizeof(*nodes));
885 drm_mm_init(&mm, 0, count * size);
887 start_n = div64_u64(start + size - 1, size);
888 end_n = div64_u64(end - size, size);
890 for (mode = insert_modes; mode->name; mode++) {
891 for (n = start_n; n <= end_n; n++) {
892 if (!expect_insert_in_range(&mm, &nodes[n],
895 pr_err("%s insert failed, size %llu, step %d [%d, %d], range [%llx, %llx]\n",
903 if (!assert_contiguous_in_range(&mm, size, start, end)) {
904 pr_err("%s: range [%llx, %llx] not full after initialisation, size=%llu\n",
905 mode->name, start, end, size);
909 /* Remove one and reinsert, it should refill itself */
910 for (n = start_n; n <= end_n; n++) {
911 u64 addr = nodes[n].start;
913 drm_mm_remove_node(&nodes[n]);
914 if (!expect_insert_in_range(&mm, &nodes[n],
917 pr_err("%s reinsert failed, step %d\n", mode->name, n);
921 if (nodes[n].start != addr) {
922 pr_err("%s reinsert node moved, step %d, expected %llx, found %llx\n",
923 mode->name, n, addr, nodes[n].start);
928 if (!assert_contiguous_in_range(&mm, size, start, end)) {
929 pr_err("%s: range [%llx, %llx] not full after reinsertion, size=%llu\n",
930 mode->name, start, end, size);
934 drm_mm_for_each_node_safe(node, next, &mm)
935 drm_mm_remove_node(node);
936 DRM_MM_BUG_ON(!drm_mm_clean(&mm));
941 drm_mm_for_each_node_safe(node, next, &mm)
942 drm_mm_remove_node(node);
943 drm_mm_takedown(&mm);
949 static int insert_outside_range(void)
952 const unsigned int start = 1024;
953 const unsigned int end = 2048;
954 const unsigned int size = end - start;
956 drm_mm_init(&mm, start, size);
958 if (!expect_insert_in_range_fail(&mm, 1, 0, start))
961 if (!expect_insert_in_range_fail(&mm, size,
962 start - size/2, start + (size+1)/2))
965 if (!expect_insert_in_range_fail(&mm, size,
966 end - (size+1)/2, end + size/2))
969 if (!expect_insert_in_range_fail(&mm, 1, end, end + size))
972 drm_mm_takedown(&mm);
976 static int igt_insert_range(void *ignored)
978 const unsigned int count = min_t(unsigned int, BIT(13), max_iterations);
982 /* Check that requests outside the bounds of drm_mm are rejected. */
983 ret = insert_outside_range();
987 for_each_prime_number_from(n, 1, 50) {
988 const u64 size = BIT_ULL(n);
989 const u64 max = count * size;
991 ret = __igt_insert_range(count, size, 0, max);
995 ret = __igt_insert_range(count, size, 1, max);
999 ret = __igt_insert_range(count, size, 0, max - 1);
1003 ret = __igt_insert_range(count, size, 0, max/2);
1007 ret = __igt_insert_range(count, size, max/2, max);
1011 ret = __igt_insert_range(count, size, max/4+1, 3*max/4-1);
1019 static int igt_align(void *ignored)
1021 const struct insert_mode *mode;
1022 const unsigned int max_count = min(8192u, max_prime);
1024 struct drm_mm_node *nodes, *node, *next;
1028 /* For each of the possible insertion modes, we pick a few
1029 * arbitrary alignments and check that the inserted node
1030 * meets our requirements.
1033 nodes = vzalloc(max_count * sizeof(*nodes));
1037 drm_mm_init(&mm, 1, U64_MAX - 2);
1039 for (mode = insert_modes; mode->name; mode++) {
1042 for_each_prime_number_from(prime, 1, max_count) {
1043 u64 size = next_prime_number(prime);
1045 if (!expect_insert(&mm, &nodes[i],
1048 pr_err("%s insert failed with alignment=%d",
1056 drm_mm_for_each_node_safe(node, next, &mm)
1057 drm_mm_remove_node(node);
1058 DRM_MM_BUG_ON(!drm_mm_clean(&mm));
1063 drm_mm_for_each_node_safe(node, next, &mm)
1064 drm_mm_remove_node(node);
1065 drm_mm_takedown(&mm);
1071 static int igt_align_pot(int max)
1074 struct drm_mm_node *node, *next;
1078 /* Check that we can align to the full u64 address space */
1080 drm_mm_init(&mm, 1, U64_MAX - 2);
1082 for (bit = max - 1; bit; bit--) {
1085 node = kzalloc(sizeof(*node), GFP_KERNEL);
1091 align = BIT_ULL(bit);
1092 size = BIT_ULL(bit-1) + 1;
1093 if (!expect_insert(&mm, node,
1095 &insert_modes[0])) {
1096 pr_err("insert failed with alignment=%llx [%d]",
1104 drm_mm_for_each_node_safe(node, next, &mm) {
1105 drm_mm_remove_node(node);
1108 drm_mm_takedown(&mm);
1112 static int igt_align32(void *ignored)
1114 return igt_align_pot(32);
1117 static int igt_align64(void *ignored)
1119 return igt_align_pot(64);
1122 static void show_scan(const struct drm_mm_scan *scan)
1124 pr_info("scan: hit [%llx, %llx], size=%lld, align=%lld, color=%ld\n",
1125 scan->hit_start, scan->hit_end,
1126 scan->size, scan->alignment, scan->color);
1129 static void show_holes(const struct drm_mm *mm, int count)
1131 u64 hole_start, hole_end;
1132 struct drm_mm_node *hole;
1134 drm_mm_for_each_hole(hole, mm, hole_start, hole_end) {
1135 struct drm_mm_node *next = list_next_entry(hole, node_list);
1136 const char *node1 = NULL, *node2 = NULL;
1138 if (hole->allocated)
1139 node1 = kasprintf(GFP_KERNEL,
1140 "[%llx + %lld, color=%ld], ",
1141 hole->start, hole->size, hole->color);
1143 if (next->allocated)
1144 node2 = kasprintf(GFP_KERNEL,
1145 ", [%llx + %lld, color=%ld]",
1146 next->start, next->size, next->color);
1148 pr_info("%sHole [%llx - %llx, size %lld]%s\n",
1150 hole_start, hole_end, hole_end - hole_start,
1162 struct drm_mm_node node;
1163 struct list_head link;
1166 static bool evict_nodes(struct drm_mm_scan *scan,
1167 struct evict_node *nodes,
1168 unsigned int *order,
1171 struct list_head *evict_list)
1173 struct evict_node *e, *en;
1176 for (i = 0; i < count; i++) {
1177 e = &nodes[order ? order[i] : i];
1178 list_add(&e->link, evict_list);
1179 if (drm_mm_scan_add_block(scan, &e->node))
1182 list_for_each_entry_safe(e, en, evict_list, link) {
1183 if (!drm_mm_scan_remove_block(scan, &e->node))
1186 if (list_empty(evict_list)) {
1187 pr_err("Failed to find eviction: size=%lld [avail=%d], align=%lld (color=%lu)\n",
1188 scan->size, count, scan->alignment, scan->color);
1192 list_for_each_entry(e, evict_list, link)
1193 drm_mm_remove_node(&e->node);
1196 struct drm_mm_node *node;
1198 while ((node = drm_mm_scan_color_evict(scan))) {
1199 e = container_of(node, typeof(*e), node);
1200 drm_mm_remove_node(&e->node);
1201 list_add(&e->link, evict_list);
1204 if (drm_mm_scan_color_evict(scan)) {
1205 pr_err("drm_mm_scan_color_evict unexpectedly reported overlapping nodes!\n");
1213 static bool evict_nothing(struct drm_mm *mm,
1214 unsigned int total_size,
1215 struct evict_node *nodes)
1217 struct drm_mm_scan scan;
1218 LIST_HEAD(evict_list);
1219 struct evict_node *e;
1220 struct drm_mm_node *node;
1223 drm_mm_scan_init(&scan, mm, 1, 0, 0, 0);
1224 for (n = 0; n < total_size; n++) {
1226 list_add(&e->link, &evict_list);
1227 drm_mm_scan_add_block(&scan, &e->node);
1229 list_for_each_entry(e, &evict_list, link)
1230 drm_mm_scan_remove_block(&scan, &e->node);
1232 for (n = 0; n < total_size; n++) {
1235 if (!drm_mm_node_allocated(&e->node)) {
1236 pr_err("node[%d] no longer allocated!\n", n);
1240 e->link.next = NULL;
1243 drm_mm_for_each_node(node, mm) {
1244 e = container_of(node, typeof(*e), node);
1245 e->link.next = &e->link;
1248 for (n = 0; n < total_size; n++) {
1251 if (!e->link.next) {
1252 pr_err("node[%d] no longer connected!\n", n);
1257 return assert_continuous(mm, nodes[0].node.size);
1260 static bool evict_everything(struct drm_mm *mm,
1261 unsigned int total_size,
1262 struct evict_node *nodes)
1264 struct drm_mm_scan scan;
1265 LIST_HEAD(evict_list);
1266 struct evict_node *e;
1270 drm_mm_scan_init(&scan, mm, total_size, 0, 0, 0);
1271 for (n = 0; n < total_size; n++) {
1273 list_add(&e->link, &evict_list);
1274 if (drm_mm_scan_add_block(&scan, &e->node))
1279 list_for_each_entry(e, &evict_list, link) {
1280 if (!drm_mm_scan_remove_block(&scan, &e->node)) {
1282 pr_err("Node %lld not marked for eviction!\n",
1291 list_for_each_entry(e, &evict_list, link)
1292 drm_mm_remove_node(&e->node);
1294 if (!assert_one_hole(mm, 0, total_size))
1297 list_for_each_entry(e, &evict_list, link) {
1298 err = drm_mm_reserve_node(mm, &e->node);
1300 pr_err("Failed to reinsert node after eviction: start=%llx\n",
1306 return assert_continuous(mm, nodes[0].node.size);
1309 static int evict_something(struct drm_mm *mm,
1310 u64 range_start, u64 range_end,
1311 struct evict_node *nodes,
1312 unsigned int *order,
1315 unsigned int alignment,
1316 const struct insert_mode *mode)
1318 struct drm_mm_scan scan;
1319 LIST_HEAD(evict_list);
1320 struct evict_node *e;
1321 struct drm_mm_node tmp;
1324 drm_mm_scan_init_with_range(&scan, mm,
1326 range_start, range_end,
1328 if (!evict_nodes(&scan,
1329 nodes, order, count, false,
1333 memset(&tmp, 0, sizeof(tmp));
1334 err = drm_mm_insert_node_generic(mm, &tmp, size, alignment, 0,
1335 DRM_MM_INSERT_EVICT);
1337 pr_err("Failed to insert into eviction hole: size=%d, align=%d\n",
1344 if (tmp.start < range_start || tmp.start + tmp.size > range_end) {
1345 pr_err("Inserted [address=%llu + %llu] did not fit into the request range [%llu, %llu]\n",
1346 tmp.start, tmp.size, range_start, range_end);
1350 if (!assert_node(&tmp, mm, size, alignment, 0) ||
1351 drm_mm_hole_follows(&tmp)) {
1352 pr_err("Inserted did not fill the eviction hole: size=%lld [%d], align=%d [rem=%lld], start=%llx, hole-follows?=%d\n",
1354 alignment, misalignment(&tmp, alignment),
1355 tmp.start, drm_mm_hole_follows(&tmp));
1359 drm_mm_remove_node(&tmp);
1363 list_for_each_entry(e, &evict_list, link) {
1364 err = drm_mm_reserve_node(mm, &e->node);
1366 pr_err("Failed to reinsert node after eviction: start=%llx\n",
1372 if (!assert_continuous(mm, nodes[0].node.size)) {
1373 pr_err("range is no longer continuous\n");
1380 static int igt_evict(void *ignored)
1382 DRM_RND_STATE(prng, random_seed);
1383 const unsigned int size = 8192;
1384 const struct insert_mode *mode;
1386 struct evict_node *nodes;
1387 struct drm_mm_node *node, *next;
1388 unsigned int *order, n;
1391 /* Here we populate a full drm_mm and then try and insert a new node
1392 * by evicting other nodes in a random order. The drm_mm_scan should
1393 * pick the first matching hole it finds from the random list. We
1394 * repeat that for different allocation strategies, alignments and
1395 * sizes to try and stress the hole finder.
1399 nodes = vzalloc(size * sizeof(*nodes));
1403 order = drm_random_order(size, &prng);
1408 drm_mm_init(&mm, 0, size);
1409 for (n = 0; n < size; n++) {
1410 err = drm_mm_insert_node(&mm, &nodes[n].node, 1);
1412 pr_err("insert failed, step %d\n", n);
1418 /* First check that using the scanner doesn't break the mm */
1419 if (!evict_nothing(&mm, size, nodes)) {
1420 pr_err("evict_nothing() failed\n");
1423 if (!evict_everything(&mm, size, nodes)) {
1424 pr_err("evict_everything() failed\n");
1428 for (mode = evict_modes; mode->name; mode++) {
1429 for (n = 1; n <= size; n <<= 1) {
1430 drm_random_reorder(order, size, &prng);
1431 err = evict_something(&mm, 0, U64_MAX,
1436 pr_err("%s evict_something(size=%u) failed\n",
1443 for (n = 1; n < size; n <<= 1) {
1444 drm_random_reorder(order, size, &prng);
1445 err = evict_something(&mm, 0, U64_MAX,
1450 pr_err("%s evict_something(size=%u, alignment=%u) failed\n",
1451 mode->name, size/2, n);
1457 for_each_prime_number_from(n, 1, min(size, max_prime)) {
1458 unsigned int nsize = (size - n + 1) / 2;
1460 DRM_MM_BUG_ON(!nsize);
1462 drm_random_reorder(order, size, &prng);
1463 err = evict_something(&mm, 0, U64_MAX,
1468 pr_err("%s evict_something(size=%u, alignment=%u) failed\n",
1469 mode->name, nsize, n);
1478 drm_mm_for_each_node_safe(node, next, &mm)
1479 drm_mm_remove_node(node);
1480 drm_mm_takedown(&mm);
1488 static int igt_evict_range(void *ignored)
1490 DRM_RND_STATE(prng, random_seed);
1491 const unsigned int size = 8192;
1492 const unsigned int range_size = size / 2;
1493 const unsigned int range_start = size / 4;
1494 const unsigned int range_end = range_start + range_size;
1495 const struct insert_mode *mode;
1497 struct evict_node *nodes;
1498 struct drm_mm_node *node, *next;
1499 unsigned int *order, n;
1502 /* Like igt_evict() but now we are limiting the search to a
1503 * small portion of the full drm_mm.
1507 nodes = vzalloc(size * sizeof(*nodes));
1511 order = drm_random_order(size, &prng);
1516 drm_mm_init(&mm, 0, size);
1517 for (n = 0; n < size; n++) {
1518 err = drm_mm_insert_node(&mm, &nodes[n].node, 1);
1520 pr_err("insert failed, step %d\n", n);
1526 for (mode = evict_modes; mode->name; mode++) {
1527 for (n = 1; n <= range_size; n <<= 1) {
1528 drm_random_reorder(order, size, &prng);
1529 err = evict_something(&mm, range_start, range_end,
1534 pr_err("%s evict_something(size=%u) failed with range [%u, %u]\n",
1535 mode->name, n, range_start, range_end);
1540 for (n = 1; n <= range_size; n <<= 1) {
1541 drm_random_reorder(order, size, &prng);
1542 err = evict_something(&mm, range_start, range_end,
1547 pr_err("%s evict_something(size=%u, alignment=%u) failed with range [%u, %u]\n",
1548 mode->name, range_size/2, n, range_start, range_end);
1553 for_each_prime_number_from(n, 1, min(range_size, max_prime)) {
1554 unsigned int nsize = (range_size - n + 1) / 2;
1556 DRM_MM_BUG_ON(!nsize);
1558 drm_random_reorder(order, size, &prng);
1559 err = evict_something(&mm, range_start, range_end,
1564 pr_err("%s evict_something(size=%u, alignment=%u) failed with range [%u, %u]\n",
1565 mode->name, nsize, n, range_start, range_end);
1573 drm_mm_for_each_node_safe(node, next, &mm)
1574 drm_mm_remove_node(node);
1575 drm_mm_takedown(&mm);
1583 static unsigned int node_index(const struct drm_mm_node *node)
1585 return div64_u64(node->start, node->size);
1588 static int igt_topdown(void *ignored)
1590 const struct insert_mode *topdown = &insert_modes[TOPDOWN];
1591 DRM_RND_STATE(prng, random_seed);
1592 const unsigned int count = 8192;
1594 unsigned long *bitmap = NULL;
1596 struct drm_mm_node *nodes, *node, *next;
1597 unsigned int *order, n, m, o = 0;
1600 /* When allocating top-down, we expect to be returned a node
1601 * from a suitable hole at the top of the drm_mm. We check that
1602 * the returned node does match the highest available slot.
1606 nodes = vzalloc(count * sizeof(*nodes));
1610 bitmap = kzalloc(count / BITS_PER_LONG * sizeof(unsigned long),
1615 order = drm_random_order(count, &prng);
1620 for (size = 1; size <= 64; size <<= 1) {
1621 drm_mm_init(&mm, 0, size*count);
1622 for (n = 0; n < count; n++) {
1623 if (!expect_insert(&mm, &nodes[n],
1626 pr_err("insert failed, size %u step %d\n", size, n);
1630 if (drm_mm_hole_follows(&nodes[n])) {
1631 pr_err("hole after topdown insert %d, start=%llx\n, size=%u",
1632 n, nodes[n].start, size);
1636 if (!assert_one_hole(&mm, 0, size*(count - n - 1)))
1640 if (!assert_continuous(&mm, size))
1643 drm_random_reorder(order, count, &prng);
1644 for_each_prime_number_from(n, 1, min(count, max_prime)) {
1645 for (m = 0; m < n; m++) {
1646 node = &nodes[order[(o + m) % count]];
1647 drm_mm_remove_node(node);
1648 __set_bit(node_index(node), bitmap);
1651 for (m = 0; m < n; m++) {
1654 node = &nodes[order[(o + m) % count]];
1655 if (!expect_insert(&mm, node,
1658 pr_err("insert failed, step %d/%d\n", m, n);
1662 if (drm_mm_hole_follows(node)) {
1663 pr_err("hole after topdown insert %d/%d, start=%llx\n",
1668 last = find_last_bit(bitmap, count);
1669 if (node_index(node) != last) {
1670 pr_err("node %d/%d, size %d, not inserted into upmost hole, expected %d, found %d\n",
1671 m, n, size, last, node_index(node));
1675 __clear_bit(last, bitmap);
1678 DRM_MM_BUG_ON(find_first_bit(bitmap, count) != count);
1683 drm_mm_for_each_node_safe(node, next, &mm)
1684 drm_mm_remove_node(node);
1685 DRM_MM_BUG_ON(!drm_mm_clean(&mm));
1690 drm_mm_for_each_node_safe(node, next, &mm)
1691 drm_mm_remove_node(node);
1692 drm_mm_takedown(&mm);
1702 static int igt_bottomup(void *ignored)
1704 const struct insert_mode *bottomup = &insert_modes[BOTTOMUP];
1705 DRM_RND_STATE(prng, random_seed);
1706 const unsigned int count = 8192;
1708 unsigned long *bitmap;
1710 struct drm_mm_node *nodes, *node, *next;
1711 unsigned int *order, n, m, o = 0;
1714 /* Like igt_topdown, but instead of searching for the last hole,
1715 * we search for the first.
1719 nodes = vzalloc(count * sizeof(*nodes));
1723 bitmap = kzalloc(count / BITS_PER_LONG * sizeof(unsigned long),
1728 order = drm_random_order(count, &prng);
1733 for (size = 1; size <= 64; size <<= 1) {
1734 drm_mm_init(&mm, 0, size*count);
1735 for (n = 0; n < count; n++) {
1736 if (!expect_insert(&mm, &nodes[n],
1739 pr_err("bottomup insert failed, size %u step %d\n", size, n);
1743 if (!assert_one_hole(&mm, size*(n + 1), size*count))
1747 if (!assert_continuous(&mm, size))
1750 drm_random_reorder(order, count, &prng);
1751 for_each_prime_number_from(n, 1, min(count, max_prime)) {
1752 for (m = 0; m < n; m++) {
1753 node = &nodes[order[(o + m) % count]];
1754 drm_mm_remove_node(node);
1755 __set_bit(node_index(node), bitmap);
1758 for (m = 0; m < n; m++) {
1761 node = &nodes[order[(o + m) % count]];
1762 if (!expect_insert(&mm, node,
1765 pr_err("insert failed, step %d/%d\n", m, n);
1769 first = find_first_bit(bitmap, count);
1770 if (node_index(node) != first) {
1771 pr_err("node %d/%d not inserted into bottom hole, expected %d, found %d\n",
1772 m, n, first, node_index(node));
1775 __clear_bit(first, bitmap);
1778 DRM_MM_BUG_ON(find_first_bit(bitmap, count) != count);
1783 drm_mm_for_each_node_safe(node, next, &mm)
1784 drm_mm_remove_node(node);
1785 DRM_MM_BUG_ON(!drm_mm_clean(&mm));
1790 drm_mm_for_each_node_safe(node, next, &mm)
1791 drm_mm_remove_node(node);
1792 drm_mm_takedown(&mm);
1802 static void separate_adjacent_colors(const struct drm_mm_node *node,
1803 unsigned long color,
1807 if (node->allocated && node->color != color)
1810 node = list_next_entry(node, node_list);
1811 if (node->allocated && node->color != color)
1815 static bool colors_abutt(const struct drm_mm_node *node)
1817 if (!drm_mm_hole_follows(node) &&
1818 list_next_entry(node, node_list)->allocated) {
1819 pr_err("colors abutt; %ld [%llx + %llx] is next to %ld [%llx + %llx]!\n",
1820 node->color, node->start, node->size,
1821 list_next_entry(node, node_list)->color,
1822 list_next_entry(node, node_list)->start,
1823 list_next_entry(node, node_list)->size);
1830 static int igt_color(void *ignored)
1832 const unsigned int count = min(4096u, max_iterations);
1833 const struct insert_mode *mode;
1835 struct drm_mm_node *node, *nn;
1837 int ret = -EINVAL, err;
1839 /* Color adjustment complicates everything. First we just check
1840 * that when we insert a node we apply any color_adjustment callback.
1841 * The callback we use should ensure that there is a gap between
1842 * any two nodes, and so after each insertion we check that those
1843 * holes are inserted and that they are preserved.
1846 drm_mm_init(&mm, 0, U64_MAX);
1848 for (n = 1; n <= count; n++) {
1849 node = kzalloc(sizeof(*node), GFP_KERNEL);
1855 if (!expect_insert(&mm, node,
1857 &insert_modes[0])) {
1858 pr_err("insert failed, step %d\n", n);
1864 drm_mm_for_each_node_safe(node, nn, &mm) {
1865 if (node->color != node->size) {
1866 pr_err("invalid color stored: expected %lld, found %ld\n",
1867 node->size, node->color);
1872 drm_mm_remove_node(node);
1876 /* Now, let's start experimenting with applying a color callback */
1877 mm.color_adjust = separate_adjacent_colors;
1878 for (mode = insert_modes; mode->name; mode++) {
1881 node = kzalloc(sizeof(*node), GFP_KERNEL);
1887 node->size = 1 + 2*count;
1888 node->color = node->size;
1890 err = drm_mm_reserve_node(&mm, node);
1892 pr_err("initial reserve failed!\n");
1897 last = node->start + node->size;
1899 for (n = 1; n <= count; n++) {
1902 node = kzalloc(sizeof(*node), GFP_KERNEL);
1909 node->size = n + count;
1910 node->color = node->size;
1912 err = drm_mm_reserve_node(&mm, node);
1913 if (err != -ENOSPC) {
1914 pr_err("reserve %d did not report color overlap! err=%d\n",
1919 node->start += n + 1;
1920 rem = misalignment(node, n + count);
1921 node->start += n + count - rem;
1923 err = drm_mm_reserve_node(&mm, node);
1925 pr_err("reserve %d failed, err=%d\n", n, err);
1930 last = node->start + node->size;
1933 for (n = 1; n <= count; n++) {
1934 node = kzalloc(sizeof(*node), GFP_KERNEL);
1940 if (!expect_insert(&mm, node,
1943 pr_err("%s insert failed, step %d\n",
1950 drm_mm_for_each_node_safe(node, nn, &mm) {
1953 if (node->color != node->size) {
1954 pr_err("%s invalid color stored: expected %lld, found %ld\n",
1955 mode->name, node->size, node->color);
1960 if (colors_abutt(node))
1963 div64_u64_rem(node->start, node->size, &rem);
1965 pr_err("%s colored node misaligned, start=%llx expected alignment=%lld [rem=%lld]\n",
1966 mode->name, node->start, node->size, rem);
1970 drm_mm_remove_node(node);
1977 drm_mm_for_each_node_safe(node, nn, &mm) {
1978 drm_mm_remove_node(node);
1981 drm_mm_takedown(&mm);
1985 static int evict_color(struct drm_mm *mm,
1986 u64 range_start, u64 range_end,
1987 struct evict_node *nodes,
1988 unsigned int *order,
1991 unsigned int alignment,
1992 unsigned long color,
1993 const struct insert_mode *mode)
1995 struct drm_mm_scan scan;
1996 LIST_HEAD(evict_list);
1997 struct evict_node *e;
1998 struct drm_mm_node tmp;
2001 drm_mm_scan_init_with_range(&scan, mm,
2002 size, alignment, color,
2003 range_start, range_end,
2005 if (!evict_nodes(&scan,
2006 nodes, order, count, true,
2010 memset(&tmp, 0, sizeof(tmp));
2011 err = drm_mm_insert_node_generic(mm, &tmp, size, alignment, color,
2012 DRM_MM_INSERT_EVICT);
2014 pr_err("Failed to insert into eviction hole: size=%d, align=%d, color=%lu, err=%d\n",
2015 size, alignment, color, err);
2021 if (tmp.start < range_start || tmp.start + tmp.size > range_end) {
2022 pr_err("Inserted [address=%llu + %llu] did not fit into the request range [%llu, %llu]\n",
2023 tmp.start, tmp.size, range_start, range_end);
2027 if (colors_abutt(&tmp))
2030 if (!assert_node(&tmp, mm, size, alignment, color)) {
2031 pr_err("Inserted did not fit the eviction hole: size=%lld [%d], align=%d [rem=%lld], start=%llx\n",
2033 alignment, misalignment(&tmp, alignment), tmp.start);
2037 drm_mm_remove_node(&tmp);
2041 list_for_each_entry(e, &evict_list, link) {
2042 err = drm_mm_reserve_node(mm, &e->node);
2044 pr_err("Failed to reinsert node after eviction: start=%llx\n",
2053 static int igt_color_evict(void *ignored)
2055 DRM_RND_STATE(prng, random_seed);
2056 const unsigned int total_size = min(8192u, max_iterations);
2057 const struct insert_mode *mode;
2058 unsigned long color = 0;
2060 struct evict_node *nodes;
2061 struct drm_mm_node *node, *next;
2062 unsigned int *order, n;
2065 /* Check that the drm_mm_scan also honours color adjustment when
2066 * choosing its victims to create a hole. Our color_adjust does not
2067 * allow two nodes to be placed together without an intervening hole
2068 * enlarging the set of victims that must be evicted.
2072 nodes = vzalloc(total_size * sizeof(*nodes));
2076 order = drm_random_order(total_size, &prng);
2081 drm_mm_init(&mm, 0, 2*total_size - 1);
2082 mm.color_adjust = separate_adjacent_colors;
2083 for (n = 0; n < total_size; n++) {
2084 if (!expect_insert(&mm, &nodes[n].node,
2086 &insert_modes[0])) {
2087 pr_err("insert failed, step %d\n", n);
2092 for (mode = evict_modes; mode->name; mode++) {
2093 for (n = 1; n <= total_size; n <<= 1) {
2094 drm_random_reorder(order, total_size, &prng);
2095 err = evict_color(&mm, 0, U64_MAX,
2096 nodes, order, total_size,
2100 pr_err("%s evict_color(size=%u) failed\n",
2106 for (n = 1; n < total_size; n <<= 1) {
2107 drm_random_reorder(order, total_size, &prng);
2108 err = evict_color(&mm, 0, U64_MAX,
2109 nodes, order, total_size,
2110 total_size/2, n, color++,
2113 pr_err("%s evict_color(size=%u, alignment=%u) failed\n",
2114 mode->name, total_size/2, n);
2119 for_each_prime_number_from(n, 1, min(total_size, max_prime)) {
2120 unsigned int nsize = (total_size - n + 1) / 2;
2122 DRM_MM_BUG_ON(!nsize);
2124 drm_random_reorder(order, total_size, &prng);
2125 err = evict_color(&mm, 0, U64_MAX,
2126 nodes, order, total_size,
2130 pr_err("%s evict_color(size=%u, alignment=%u) failed\n",
2131 mode->name, nsize, n);
2141 drm_mm_for_each_node_safe(node, next, &mm)
2142 drm_mm_remove_node(node);
2143 drm_mm_takedown(&mm);
2151 static int igt_color_evict_range(void *ignored)
2153 DRM_RND_STATE(prng, random_seed);
2154 const unsigned int total_size = 8192;
2155 const unsigned int range_size = total_size / 2;
2156 const unsigned int range_start = total_size / 4;
2157 const unsigned int range_end = range_start + range_size;
2158 const struct insert_mode *mode;
2159 unsigned long color = 0;
2161 struct evict_node *nodes;
2162 struct drm_mm_node *node, *next;
2163 unsigned int *order, n;
2166 /* Like igt_color_evict(), but limited to small portion of the full
2171 nodes = vzalloc(total_size * sizeof(*nodes));
2175 order = drm_random_order(total_size, &prng);
2180 drm_mm_init(&mm, 0, 2*total_size - 1);
2181 mm.color_adjust = separate_adjacent_colors;
2182 for (n = 0; n < total_size; n++) {
2183 if (!expect_insert(&mm, &nodes[n].node,
2185 &insert_modes[0])) {
2186 pr_err("insert failed, step %d\n", n);
2191 for (mode = evict_modes; mode->name; mode++) {
2192 for (n = 1; n <= range_size; n <<= 1) {
2193 drm_random_reorder(order, range_size, &prng);
2194 err = evict_color(&mm, range_start, range_end,
2195 nodes, order, total_size,
2199 pr_err("%s evict_color(size=%u) failed for range [%x, %x]\n",
2200 mode->name, n, range_start, range_end);
2205 for (n = 1; n < range_size; n <<= 1) {
2206 drm_random_reorder(order, total_size, &prng);
2207 err = evict_color(&mm, range_start, range_end,
2208 nodes, order, total_size,
2209 range_size/2, n, color++,
2212 pr_err("%s evict_color(size=%u, alignment=%u) failed for range [%x, %x]\n",
2213 mode->name, total_size/2, n, range_start, range_end);
2218 for_each_prime_number_from(n, 1, min(range_size, max_prime)) {
2219 unsigned int nsize = (range_size - n + 1) / 2;
2221 DRM_MM_BUG_ON(!nsize);
2223 drm_random_reorder(order, total_size, &prng);
2224 err = evict_color(&mm, range_start, range_end,
2225 nodes, order, total_size,
2229 pr_err("%s evict_color(size=%u, alignment=%u) failed for range [%x, %x]\n",
2230 mode->name, nsize, n, range_start, range_end);
2240 drm_mm_for_each_node_safe(node, next, &mm)
2241 drm_mm_remove_node(node);
2242 drm_mm_takedown(&mm);
2250 #include "drm_selftest.c"
2252 static int __init test_drm_mm_init(void)
2256 while (!random_seed)
2257 random_seed = get_random_int();
2259 pr_info("Testing DRM range manger (struct drm_mm), with random_seed=0x%x max_iterations=%u max_prime=%u\n",
2260 random_seed, max_iterations, max_prime);
2261 err = run_selftests(selftests, ARRAY_SIZE(selftests), NULL);
2263 return err > 0 ? 0 : err;
2266 static void __exit test_drm_mm_exit(void)
2270 module_init(test_drm_mm_init);
2271 module_exit(test_drm_mm_exit);
2273 module_param(random_seed, uint, 0400);
2274 module_param(max_iterations, uint, 0400);
2275 module_param(max_prime, uint, 0400);
2277 MODULE_AUTHOR("Intel Corporation");
2278 MODULE_LICENSE("GPL");