1 // SPDX-License-Identifier: GPL-2.0
3 * HMM stands for Heterogeneous Memory Management, it is a helper layer inside
4 * the linux kernel to help device drivers mirror a process address space in
5 * the device. This allows the device to use the same address space which
6 * makes communication and data exchange a lot easier.
8 * This framework's sole purpose is to exercise various code paths inside
9 * the kernel to make sure that HMM performs as expected and to flush out any
13 #include "../kselftest_harness.h"
24 #include <sys/types.h>
27 #include <sys/ioctl.h>
29 #include "./local_config.h"
30 #ifdef LOCAL_CONFIG_HAVE_LIBHUGETLBFS
31 #include <hugetlbfs.h>
35 * This is a private UAPI to the kernel test module so it isn't exported
36 * in the usual include/uapi/... directory.
38 #include "../../../../lib/test_hmm_uapi.h"
39 #include "../../../../mm/gup_test.h"
51 HMM_PRIVATE_DEVICE_ONE,
52 HMM_PRIVATE_DEVICE_TWO,
53 HMM_COHERENCE_DEVICE_ONE,
54 HMM_COHERENCE_DEVICE_TWO,
57 #define TWOMEG (1 << 21)
58 #define HMM_BUFFER_SIZE (1024 << 12)
59 #define HMM_PATH_MAX 64
62 #define ALIGN(x, a) (((x) + (a - 1)) & (~((a) - 1)))
63 /* Just the flags we need, copied from mm.h: */
64 #define FOLL_WRITE 0x01 /* check pte is writable */
65 #define FOLL_LONGTERM 0x10000 /* mapping lifetime is indefinite */
70 unsigned int page_size;
71 unsigned int page_shift;
79 FIXTURE_VARIANT_ADD(hmm, hmm_device_private)
81 .device_number = HMM_PRIVATE_DEVICE_ONE,
84 FIXTURE_VARIANT_ADD(hmm, hmm_device_coherent)
86 .device_number = HMM_COHERENCE_DEVICE_ONE,
93 unsigned int page_size;
94 unsigned int page_shift;
103 FIXTURE_VARIANT_ADD(hmm2, hmm2_device_private)
105 .device_number0 = HMM_PRIVATE_DEVICE_ONE,
106 .device_number1 = HMM_PRIVATE_DEVICE_TWO,
109 FIXTURE_VARIANT_ADD(hmm2, hmm2_device_coherent)
111 .device_number0 = HMM_COHERENCE_DEVICE_ONE,
112 .device_number1 = HMM_COHERENCE_DEVICE_TWO,
115 static int hmm_open(int unit)
117 char pathname[HMM_PATH_MAX];
120 snprintf(pathname, sizeof(pathname), "/dev/hmm_dmirror%d", unit);
121 fd = open(pathname, O_RDWR, 0);
123 fprintf(stderr, "could not open hmm dmirror driver (%s)\n",
128 static bool hmm_is_coherent_type(int dev_num)
130 return (dev_num >= HMM_COHERENCE_DEVICE_ONE);
135 self->page_size = sysconf(_SC_PAGE_SIZE);
136 self->page_shift = ffs(self->page_size) - 1;
138 self->fd = hmm_open(variant->device_number);
139 if (self->fd < 0 && hmm_is_coherent_type(variant->device_number))
140 SKIP(exit(0), "DEVICE_COHERENT not available");
141 ASSERT_GE(self->fd, 0);
146 self->page_size = sysconf(_SC_PAGE_SIZE);
147 self->page_shift = ffs(self->page_size) - 1;
149 self->fd0 = hmm_open(variant->device_number0);
150 if (self->fd0 < 0 && hmm_is_coherent_type(variant->device_number0))
151 SKIP(exit(0), "DEVICE_COHERENT not available");
152 ASSERT_GE(self->fd0, 0);
153 self->fd1 = hmm_open(variant->device_number1);
154 ASSERT_GE(self->fd1, 0);
157 FIXTURE_TEARDOWN(hmm)
159 int ret = close(self->fd);
165 FIXTURE_TEARDOWN(hmm2)
167 int ret = close(self->fd0);
172 ret = close(self->fd1);
177 static int hmm_dmirror_cmd(int fd,
178 unsigned long request,
179 struct hmm_buffer *buffer,
180 unsigned long npages)
182 struct hmm_dmirror_cmd cmd;
185 /* Simulate a device reading system memory. */
186 cmd.addr = (__u64)buffer->ptr;
187 cmd.ptr = (__u64)buffer->mirror;
191 ret = ioctl(fd, request, &cmd);
198 buffer->cpages = cmd.cpages;
199 buffer->faults = cmd.faults;
204 static void hmm_buffer_free(struct hmm_buffer *buffer)
210 munmap(buffer->ptr, buffer->size);
211 free(buffer->mirror);
216 * Create a temporary file that will be deleted on close.
218 static int hmm_create_file(unsigned long size)
220 char path[HMM_PATH_MAX];
223 strcpy(path, "/tmp");
224 fd = open(path, O_TMPFILE | O_EXCL | O_RDWR, 0600);
229 r = ftruncate(fd, size);
230 } while (r == -1 && errno == EINTR);
239 * Return a random unsigned number.
241 static unsigned int hmm_random(void)
247 fd = open("/dev/urandom", O_RDONLY);
249 fprintf(stderr, "%s:%d failed to open /dev/urandom\n",
254 read(fd, &r, sizeof(r));
258 static void hmm_nanosleep(unsigned int n)
267 static int hmm_migrate_sys_to_dev(int fd,
268 struct hmm_buffer *buffer,
269 unsigned long npages)
271 return hmm_dmirror_cmd(fd, HMM_DMIRROR_MIGRATE_TO_DEV, buffer, npages);
274 static int hmm_migrate_dev_to_sys(int fd,
275 struct hmm_buffer *buffer,
276 unsigned long npages)
278 return hmm_dmirror_cmd(fd, HMM_DMIRROR_MIGRATE_TO_SYS, buffer, npages);
282 * Simple NULL test of device open/close.
284 TEST_F(hmm, open_close)
289 * Read private anonymous memory.
291 TEST_F(hmm, anon_read)
293 struct hmm_buffer *buffer;
294 unsigned long npages;
301 npages = ALIGN(HMM_BUFFER_SIZE, self->page_size) >> self->page_shift;
302 ASSERT_NE(npages, 0);
303 size = npages << self->page_shift;
305 buffer = malloc(sizeof(*buffer));
306 ASSERT_NE(buffer, NULL);
310 buffer->mirror = malloc(size);
311 ASSERT_NE(buffer->mirror, NULL);
313 buffer->ptr = mmap(NULL, size,
314 PROT_READ | PROT_WRITE,
315 MAP_PRIVATE | MAP_ANONYMOUS,
317 ASSERT_NE(buffer->ptr, MAP_FAILED);
320 * Initialize buffer in system memory but leave the first two pages
321 * zero (pte_none and pfn_zero).
323 i = 2 * self->page_size / sizeof(*ptr);
324 for (ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
327 /* Set buffer permission to read-only. */
328 ret = mprotect(buffer->ptr, size, PROT_READ);
331 /* Populate the CPU page table with a special zero page. */
332 val = *(int *)(buffer->ptr + self->page_size);
335 /* Simulate a device reading system memory. */
336 ret = hmm_dmirror_cmd(self->fd, HMM_DMIRROR_READ, buffer, npages);
338 ASSERT_EQ(buffer->cpages, npages);
339 ASSERT_EQ(buffer->faults, 1);
341 /* Check what the device read. */
342 ptr = buffer->mirror;
343 for (i = 0; i < 2 * self->page_size / sizeof(*ptr); ++i)
344 ASSERT_EQ(ptr[i], 0);
345 for (; i < size / sizeof(*ptr); ++i)
346 ASSERT_EQ(ptr[i], i);
348 hmm_buffer_free(buffer);
352 * Read private anonymous memory which has been protected with
353 * mprotect() PROT_NONE.
355 TEST_F(hmm, anon_read_prot)
357 struct hmm_buffer *buffer;
358 unsigned long npages;
364 npages = ALIGN(HMM_BUFFER_SIZE, self->page_size) >> self->page_shift;
365 ASSERT_NE(npages, 0);
366 size = npages << self->page_shift;
368 buffer = malloc(sizeof(*buffer));
369 ASSERT_NE(buffer, NULL);
373 buffer->mirror = malloc(size);
374 ASSERT_NE(buffer->mirror, NULL);
376 buffer->ptr = mmap(NULL, size,
377 PROT_READ | PROT_WRITE,
378 MAP_PRIVATE | MAP_ANONYMOUS,
380 ASSERT_NE(buffer->ptr, MAP_FAILED);
382 /* Initialize buffer in system memory. */
383 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
386 /* Initialize mirror buffer so we can verify it isn't written. */
387 for (i = 0, ptr = buffer->mirror; i < size / sizeof(*ptr); ++i)
390 /* Protect buffer from reading. */
391 ret = mprotect(buffer->ptr, size, PROT_NONE);
394 /* Simulate a device reading system memory. */
395 ret = hmm_dmirror_cmd(self->fd, HMM_DMIRROR_READ, buffer, npages);
396 ASSERT_EQ(ret, -EFAULT);
398 /* Allow CPU to read the buffer so we can check it. */
399 ret = mprotect(buffer->ptr, size, PROT_READ);
401 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
402 ASSERT_EQ(ptr[i], i);
404 /* Check what the device read. */
405 for (i = 0, ptr = buffer->mirror; i < size / sizeof(*ptr); ++i)
406 ASSERT_EQ(ptr[i], -i);
408 hmm_buffer_free(buffer);
412 * Write private anonymous memory.
414 TEST_F(hmm, anon_write)
416 struct hmm_buffer *buffer;
417 unsigned long npages;
423 npages = ALIGN(HMM_BUFFER_SIZE, self->page_size) >> self->page_shift;
424 ASSERT_NE(npages, 0);
425 size = npages << self->page_shift;
427 buffer = malloc(sizeof(*buffer));
428 ASSERT_NE(buffer, NULL);
432 buffer->mirror = malloc(size);
433 ASSERT_NE(buffer->mirror, NULL);
435 buffer->ptr = mmap(NULL, size,
436 PROT_READ | PROT_WRITE,
437 MAP_PRIVATE | MAP_ANONYMOUS,
439 ASSERT_NE(buffer->ptr, MAP_FAILED);
441 /* Initialize data that the device will write to buffer->ptr. */
442 for (i = 0, ptr = buffer->mirror; i < size / sizeof(*ptr); ++i)
445 /* Simulate a device writing system memory. */
446 ret = hmm_dmirror_cmd(self->fd, HMM_DMIRROR_WRITE, buffer, npages);
448 ASSERT_EQ(buffer->cpages, npages);
449 ASSERT_EQ(buffer->faults, 1);
451 /* Check what the device wrote. */
452 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
453 ASSERT_EQ(ptr[i], i);
455 hmm_buffer_free(buffer);
459 * Write private anonymous memory which has been protected with
460 * mprotect() PROT_READ.
462 TEST_F(hmm, anon_write_prot)
464 struct hmm_buffer *buffer;
465 unsigned long npages;
471 npages = ALIGN(HMM_BUFFER_SIZE, self->page_size) >> self->page_shift;
472 ASSERT_NE(npages, 0);
473 size = npages << self->page_shift;
475 buffer = malloc(sizeof(*buffer));
476 ASSERT_NE(buffer, NULL);
480 buffer->mirror = malloc(size);
481 ASSERT_NE(buffer->mirror, NULL);
483 buffer->ptr = mmap(NULL, size,
485 MAP_PRIVATE | MAP_ANONYMOUS,
487 ASSERT_NE(buffer->ptr, MAP_FAILED);
489 /* Simulate a device reading a zero page of memory. */
490 ret = hmm_dmirror_cmd(self->fd, HMM_DMIRROR_READ, buffer, 1);
492 ASSERT_EQ(buffer->cpages, 1);
493 ASSERT_EQ(buffer->faults, 1);
495 /* Initialize data that the device will write to buffer->ptr. */
496 for (i = 0, ptr = buffer->mirror; i < size / sizeof(*ptr); ++i)
499 /* Simulate a device writing system memory. */
500 ret = hmm_dmirror_cmd(self->fd, HMM_DMIRROR_WRITE, buffer, npages);
501 ASSERT_EQ(ret, -EPERM);
503 /* Check what the device wrote. */
504 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
505 ASSERT_EQ(ptr[i], 0);
507 /* Now allow writing and see that the zero page is replaced. */
508 ret = mprotect(buffer->ptr, size, PROT_WRITE | PROT_READ);
511 /* Simulate a device writing system memory. */
512 ret = hmm_dmirror_cmd(self->fd, HMM_DMIRROR_WRITE, buffer, npages);
514 ASSERT_EQ(buffer->cpages, npages);
515 ASSERT_EQ(buffer->faults, 1);
517 /* Check what the device wrote. */
518 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
519 ASSERT_EQ(ptr[i], i);
521 hmm_buffer_free(buffer);
525 * Check that a device writing an anonymous private mapping
526 * will copy-on-write if a child process inherits the mapping.
528 TEST_F(hmm, anon_write_child)
530 struct hmm_buffer *buffer;
531 unsigned long npages;
539 npages = ALIGN(HMM_BUFFER_SIZE, self->page_size) >> self->page_shift;
540 ASSERT_NE(npages, 0);
541 size = npages << self->page_shift;
543 buffer = malloc(sizeof(*buffer));
544 ASSERT_NE(buffer, NULL);
548 buffer->mirror = malloc(size);
549 ASSERT_NE(buffer->mirror, NULL);
551 buffer->ptr = mmap(NULL, size,
552 PROT_READ | PROT_WRITE,
553 MAP_PRIVATE | MAP_ANONYMOUS,
555 ASSERT_NE(buffer->ptr, MAP_FAILED);
557 /* Initialize buffer->ptr so we can tell if it is written. */
558 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
561 /* Initialize data that the device will write to buffer->ptr. */
562 for (i = 0, ptr = buffer->mirror; i < size / sizeof(*ptr); ++i)
569 waitpid(pid, &ret, 0);
570 ASSERT_EQ(WIFEXITED(ret), 1);
572 /* Check that the parent's buffer did not change. */
573 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
574 ASSERT_EQ(ptr[i], i);
578 /* Check that we see the parent's values. */
579 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
580 ASSERT_EQ(ptr[i], i);
581 for (i = 0, ptr = buffer->mirror; i < size / sizeof(*ptr); ++i)
582 ASSERT_EQ(ptr[i], -i);
584 /* The child process needs its own mirror to its own mm. */
585 child_fd = hmm_open(0);
586 ASSERT_GE(child_fd, 0);
588 /* Simulate a device writing system memory. */
589 ret = hmm_dmirror_cmd(child_fd, HMM_DMIRROR_WRITE, buffer, npages);
591 ASSERT_EQ(buffer->cpages, npages);
592 ASSERT_EQ(buffer->faults, 1);
594 /* Check what the device wrote. */
595 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
596 ASSERT_EQ(ptr[i], -i);
603 * Check that a device writing an anonymous shared mapping
604 * will not copy-on-write if a child process inherits the mapping.
606 TEST_F(hmm, anon_write_child_shared)
608 struct hmm_buffer *buffer;
609 unsigned long npages;
617 npages = ALIGN(HMM_BUFFER_SIZE, self->page_size) >> self->page_shift;
618 ASSERT_NE(npages, 0);
619 size = npages << self->page_shift;
621 buffer = malloc(sizeof(*buffer));
622 ASSERT_NE(buffer, NULL);
626 buffer->mirror = malloc(size);
627 ASSERT_NE(buffer->mirror, NULL);
629 buffer->ptr = mmap(NULL, size,
630 PROT_READ | PROT_WRITE,
631 MAP_SHARED | MAP_ANONYMOUS,
633 ASSERT_NE(buffer->ptr, MAP_FAILED);
635 /* Initialize buffer->ptr so we can tell if it is written. */
636 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
639 /* Initialize data that the device will write to buffer->ptr. */
640 for (i = 0, ptr = buffer->mirror; i < size / sizeof(*ptr); ++i)
647 waitpid(pid, &ret, 0);
648 ASSERT_EQ(WIFEXITED(ret), 1);
650 /* Check that the parent's buffer did change. */
651 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
652 ASSERT_EQ(ptr[i], -i);
656 /* Check that we see the parent's values. */
657 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
658 ASSERT_EQ(ptr[i], i);
659 for (i = 0, ptr = buffer->mirror; i < size / sizeof(*ptr); ++i)
660 ASSERT_EQ(ptr[i], -i);
662 /* The child process needs its own mirror to its own mm. */
663 child_fd = hmm_open(0);
664 ASSERT_GE(child_fd, 0);
666 /* Simulate a device writing system memory. */
667 ret = hmm_dmirror_cmd(child_fd, HMM_DMIRROR_WRITE, buffer, npages);
669 ASSERT_EQ(buffer->cpages, npages);
670 ASSERT_EQ(buffer->faults, 1);
672 /* Check what the device wrote. */
673 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
674 ASSERT_EQ(ptr[i], -i);
681 * Write private anonymous huge page.
683 TEST_F(hmm, anon_write_huge)
685 struct hmm_buffer *buffer;
686 unsigned long npages;
696 buffer = malloc(sizeof(*buffer));
697 ASSERT_NE(buffer, NULL);
701 buffer->mirror = malloc(size);
702 ASSERT_NE(buffer->mirror, NULL);
704 buffer->ptr = mmap(NULL, size,
705 PROT_READ | PROT_WRITE,
706 MAP_PRIVATE | MAP_ANONYMOUS,
708 ASSERT_NE(buffer->ptr, MAP_FAILED);
711 npages = size >> self->page_shift;
712 map = (void *)ALIGN((uintptr_t)buffer->ptr, size);
713 ret = madvise(map, size, MADV_HUGEPAGE);
715 old_ptr = buffer->ptr;
718 /* Initialize data that the device will write to buffer->ptr. */
719 for (i = 0, ptr = buffer->mirror; i < size / sizeof(*ptr); ++i)
722 /* Simulate a device writing system memory. */
723 ret = hmm_dmirror_cmd(self->fd, HMM_DMIRROR_WRITE, buffer, npages);
725 ASSERT_EQ(buffer->cpages, npages);
726 ASSERT_EQ(buffer->faults, 1);
728 /* Check what the device wrote. */
729 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
730 ASSERT_EQ(ptr[i], i);
732 buffer->ptr = old_ptr;
733 hmm_buffer_free(buffer);
736 #ifdef LOCAL_CONFIG_HAVE_LIBHUGETLBFS
738 * Write huge TLBFS page.
740 TEST_F(hmm, anon_write_hugetlbfs)
742 struct hmm_buffer *buffer;
743 unsigned long npages;
751 /* Skip test if we can't allocate a hugetlbfs page. */
753 n = gethugepagesizes(pagesizes, 4);
755 SKIP(return, "Huge page size could not be determined");
756 for (idx = 0; --n > 0; ) {
757 if (pagesizes[n] < pagesizes[idx])
760 size = ALIGN(TWOMEG, pagesizes[idx]);
761 npages = size >> self->page_shift;
763 buffer = malloc(sizeof(*buffer));
764 ASSERT_NE(buffer, NULL);
766 buffer->ptr = get_hugepage_region(size, GHR_STRICT);
767 if (buffer->ptr == NULL) {
769 SKIP(return, "Huge page could not be allocated");
774 buffer->mirror = malloc(size);
775 ASSERT_NE(buffer->mirror, NULL);
777 /* Initialize data that the device will write to buffer->ptr. */
778 for (i = 0, ptr = buffer->mirror; i < size / sizeof(*ptr); ++i)
781 /* Simulate a device writing system memory. */
782 ret = hmm_dmirror_cmd(self->fd, HMM_DMIRROR_WRITE, buffer, npages);
784 ASSERT_EQ(buffer->cpages, npages);
785 ASSERT_EQ(buffer->faults, 1);
787 /* Check what the device wrote. */
788 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
789 ASSERT_EQ(ptr[i], i);
791 free_hugepage_region(buffer->ptr);
793 hmm_buffer_free(buffer);
795 #endif /* LOCAL_CONFIG_HAVE_LIBHUGETLBFS */
798 * Read mmap'ed file memory.
800 TEST_F(hmm, file_read)
802 struct hmm_buffer *buffer;
803 unsigned long npages;
811 npages = ALIGN(HMM_BUFFER_SIZE, self->page_size) >> self->page_shift;
812 ASSERT_NE(npages, 0);
813 size = npages << self->page_shift;
815 fd = hmm_create_file(size);
818 buffer = malloc(sizeof(*buffer));
819 ASSERT_NE(buffer, NULL);
823 buffer->mirror = malloc(size);
824 ASSERT_NE(buffer->mirror, NULL);
826 /* Write initial contents of the file. */
827 for (i = 0, ptr = buffer->mirror; i < size / sizeof(*ptr); ++i)
829 len = pwrite(fd, buffer->mirror, size, 0);
830 ASSERT_EQ(len, size);
831 memset(buffer->mirror, 0, size);
833 buffer->ptr = mmap(NULL, size,
837 ASSERT_NE(buffer->ptr, MAP_FAILED);
839 /* Simulate a device reading system memory. */
840 ret = hmm_dmirror_cmd(self->fd, HMM_DMIRROR_READ, buffer, npages);
842 ASSERT_EQ(buffer->cpages, npages);
843 ASSERT_EQ(buffer->faults, 1);
845 /* Check what the device read. */
846 for (i = 0, ptr = buffer->mirror; i < size / sizeof(*ptr); ++i)
847 ASSERT_EQ(ptr[i], i);
849 hmm_buffer_free(buffer);
853 * Write mmap'ed file memory.
855 TEST_F(hmm, file_write)
857 struct hmm_buffer *buffer;
858 unsigned long npages;
866 npages = ALIGN(HMM_BUFFER_SIZE, self->page_size) >> self->page_shift;
867 ASSERT_NE(npages, 0);
868 size = npages << self->page_shift;
870 fd = hmm_create_file(size);
873 buffer = malloc(sizeof(*buffer));
874 ASSERT_NE(buffer, NULL);
878 buffer->mirror = malloc(size);
879 ASSERT_NE(buffer->mirror, NULL);
881 buffer->ptr = mmap(NULL, size,
882 PROT_READ | PROT_WRITE,
885 ASSERT_NE(buffer->ptr, MAP_FAILED);
887 /* Initialize data that the device will write to buffer->ptr. */
888 for (i = 0, ptr = buffer->mirror; i < size / sizeof(*ptr); ++i)
891 /* Simulate a device writing system memory. */
892 ret = hmm_dmirror_cmd(self->fd, HMM_DMIRROR_WRITE, buffer, npages);
894 ASSERT_EQ(buffer->cpages, npages);
895 ASSERT_EQ(buffer->faults, 1);
897 /* Check what the device wrote. */
898 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
899 ASSERT_EQ(ptr[i], i);
901 /* Check that the device also wrote the file. */
902 len = pread(fd, buffer->mirror, size, 0);
903 ASSERT_EQ(len, size);
904 for (i = 0, ptr = buffer->mirror; i < size / sizeof(*ptr); ++i)
905 ASSERT_EQ(ptr[i], i);
907 hmm_buffer_free(buffer);
911 * Migrate anonymous memory to device private memory.
915 struct hmm_buffer *buffer;
916 unsigned long npages;
922 npages = ALIGN(HMM_BUFFER_SIZE, self->page_size) >> self->page_shift;
923 ASSERT_NE(npages, 0);
924 size = npages << self->page_shift;
926 buffer = malloc(sizeof(*buffer));
927 ASSERT_NE(buffer, NULL);
931 buffer->mirror = malloc(size);
932 ASSERT_NE(buffer->mirror, NULL);
934 buffer->ptr = mmap(NULL, size,
935 PROT_READ | PROT_WRITE,
936 MAP_PRIVATE | MAP_ANONYMOUS,
938 ASSERT_NE(buffer->ptr, MAP_FAILED);
940 /* Initialize buffer in system memory. */
941 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
944 /* Migrate memory to device. */
945 ret = hmm_migrate_sys_to_dev(self->fd, buffer, npages);
947 ASSERT_EQ(buffer->cpages, npages);
949 /* Check what the device read. */
950 for (i = 0, ptr = buffer->mirror; i < size / sizeof(*ptr); ++i)
951 ASSERT_EQ(ptr[i], i);
953 hmm_buffer_free(buffer);
957 * Migrate anonymous memory to device private memory and fault some of it back
958 * to system memory, then try migrating the resulting mix of system and device
959 * private memory to the device.
961 TEST_F(hmm, migrate_fault)
963 struct hmm_buffer *buffer;
964 unsigned long npages;
970 npages = ALIGN(HMM_BUFFER_SIZE, self->page_size) >> self->page_shift;
971 ASSERT_NE(npages, 0);
972 size = npages << self->page_shift;
974 buffer = malloc(sizeof(*buffer));
975 ASSERT_NE(buffer, NULL);
979 buffer->mirror = malloc(size);
980 ASSERT_NE(buffer->mirror, NULL);
982 buffer->ptr = mmap(NULL, size,
983 PROT_READ | PROT_WRITE,
984 MAP_PRIVATE | MAP_ANONYMOUS,
986 ASSERT_NE(buffer->ptr, MAP_FAILED);
988 /* Initialize buffer in system memory. */
989 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
992 /* Migrate memory to device. */
993 ret = hmm_migrate_sys_to_dev(self->fd, buffer, npages);
995 ASSERT_EQ(buffer->cpages, npages);
997 /* Check what the device read. */
998 for (i = 0, ptr = buffer->mirror; i < size / sizeof(*ptr); ++i)
999 ASSERT_EQ(ptr[i], i);
1001 /* Fault half the pages back to system memory and check them. */
1002 for (i = 0, ptr = buffer->ptr; i < size / (2 * sizeof(*ptr)); ++i)
1003 ASSERT_EQ(ptr[i], i);
1005 /* Migrate memory to the device again. */
1006 ret = hmm_migrate_sys_to_dev(self->fd, buffer, npages);
1008 ASSERT_EQ(buffer->cpages, npages);
1010 /* Check what the device read. */
1011 for (i = 0, ptr = buffer->mirror; i < size / sizeof(*ptr); ++i)
1012 ASSERT_EQ(ptr[i], i);
1014 hmm_buffer_free(buffer);
1018 * Migrate anonymous shared memory to device private memory.
1020 TEST_F(hmm, migrate_shared)
1022 struct hmm_buffer *buffer;
1023 unsigned long npages;
1027 npages = ALIGN(HMM_BUFFER_SIZE, self->page_size) >> self->page_shift;
1028 ASSERT_NE(npages, 0);
1029 size = npages << self->page_shift;
1031 buffer = malloc(sizeof(*buffer));
1032 ASSERT_NE(buffer, NULL);
1035 buffer->size = size;
1036 buffer->mirror = malloc(size);
1037 ASSERT_NE(buffer->mirror, NULL);
1039 buffer->ptr = mmap(NULL, size,
1040 PROT_READ | PROT_WRITE,
1041 MAP_SHARED | MAP_ANONYMOUS,
1043 ASSERT_NE(buffer->ptr, MAP_FAILED);
1045 /* Migrate memory to device. */
1046 ret = hmm_migrate_sys_to_dev(self->fd, buffer, npages);
1047 ASSERT_EQ(ret, -ENOENT);
1049 hmm_buffer_free(buffer);
1053 * Try to migrate various memory types to device private memory.
1055 TEST_F(hmm2, migrate_mixed)
1057 struct hmm_buffer *buffer;
1058 unsigned long npages;
1066 size = npages << self->page_shift;
1068 buffer = malloc(sizeof(*buffer));
1069 ASSERT_NE(buffer, NULL);
1072 buffer->size = size;
1073 buffer->mirror = malloc(size);
1074 ASSERT_NE(buffer->mirror, NULL);
1076 /* Reserve a range of addresses. */
1077 buffer->ptr = mmap(NULL, size,
1079 MAP_PRIVATE | MAP_ANONYMOUS,
1081 ASSERT_NE(buffer->ptr, MAP_FAILED);
1084 /* Migrating a protected area should be an error. */
1085 ret = hmm_migrate_sys_to_dev(self->fd1, buffer, npages);
1086 ASSERT_EQ(ret, -EINVAL);
1088 /* Punch a hole after the first page address. */
1089 ret = munmap(buffer->ptr + self->page_size, self->page_size);
1092 /* We expect an error if the vma doesn't cover the range. */
1093 ret = hmm_migrate_sys_to_dev(self->fd1, buffer, 3);
1094 ASSERT_EQ(ret, -EINVAL);
1096 /* Page 2 will be a read-only zero page. */
1097 ret = mprotect(buffer->ptr + 2 * self->page_size, self->page_size,
1100 ptr = (int *)(buffer->ptr + 2 * self->page_size);
1104 /* Page 3 will be read-only. */
1105 ret = mprotect(buffer->ptr + 3 * self->page_size, self->page_size,
1106 PROT_READ | PROT_WRITE);
1108 ptr = (int *)(buffer->ptr + 3 * self->page_size);
1110 ret = mprotect(buffer->ptr + 3 * self->page_size, self->page_size,
1114 /* Page 4-5 will be read-write. */
1115 ret = mprotect(buffer->ptr + 4 * self->page_size, 2 * self->page_size,
1116 PROT_READ | PROT_WRITE);
1118 ptr = (int *)(buffer->ptr + 4 * self->page_size);
1120 ptr = (int *)(buffer->ptr + 5 * self->page_size);
1123 /* Now try to migrate pages 2-5 to device 1. */
1124 buffer->ptr = p + 2 * self->page_size;
1125 ret = hmm_migrate_sys_to_dev(self->fd1, buffer, 4);
1127 ASSERT_EQ(buffer->cpages, 4);
1129 /* Page 5 won't be migrated to device 0 because it's on device 1. */
1130 buffer->ptr = p + 5 * self->page_size;
1131 ret = hmm_migrate_sys_to_dev(self->fd0, buffer, 1);
1132 ASSERT_EQ(ret, -ENOENT);
1136 hmm_buffer_free(buffer);
1140 * Migrate anonymous memory to device memory and back to system memory
1141 * multiple times. In case of private zone configuration, this is done
1142 * through fault pages accessed by CPU. In case of coherent zone configuration,
1143 * the pages from the device should be explicitly migrated back to system memory.
1144 * The reason is Coherent device zone has coherent access by CPU, therefore
1145 * it will not generate any page fault.
1147 TEST_F(hmm, migrate_multiple)
1149 struct hmm_buffer *buffer;
1150 unsigned long npages;
1157 npages = ALIGN(HMM_BUFFER_SIZE, self->page_size) >> self->page_shift;
1158 ASSERT_NE(npages, 0);
1159 size = npages << self->page_shift;
1161 for (c = 0; c < NTIMES; c++) {
1162 buffer = malloc(sizeof(*buffer));
1163 ASSERT_NE(buffer, NULL);
1166 buffer->size = size;
1167 buffer->mirror = malloc(size);
1168 ASSERT_NE(buffer->mirror, NULL);
1170 buffer->ptr = mmap(NULL, size,
1171 PROT_READ | PROT_WRITE,
1172 MAP_PRIVATE | MAP_ANONYMOUS,
1174 ASSERT_NE(buffer->ptr, MAP_FAILED);
1176 /* Initialize buffer in system memory. */
1177 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
1180 /* Migrate memory to device. */
1181 ret = hmm_migrate_sys_to_dev(self->fd, buffer, npages);
1183 ASSERT_EQ(buffer->cpages, npages);
1185 /* Check what the device read. */
1186 for (i = 0, ptr = buffer->mirror; i < size / sizeof(*ptr); ++i)
1187 ASSERT_EQ(ptr[i], i);
1189 /* Migrate back to system memory and check them. */
1190 if (hmm_is_coherent_type(variant->device_number)) {
1191 ret = hmm_migrate_dev_to_sys(self->fd, buffer, npages);
1193 ASSERT_EQ(buffer->cpages, npages);
1196 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
1197 ASSERT_EQ(ptr[i], i);
1199 hmm_buffer_free(buffer);
1204 * Read anonymous memory multiple times.
1206 TEST_F(hmm, anon_read_multiple)
1208 struct hmm_buffer *buffer;
1209 unsigned long npages;
1216 npages = ALIGN(HMM_BUFFER_SIZE, self->page_size) >> self->page_shift;
1217 ASSERT_NE(npages, 0);
1218 size = npages << self->page_shift;
1220 for (c = 0; c < NTIMES; c++) {
1221 buffer = malloc(sizeof(*buffer));
1222 ASSERT_NE(buffer, NULL);
1225 buffer->size = size;
1226 buffer->mirror = malloc(size);
1227 ASSERT_NE(buffer->mirror, NULL);
1229 buffer->ptr = mmap(NULL, size,
1230 PROT_READ | PROT_WRITE,
1231 MAP_PRIVATE | MAP_ANONYMOUS,
1233 ASSERT_NE(buffer->ptr, MAP_FAILED);
1235 /* Initialize buffer in system memory. */
1236 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
1239 /* Simulate a device reading system memory. */
1240 ret = hmm_dmirror_cmd(self->fd, HMM_DMIRROR_READ, buffer,
1243 ASSERT_EQ(buffer->cpages, npages);
1244 ASSERT_EQ(buffer->faults, 1);
1246 /* Check what the device read. */
1247 for (i = 0, ptr = buffer->mirror; i < size / sizeof(*ptr); ++i)
1248 ASSERT_EQ(ptr[i], i + c);
1250 hmm_buffer_free(buffer);
1254 void *unmap_buffer(void *p)
1256 struct hmm_buffer *buffer = p;
1258 /* Delay for a bit and then unmap buffer while it is being read. */
1259 hmm_nanosleep(hmm_random() % 32000);
1260 munmap(buffer->ptr + buffer->size / 2, buffer->size / 2);
1267 * Try reading anonymous memory while it is being unmapped.
1269 TEST_F(hmm, anon_teardown)
1271 unsigned long npages;
1276 npages = ALIGN(HMM_BUFFER_SIZE, self->page_size) >> self->page_shift;
1277 ASSERT_NE(npages, 0);
1278 size = npages << self->page_shift;
1280 for (c = 0; c < NTIMES; ++c) {
1282 struct hmm_buffer *buffer;
1287 buffer = malloc(sizeof(*buffer));
1288 ASSERT_NE(buffer, NULL);
1291 buffer->size = size;
1292 buffer->mirror = malloc(size);
1293 ASSERT_NE(buffer->mirror, NULL);
1295 buffer->ptr = mmap(NULL, size,
1296 PROT_READ | PROT_WRITE,
1297 MAP_PRIVATE | MAP_ANONYMOUS,
1299 ASSERT_NE(buffer->ptr, MAP_FAILED);
1301 /* Initialize buffer in system memory. */
1302 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
1305 rc = pthread_create(&thread, NULL, unmap_buffer, buffer);
1308 /* Simulate a device reading system memory. */
1309 rc = hmm_dmirror_cmd(self->fd, HMM_DMIRROR_READ, buffer,
1312 ASSERT_EQ(buffer->cpages, npages);
1313 ASSERT_EQ(buffer->faults, 1);
1315 /* Check what the device read. */
1316 for (i = 0, ptr = buffer->mirror;
1317 i < size / sizeof(*ptr);
1319 ASSERT_EQ(ptr[i], i + c);
1322 pthread_join(thread, &ret);
1323 hmm_buffer_free(buffer);
1328 * Test memory snapshot without faulting in pages accessed by the device.
1330 TEST_F(hmm, mixedmap)
1332 struct hmm_buffer *buffer;
1333 unsigned long npages;
1339 size = npages << self->page_shift;
1341 buffer = malloc(sizeof(*buffer));
1342 ASSERT_NE(buffer, NULL);
1345 buffer->size = size;
1346 buffer->mirror = malloc(npages);
1347 ASSERT_NE(buffer->mirror, NULL);
1350 /* Reserve a range of addresses. */
1351 buffer->ptr = mmap(NULL, size,
1352 PROT_READ | PROT_WRITE,
1355 ASSERT_NE(buffer->ptr, MAP_FAILED);
1357 /* Simulate a device snapshotting CPU pagetables. */
1358 ret = hmm_dmirror_cmd(self->fd, HMM_DMIRROR_SNAPSHOT, buffer, npages);
1360 ASSERT_EQ(buffer->cpages, npages);
1362 /* Check what the device saw. */
1364 ASSERT_EQ(m[0], HMM_DMIRROR_PROT_READ);
1366 hmm_buffer_free(buffer);
1370 * Test memory snapshot without faulting in pages accessed by the device.
1372 TEST_F(hmm2, snapshot)
1374 struct hmm_buffer *buffer;
1375 unsigned long npages;
1384 size = npages << self->page_shift;
1386 buffer = malloc(sizeof(*buffer));
1387 ASSERT_NE(buffer, NULL);
1390 buffer->size = size;
1391 buffer->mirror = malloc(npages);
1392 ASSERT_NE(buffer->mirror, NULL);
1394 /* Reserve a range of addresses. */
1395 buffer->ptr = mmap(NULL, size,
1397 MAP_PRIVATE | MAP_ANONYMOUS,
1399 ASSERT_NE(buffer->ptr, MAP_FAILED);
1402 /* Punch a hole after the first page address. */
1403 ret = munmap(buffer->ptr + self->page_size, self->page_size);
1406 /* Page 2 will be read-only zero page. */
1407 ret = mprotect(buffer->ptr + 2 * self->page_size, self->page_size,
1410 ptr = (int *)(buffer->ptr + 2 * self->page_size);
1414 /* Page 3 will be read-only. */
1415 ret = mprotect(buffer->ptr + 3 * self->page_size, self->page_size,
1416 PROT_READ | PROT_WRITE);
1418 ptr = (int *)(buffer->ptr + 3 * self->page_size);
1420 ret = mprotect(buffer->ptr + 3 * self->page_size, self->page_size,
1424 /* Page 4-6 will be read-write. */
1425 ret = mprotect(buffer->ptr + 4 * self->page_size, 3 * self->page_size,
1426 PROT_READ | PROT_WRITE);
1428 ptr = (int *)(buffer->ptr + 4 * self->page_size);
1431 /* Page 5 will be migrated to device 0. */
1432 buffer->ptr = p + 5 * self->page_size;
1433 ret = hmm_migrate_sys_to_dev(self->fd0, buffer, 1);
1435 ASSERT_EQ(buffer->cpages, 1);
1437 /* Page 6 will be migrated to device 1. */
1438 buffer->ptr = p + 6 * self->page_size;
1439 ret = hmm_migrate_sys_to_dev(self->fd1, buffer, 1);
1441 ASSERT_EQ(buffer->cpages, 1);
1443 /* Simulate a device snapshotting CPU pagetables. */
1445 ret = hmm_dmirror_cmd(self->fd0, HMM_DMIRROR_SNAPSHOT, buffer, npages);
1447 ASSERT_EQ(buffer->cpages, npages);
1449 /* Check what the device saw. */
1451 ASSERT_EQ(m[0], HMM_DMIRROR_PROT_ERROR);
1452 ASSERT_EQ(m[1], HMM_DMIRROR_PROT_ERROR);
1453 ASSERT_EQ(m[2], HMM_DMIRROR_PROT_ZERO | HMM_DMIRROR_PROT_READ);
1454 ASSERT_EQ(m[3], HMM_DMIRROR_PROT_READ);
1455 ASSERT_EQ(m[4], HMM_DMIRROR_PROT_WRITE);
1456 if (!hmm_is_coherent_type(variant->device_number0)) {
1457 ASSERT_EQ(m[5], HMM_DMIRROR_PROT_DEV_PRIVATE_LOCAL |
1458 HMM_DMIRROR_PROT_WRITE);
1459 ASSERT_EQ(m[6], HMM_DMIRROR_PROT_NONE);
1461 ASSERT_EQ(m[5], HMM_DMIRROR_PROT_DEV_COHERENT_LOCAL |
1462 HMM_DMIRROR_PROT_WRITE);
1463 ASSERT_EQ(m[6], HMM_DMIRROR_PROT_DEV_COHERENT_REMOTE |
1464 HMM_DMIRROR_PROT_WRITE);
1467 hmm_buffer_free(buffer);
1470 #ifdef LOCAL_CONFIG_HAVE_LIBHUGETLBFS
1472 * Test the hmm_range_fault() HMM_PFN_PMD flag for large pages that
1473 * should be mapped by a large page table entry.
1475 TEST_F(hmm, compound)
1477 struct hmm_buffer *buffer;
1478 unsigned long npages;
1487 /* Skip test if we can't allocate a hugetlbfs page. */
1489 n = gethugepagesizes(pagesizes, 4);
1492 for (idx = 0; --n > 0; ) {
1493 if (pagesizes[n] < pagesizes[idx])
1496 size = ALIGN(TWOMEG, pagesizes[idx]);
1497 npages = size >> self->page_shift;
1499 buffer = malloc(sizeof(*buffer));
1500 ASSERT_NE(buffer, NULL);
1502 buffer->ptr = get_hugepage_region(size, GHR_STRICT);
1503 if (buffer->ptr == NULL) {
1508 buffer->size = size;
1509 buffer->mirror = malloc(npages);
1510 ASSERT_NE(buffer->mirror, NULL);
1512 /* Initialize the pages the device will snapshot in buffer->ptr. */
1513 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
1516 /* Simulate a device snapshotting CPU pagetables. */
1517 ret = hmm_dmirror_cmd(self->fd, HMM_DMIRROR_SNAPSHOT, buffer, npages);
1519 ASSERT_EQ(buffer->cpages, npages);
1521 /* Check what the device saw. */
1523 for (i = 0; i < npages; ++i)
1524 ASSERT_EQ(m[i], HMM_DMIRROR_PROT_WRITE |
1525 HMM_DMIRROR_PROT_PMD);
1527 /* Make the region read-only. */
1528 ret = mprotect(buffer->ptr, size, PROT_READ);
1531 /* Simulate a device snapshotting CPU pagetables. */
1532 ret = hmm_dmirror_cmd(self->fd, HMM_DMIRROR_SNAPSHOT, buffer, npages);
1534 ASSERT_EQ(buffer->cpages, npages);
1536 /* Check what the device saw. */
1538 for (i = 0; i < npages; ++i)
1539 ASSERT_EQ(m[i], HMM_DMIRROR_PROT_READ |
1540 HMM_DMIRROR_PROT_PMD);
1542 free_hugepage_region(buffer->ptr);
1544 hmm_buffer_free(buffer);
1546 #endif /* LOCAL_CONFIG_HAVE_LIBHUGETLBFS */
1549 * Test two devices reading the same memory (double mapped).
1551 TEST_F(hmm2, double_map)
1553 struct hmm_buffer *buffer;
1554 unsigned long npages;
1561 size = npages << self->page_shift;
1563 buffer = malloc(sizeof(*buffer));
1564 ASSERT_NE(buffer, NULL);
1567 buffer->size = size;
1568 buffer->mirror = malloc(npages);
1569 ASSERT_NE(buffer->mirror, NULL);
1571 /* Reserve a range of addresses. */
1572 buffer->ptr = mmap(NULL, size,
1573 PROT_READ | PROT_WRITE,
1574 MAP_PRIVATE | MAP_ANONYMOUS,
1576 ASSERT_NE(buffer->ptr, MAP_FAILED);
1578 /* Initialize buffer in system memory. */
1579 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
1582 /* Make region read-only. */
1583 ret = mprotect(buffer->ptr, size, PROT_READ);
1586 /* Simulate device 0 reading system memory. */
1587 ret = hmm_dmirror_cmd(self->fd0, HMM_DMIRROR_READ, buffer, npages);
1589 ASSERT_EQ(buffer->cpages, npages);
1590 ASSERT_EQ(buffer->faults, 1);
1592 /* Check what the device read. */
1593 for (i = 0, ptr = buffer->mirror; i < size / sizeof(*ptr); ++i)
1594 ASSERT_EQ(ptr[i], i);
1596 /* Simulate device 1 reading system memory. */
1597 ret = hmm_dmirror_cmd(self->fd1, HMM_DMIRROR_READ, buffer, npages);
1599 ASSERT_EQ(buffer->cpages, npages);
1600 ASSERT_EQ(buffer->faults, 1);
1602 /* Check what the device read. */
1603 for (i = 0, ptr = buffer->mirror; i < size / sizeof(*ptr); ++i)
1604 ASSERT_EQ(ptr[i], i);
1606 /* Migrate pages to device 1 and try to read from device 0. */
1607 ret = hmm_migrate_sys_to_dev(self->fd1, buffer, npages);
1609 ASSERT_EQ(buffer->cpages, npages);
1611 ret = hmm_dmirror_cmd(self->fd0, HMM_DMIRROR_READ, buffer, npages);
1613 ASSERT_EQ(buffer->cpages, npages);
1614 ASSERT_EQ(buffer->faults, 1);
1616 /* Check what device 0 read. */
1617 for (i = 0, ptr = buffer->mirror; i < size / sizeof(*ptr); ++i)
1618 ASSERT_EQ(ptr[i], i);
1620 hmm_buffer_free(buffer);
1624 * Basic check of exclusive faulting.
1626 TEST_F(hmm, exclusive)
1628 struct hmm_buffer *buffer;
1629 unsigned long npages;
1635 npages = ALIGN(HMM_BUFFER_SIZE, self->page_size) >> self->page_shift;
1636 ASSERT_NE(npages, 0);
1637 size = npages << self->page_shift;
1639 buffer = malloc(sizeof(*buffer));
1640 ASSERT_NE(buffer, NULL);
1643 buffer->size = size;
1644 buffer->mirror = malloc(size);
1645 ASSERT_NE(buffer->mirror, NULL);
1647 buffer->ptr = mmap(NULL, size,
1648 PROT_READ | PROT_WRITE,
1649 MAP_PRIVATE | MAP_ANONYMOUS,
1651 ASSERT_NE(buffer->ptr, MAP_FAILED);
1653 /* Initialize buffer in system memory. */
1654 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
1657 /* Map memory exclusively for device access. */
1658 ret = hmm_dmirror_cmd(self->fd, HMM_DMIRROR_EXCLUSIVE, buffer, npages);
1660 ASSERT_EQ(buffer->cpages, npages);
1662 /* Check what the device read. */
1663 for (i = 0, ptr = buffer->mirror; i < size / sizeof(*ptr); ++i)
1664 ASSERT_EQ(ptr[i], i);
1666 /* Fault pages back to system memory and check them. */
1667 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
1668 ASSERT_EQ(ptr[i]++, i);
1670 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
1671 ASSERT_EQ(ptr[i], i+1);
1673 /* Check atomic access revoked */
1674 ret = hmm_dmirror_cmd(self->fd, HMM_DMIRROR_CHECK_EXCLUSIVE, buffer, npages);
1677 hmm_buffer_free(buffer);
1680 TEST_F(hmm, exclusive_mprotect)
1682 struct hmm_buffer *buffer;
1683 unsigned long npages;
1689 npages = ALIGN(HMM_BUFFER_SIZE, self->page_size) >> self->page_shift;
1690 ASSERT_NE(npages, 0);
1691 size = npages << self->page_shift;
1693 buffer = malloc(sizeof(*buffer));
1694 ASSERT_NE(buffer, NULL);
1697 buffer->size = size;
1698 buffer->mirror = malloc(size);
1699 ASSERT_NE(buffer->mirror, NULL);
1701 buffer->ptr = mmap(NULL, size,
1702 PROT_READ | PROT_WRITE,
1703 MAP_PRIVATE | MAP_ANONYMOUS,
1705 ASSERT_NE(buffer->ptr, MAP_FAILED);
1707 /* Initialize buffer in system memory. */
1708 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
1711 /* Map memory exclusively for device access. */
1712 ret = hmm_dmirror_cmd(self->fd, HMM_DMIRROR_EXCLUSIVE, buffer, npages);
1714 ASSERT_EQ(buffer->cpages, npages);
1716 /* Check what the device read. */
1717 for (i = 0, ptr = buffer->mirror; i < size / sizeof(*ptr); ++i)
1718 ASSERT_EQ(ptr[i], i);
1720 ret = mprotect(buffer->ptr, size, PROT_READ);
1723 /* Simulate a device writing system memory. */
1724 ret = hmm_dmirror_cmd(self->fd, HMM_DMIRROR_WRITE, buffer, npages);
1725 ASSERT_EQ(ret, -EPERM);
1727 hmm_buffer_free(buffer);
1731 * Check copy-on-write works.
1733 TEST_F(hmm, exclusive_cow)
1735 struct hmm_buffer *buffer;
1736 unsigned long npages;
1742 npages = ALIGN(HMM_BUFFER_SIZE, self->page_size) >> self->page_shift;
1743 ASSERT_NE(npages, 0);
1744 size = npages << self->page_shift;
1746 buffer = malloc(sizeof(*buffer));
1747 ASSERT_NE(buffer, NULL);
1750 buffer->size = size;
1751 buffer->mirror = malloc(size);
1752 ASSERT_NE(buffer->mirror, NULL);
1754 buffer->ptr = mmap(NULL, size,
1755 PROT_READ | PROT_WRITE,
1756 MAP_PRIVATE | MAP_ANONYMOUS,
1758 ASSERT_NE(buffer->ptr, MAP_FAILED);
1760 /* Initialize buffer in system memory. */
1761 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
1764 /* Map memory exclusively for device access. */
1765 ret = hmm_dmirror_cmd(self->fd, HMM_DMIRROR_EXCLUSIVE, buffer, npages);
1767 ASSERT_EQ(buffer->cpages, npages);
1771 /* Fault pages back to system memory and check them. */
1772 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
1773 ASSERT_EQ(ptr[i]++, i);
1775 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
1776 ASSERT_EQ(ptr[i], i+1);
1778 hmm_buffer_free(buffer);
1781 static int gup_test_exec(int gup_fd, unsigned long addr, int cmd,
1782 int npages, int size, int flags)
1784 struct gup_test gup = {
1785 .nr_pages_per_call = npages,
1787 .gup_flags = FOLL_WRITE | flags,
1791 if (ioctl(gup_fd, cmd, &gup)) {
1792 perror("ioctl on error\n");
1800 * Test get user device pages through gup_test. Setting PIN_LONGTERM flag.
1801 * This should trigger a migration back to system memory for both, private
1802 * and coherent type pages.
1803 * This test makes use of gup_test module. Make sure GUP_TEST_CONFIG is added
1804 * to your configuration before you run it.
1806 TEST_F(hmm, hmm_gup_test)
1808 struct hmm_buffer *buffer;
1810 unsigned long npages;
1817 gup_fd = open("/sys/kernel/debug/gup_test", O_RDWR);
1819 SKIP(return, "Skipping test, could not find gup_test driver");
1822 size = npages << self->page_shift;
1824 buffer = malloc(sizeof(*buffer));
1825 ASSERT_NE(buffer, NULL);
1828 buffer->size = size;
1829 buffer->mirror = malloc(size);
1830 ASSERT_NE(buffer->mirror, NULL);
1832 buffer->ptr = mmap(NULL, size,
1833 PROT_READ | PROT_WRITE,
1834 MAP_PRIVATE | MAP_ANONYMOUS,
1836 ASSERT_NE(buffer->ptr, MAP_FAILED);
1838 /* Initialize buffer in system memory. */
1839 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
1842 /* Migrate memory to device. */
1843 ret = hmm_migrate_sys_to_dev(self->fd, buffer, npages);
1845 ASSERT_EQ(buffer->cpages, npages);
1846 /* Check what the device read. */
1847 for (i = 0, ptr = buffer->mirror; i < size / sizeof(*ptr); ++i)
1848 ASSERT_EQ(ptr[i], i);
1850 ASSERT_EQ(gup_test_exec(gup_fd,
1851 (unsigned long)buffer->ptr,
1852 GUP_BASIC_TEST, 1, self->page_size, 0), 0);
1853 ASSERT_EQ(gup_test_exec(gup_fd,
1854 (unsigned long)buffer->ptr + 1 * self->page_size,
1855 GUP_FAST_BENCHMARK, 1, self->page_size, 0), 0);
1856 ASSERT_EQ(gup_test_exec(gup_fd,
1857 (unsigned long)buffer->ptr + 2 * self->page_size,
1858 PIN_FAST_BENCHMARK, 1, self->page_size, FOLL_LONGTERM), 0);
1859 ASSERT_EQ(gup_test_exec(gup_fd,
1860 (unsigned long)buffer->ptr + 3 * self->page_size,
1861 PIN_LONGTERM_BENCHMARK, 1, self->page_size, 0), 0);
1863 /* Take snapshot to CPU pagetables */
1864 ret = hmm_dmirror_cmd(self->fd, HMM_DMIRROR_SNAPSHOT, buffer, npages);
1866 ASSERT_EQ(buffer->cpages, npages);
1868 if (hmm_is_coherent_type(variant->device_number)) {
1869 ASSERT_EQ(HMM_DMIRROR_PROT_DEV_COHERENT_LOCAL | HMM_DMIRROR_PROT_WRITE, m[0]);
1870 ASSERT_EQ(HMM_DMIRROR_PROT_DEV_COHERENT_LOCAL | HMM_DMIRROR_PROT_WRITE, m[1]);
1872 ASSERT_EQ(HMM_DMIRROR_PROT_WRITE, m[0]);
1873 ASSERT_EQ(HMM_DMIRROR_PROT_WRITE, m[1]);
1875 ASSERT_EQ(HMM_DMIRROR_PROT_WRITE, m[2]);
1876 ASSERT_EQ(HMM_DMIRROR_PROT_WRITE, m[3]);
1878 * Check again the content on the pages. Make sure there's no
1881 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
1882 ASSERT_EQ(ptr[i], i);
1885 hmm_buffer_free(buffer);
1889 * Test copy-on-write in device pages.
1890 * In case of writing to COW private page(s), a page fault will migrate pages
1891 * back to system memory first. Then, these pages will be duplicated. In case
1892 * of COW device coherent type, pages are duplicated directly from device
1895 TEST_F(hmm, hmm_cow_in_device)
1897 struct hmm_buffer *buffer;
1898 unsigned long npages;
1908 size = npages << self->page_shift;
1910 buffer = malloc(sizeof(*buffer));
1911 ASSERT_NE(buffer, NULL);
1914 buffer->size = size;
1915 buffer->mirror = malloc(size);
1916 ASSERT_NE(buffer->mirror, NULL);
1918 buffer->ptr = mmap(NULL, size,
1919 PROT_READ | PROT_WRITE,
1920 MAP_PRIVATE | MAP_ANONYMOUS,
1922 ASSERT_NE(buffer->ptr, MAP_FAILED);
1924 /* Initialize buffer in system memory. */
1925 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
1928 /* Migrate memory to device. */
1930 ret = hmm_migrate_sys_to_dev(self->fd, buffer, npages);
1932 ASSERT_EQ(buffer->cpages, npages);
1938 /* Child process waitd for SIGTERM from the parent. */
1941 perror("Should not reach this\n");
1944 /* Parent process writes to COW pages(s) and gets a
1945 * new copy in system. In case of device private pages,
1946 * this write causes a migration to system mem first.
1948 for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
1951 /* Terminate child and wait */
1952 EXPECT_EQ(0, kill(pid, SIGTERM));
1953 EXPECT_EQ(pid, waitpid(pid, &status, 0));
1954 EXPECT_NE(0, WIFSIGNALED(status));
1955 EXPECT_EQ(SIGTERM, WTERMSIG(status));
1957 /* Take snapshot to CPU pagetables */
1958 ret = hmm_dmirror_cmd(self->fd, HMM_DMIRROR_SNAPSHOT, buffer, npages);
1960 ASSERT_EQ(buffer->cpages, npages);
1962 for (i = 0; i < npages; i++)
1963 ASSERT_EQ(HMM_DMIRROR_PROT_WRITE, m[i]);
1965 hmm_buffer_free(buffer);