Commit f70dab3c authored by Alex Sierra's avatar Alex Sierra Committed by akpm

tools: update hmm-test to support device coherent type

Test cases such as migrate_fault and migrate_multiple, were modified to
explicit migrate from device to sys memory without the need of page
faults, when using device coherent type.

Snapshot test case updated to read memory device type first and based on
that, get the proper returned results migrate_ping_pong test case added to
test explicit migration from device to sys memory for both private and
coherent zone types.

Helpers to migrate from device to sys memory and vicerversa were also
added.

Link: https://lkml.kernel.org/r/20220715150521.18165-12-alex.sierra@amd.comSigned-off-by: default avatarAlex Sierra <alex.sierra@amd.com>
Signed-off-by: default avatarChristoph Hellwig <hch@lst.de>
Acked-by: default avatarFelix Kuehling <Felix.Kuehling@amd.com>
Reviewed-by: default avatarAlistair Popple <apopple@nvidia.com>
Cc: David Hildenbrand <david@redhat.com>
Cc: Jason Gunthorpe <jgg@nvidia.com>
Cc: Jerome Glisse <jglisse@redhat.com>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: Ralph Campbell <rcampbell@nvidia.com>
Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
parent 4c2e0f76
...@@ -46,6 +46,13 @@ struct hmm_buffer { ...@@ -46,6 +46,13 @@ struct hmm_buffer {
uint64_t faults; uint64_t faults;
}; };
enum {
HMM_PRIVATE_DEVICE_ONE,
HMM_PRIVATE_DEVICE_TWO,
HMM_COHERENCE_DEVICE_ONE,
HMM_COHERENCE_DEVICE_TWO,
};
#define TWOMEG (1 << 21) #define TWOMEG (1 << 21)
#define HMM_BUFFER_SIZE (1024 << 12) #define HMM_BUFFER_SIZE (1024 << 12)
#define HMM_PATH_MAX 64 #define HMM_PATH_MAX 64
...@@ -60,6 +67,21 @@ FIXTURE(hmm) ...@@ -60,6 +67,21 @@ FIXTURE(hmm)
unsigned int page_shift; unsigned int page_shift;
}; };
FIXTURE_VARIANT(hmm)
{
int device_number;
};
FIXTURE_VARIANT_ADD(hmm, hmm_device_private)
{
.device_number = HMM_PRIVATE_DEVICE_ONE,
};
FIXTURE_VARIANT_ADD(hmm, hmm_device_coherent)
{
.device_number = HMM_COHERENCE_DEVICE_ONE,
};
FIXTURE(hmm2) FIXTURE(hmm2)
{ {
int fd0; int fd0;
...@@ -68,6 +90,24 @@ FIXTURE(hmm2) ...@@ -68,6 +90,24 @@ FIXTURE(hmm2)
unsigned int page_shift; unsigned int page_shift;
}; };
FIXTURE_VARIANT(hmm2)
{
int device_number0;
int device_number1;
};
FIXTURE_VARIANT_ADD(hmm2, hmm2_device_private)
{
.device_number0 = HMM_PRIVATE_DEVICE_ONE,
.device_number1 = HMM_PRIVATE_DEVICE_TWO,
};
FIXTURE_VARIANT_ADD(hmm2, hmm2_device_coherent)
{
.device_number0 = HMM_COHERENCE_DEVICE_ONE,
.device_number1 = HMM_COHERENCE_DEVICE_TWO,
};
static int hmm_open(int unit) static int hmm_open(int unit)
{ {
char pathname[HMM_PATH_MAX]; char pathname[HMM_PATH_MAX];
...@@ -81,12 +121,19 @@ static int hmm_open(int unit) ...@@ -81,12 +121,19 @@ static int hmm_open(int unit)
return fd; return fd;
} }
static bool hmm_is_coherent_type(int dev_num)
{
return (dev_num >= HMM_COHERENCE_DEVICE_ONE);
}
FIXTURE_SETUP(hmm) FIXTURE_SETUP(hmm)
{ {
self->page_size = sysconf(_SC_PAGE_SIZE); self->page_size = sysconf(_SC_PAGE_SIZE);
self->page_shift = ffs(self->page_size) - 1; self->page_shift = ffs(self->page_size) - 1;
self->fd = hmm_open(0); self->fd = hmm_open(variant->device_number);
if (self->fd < 0 && hmm_is_coherent_type(variant->device_number))
SKIP(exit(0), "DEVICE_COHERENT not available");
ASSERT_GE(self->fd, 0); ASSERT_GE(self->fd, 0);
} }
...@@ -95,9 +142,11 @@ FIXTURE_SETUP(hmm2) ...@@ -95,9 +142,11 @@ FIXTURE_SETUP(hmm2)
self->page_size = sysconf(_SC_PAGE_SIZE); self->page_size = sysconf(_SC_PAGE_SIZE);
self->page_shift = ffs(self->page_size) - 1; self->page_shift = ffs(self->page_size) - 1;
self->fd0 = hmm_open(0); self->fd0 = hmm_open(variant->device_number0);
if (self->fd0 < 0 && hmm_is_coherent_type(variant->device_number0))
SKIP(exit(0), "DEVICE_COHERENT not available");
ASSERT_GE(self->fd0, 0); ASSERT_GE(self->fd0, 0);
self->fd1 = hmm_open(1); self->fd1 = hmm_open(variant->device_number1);
ASSERT_GE(self->fd1, 0); ASSERT_GE(self->fd1, 0);
} }
...@@ -211,6 +260,20 @@ static void hmm_nanosleep(unsigned int n) ...@@ -211,6 +260,20 @@ static void hmm_nanosleep(unsigned int n)
nanosleep(&t, NULL); nanosleep(&t, NULL);
} }
static int hmm_migrate_sys_to_dev(int fd,
struct hmm_buffer *buffer,
unsigned long npages)
{
return hmm_dmirror_cmd(fd, HMM_DMIRROR_MIGRATE_TO_DEV, buffer, npages);
}
static int hmm_migrate_dev_to_sys(int fd,
struct hmm_buffer *buffer,
unsigned long npages)
{
return hmm_dmirror_cmd(fd, HMM_DMIRROR_MIGRATE_TO_SYS, buffer, npages);
}
/* /*
* Simple NULL test of device open/close. * Simple NULL test of device open/close.
*/ */
...@@ -875,7 +938,7 @@ TEST_F(hmm, migrate) ...@@ -875,7 +938,7 @@ TEST_F(hmm, migrate)
ptr[i] = i; ptr[i] = i;
/* Migrate memory to device. */ /* Migrate memory to device. */
ret = hmm_dmirror_cmd(self->fd, HMM_DMIRROR_MIGRATE, buffer, npages); ret = hmm_migrate_sys_to_dev(self->fd, buffer, npages);
ASSERT_EQ(ret, 0); ASSERT_EQ(ret, 0);
ASSERT_EQ(buffer->cpages, npages); ASSERT_EQ(buffer->cpages, npages);
...@@ -923,7 +986,7 @@ TEST_F(hmm, migrate_fault) ...@@ -923,7 +986,7 @@ TEST_F(hmm, migrate_fault)
ptr[i] = i; ptr[i] = i;
/* Migrate memory to device. */ /* Migrate memory to device. */
ret = hmm_dmirror_cmd(self->fd, HMM_DMIRROR_MIGRATE, buffer, npages); ret = hmm_migrate_sys_to_dev(self->fd, buffer, npages);
ASSERT_EQ(ret, 0); ASSERT_EQ(ret, 0);
ASSERT_EQ(buffer->cpages, npages); ASSERT_EQ(buffer->cpages, npages);
...@@ -936,7 +999,7 @@ TEST_F(hmm, migrate_fault) ...@@ -936,7 +999,7 @@ TEST_F(hmm, migrate_fault)
ASSERT_EQ(ptr[i], i); ASSERT_EQ(ptr[i], i);
/* Migrate memory to the device again. */ /* Migrate memory to the device again. */
ret = hmm_dmirror_cmd(self->fd, HMM_DMIRROR_MIGRATE, buffer, npages); ret = hmm_migrate_sys_to_dev(self->fd, buffer, npages);
ASSERT_EQ(ret, 0); ASSERT_EQ(ret, 0);
ASSERT_EQ(buffer->cpages, npages); ASSERT_EQ(buffer->cpages, npages);
...@@ -976,7 +1039,7 @@ TEST_F(hmm, migrate_shared) ...@@ -976,7 +1039,7 @@ TEST_F(hmm, migrate_shared)
ASSERT_NE(buffer->ptr, MAP_FAILED); ASSERT_NE(buffer->ptr, MAP_FAILED);
/* Migrate memory to device. */ /* Migrate memory to device. */
ret = hmm_dmirror_cmd(self->fd, HMM_DMIRROR_MIGRATE, buffer, npages); ret = hmm_migrate_sys_to_dev(self->fd, buffer, npages);
ASSERT_EQ(ret, -ENOENT); ASSERT_EQ(ret, -ENOENT);
hmm_buffer_free(buffer); hmm_buffer_free(buffer);
...@@ -1015,7 +1078,7 @@ TEST_F(hmm2, migrate_mixed) ...@@ -1015,7 +1078,7 @@ TEST_F(hmm2, migrate_mixed)
p = buffer->ptr; p = buffer->ptr;
/* Migrating a protected area should be an error. */ /* Migrating a protected area should be an error. */
ret = hmm_dmirror_cmd(self->fd1, HMM_DMIRROR_MIGRATE, buffer, npages); ret = hmm_migrate_sys_to_dev(self->fd1, buffer, npages);
ASSERT_EQ(ret, -EINVAL); ASSERT_EQ(ret, -EINVAL);
/* Punch a hole after the first page address. */ /* Punch a hole after the first page address. */
...@@ -1023,7 +1086,7 @@ TEST_F(hmm2, migrate_mixed) ...@@ -1023,7 +1086,7 @@ TEST_F(hmm2, migrate_mixed)
ASSERT_EQ(ret, 0); ASSERT_EQ(ret, 0);
/* We expect an error if the vma doesn't cover the range. */ /* We expect an error if the vma doesn't cover the range. */
ret = hmm_dmirror_cmd(self->fd1, HMM_DMIRROR_MIGRATE, buffer, 3); ret = hmm_migrate_sys_to_dev(self->fd1, buffer, 3);
ASSERT_EQ(ret, -EINVAL); ASSERT_EQ(ret, -EINVAL);
/* Page 2 will be a read-only zero page. */ /* Page 2 will be a read-only zero page. */
...@@ -1055,13 +1118,13 @@ TEST_F(hmm2, migrate_mixed) ...@@ -1055,13 +1118,13 @@ TEST_F(hmm2, migrate_mixed)
/* Now try to migrate pages 2-5 to device 1. */ /* Now try to migrate pages 2-5 to device 1. */
buffer->ptr = p + 2 * self->page_size; buffer->ptr = p + 2 * self->page_size;
ret = hmm_dmirror_cmd(self->fd1, HMM_DMIRROR_MIGRATE, buffer, 4); ret = hmm_migrate_sys_to_dev(self->fd1, buffer, 4);
ASSERT_EQ(ret, 0); ASSERT_EQ(ret, 0);
ASSERT_EQ(buffer->cpages, 4); ASSERT_EQ(buffer->cpages, 4);
/* Page 5 won't be migrated to device 0 because it's on device 1. */ /* Page 5 won't be migrated to device 0 because it's on device 1. */
buffer->ptr = p + 5 * self->page_size; buffer->ptr = p + 5 * self->page_size;
ret = hmm_dmirror_cmd(self->fd0, HMM_DMIRROR_MIGRATE, buffer, 1); ret = hmm_migrate_sys_to_dev(self->fd0, buffer, 1);
ASSERT_EQ(ret, -ENOENT); ASSERT_EQ(ret, -ENOENT);
buffer->ptr = p; buffer->ptr = p;
...@@ -1070,8 +1133,12 @@ TEST_F(hmm2, migrate_mixed) ...@@ -1070,8 +1133,12 @@ TEST_F(hmm2, migrate_mixed)
} }
/* /*
* Migrate anonymous memory to device private memory and fault it back to system * Migrate anonymous memory to device memory and back to system memory
* memory multiple times. * multiple times. In case of private zone configuration, this is done
* through fault pages accessed by CPU. In case of coherent zone configuration,
* the pages from the device should be explicitly migrated back to system memory.
* The reason is Coherent device zone has coherent access by CPU, therefore
* it will not generate any page fault.
*/ */
TEST_F(hmm, migrate_multiple) TEST_F(hmm, migrate_multiple)
{ {
...@@ -1107,8 +1174,7 @@ TEST_F(hmm, migrate_multiple) ...@@ -1107,8 +1174,7 @@ TEST_F(hmm, migrate_multiple)
ptr[i] = i; ptr[i] = i;
/* Migrate memory to device. */ /* Migrate memory to device. */
ret = hmm_dmirror_cmd(self->fd, HMM_DMIRROR_MIGRATE, buffer, ret = hmm_migrate_sys_to_dev(self->fd, buffer, npages);
npages);
ASSERT_EQ(ret, 0); ASSERT_EQ(ret, 0);
ASSERT_EQ(buffer->cpages, npages); ASSERT_EQ(buffer->cpages, npages);
...@@ -1116,7 +1182,13 @@ TEST_F(hmm, migrate_multiple) ...@@ -1116,7 +1182,13 @@ TEST_F(hmm, migrate_multiple)
for (i = 0, ptr = buffer->mirror; i < size / sizeof(*ptr); ++i) for (i = 0, ptr = buffer->mirror; i < size / sizeof(*ptr); ++i)
ASSERT_EQ(ptr[i], i); ASSERT_EQ(ptr[i], i);
/* Fault pages back to system memory and check them. */ /* Migrate back to system memory and check them. */
if (hmm_is_coherent_type(variant->device_number)) {
ret = hmm_migrate_dev_to_sys(self->fd, buffer, npages);
ASSERT_EQ(ret, 0);
ASSERT_EQ(buffer->cpages, npages);
}
for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i) for (i = 0, ptr = buffer->ptr; i < size / sizeof(*ptr); ++i)
ASSERT_EQ(ptr[i], i); ASSERT_EQ(ptr[i], i);
...@@ -1354,13 +1426,13 @@ TEST_F(hmm2, snapshot) ...@@ -1354,13 +1426,13 @@ TEST_F(hmm2, snapshot)
/* Page 5 will be migrated to device 0. */ /* Page 5 will be migrated to device 0. */
buffer->ptr = p + 5 * self->page_size; buffer->ptr = p + 5 * self->page_size;
ret = hmm_dmirror_cmd(self->fd0, HMM_DMIRROR_MIGRATE, buffer, 1); ret = hmm_migrate_sys_to_dev(self->fd0, buffer, 1);
ASSERT_EQ(ret, 0); ASSERT_EQ(ret, 0);
ASSERT_EQ(buffer->cpages, 1); ASSERT_EQ(buffer->cpages, 1);
/* Page 6 will be migrated to device 1. */ /* Page 6 will be migrated to device 1. */
buffer->ptr = p + 6 * self->page_size; buffer->ptr = p + 6 * self->page_size;
ret = hmm_dmirror_cmd(self->fd1, HMM_DMIRROR_MIGRATE, buffer, 1); ret = hmm_migrate_sys_to_dev(self->fd1, buffer, 1);
ASSERT_EQ(ret, 0); ASSERT_EQ(ret, 0);
ASSERT_EQ(buffer->cpages, 1); ASSERT_EQ(buffer->cpages, 1);
...@@ -1377,9 +1449,16 @@ TEST_F(hmm2, snapshot) ...@@ -1377,9 +1449,16 @@ TEST_F(hmm2, snapshot)
ASSERT_EQ(m[2], HMM_DMIRROR_PROT_ZERO | HMM_DMIRROR_PROT_READ); ASSERT_EQ(m[2], HMM_DMIRROR_PROT_ZERO | HMM_DMIRROR_PROT_READ);
ASSERT_EQ(m[3], HMM_DMIRROR_PROT_READ); ASSERT_EQ(m[3], HMM_DMIRROR_PROT_READ);
ASSERT_EQ(m[4], HMM_DMIRROR_PROT_WRITE); ASSERT_EQ(m[4], HMM_DMIRROR_PROT_WRITE);
ASSERT_EQ(m[5], HMM_DMIRROR_PROT_DEV_PRIVATE_LOCAL | if (!hmm_is_coherent_type(variant->device_number0)) {
HMM_DMIRROR_PROT_WRITE); ASSERT_EQ(m[5], HMM_DMIRROR_PROT_DEV_PRIVATE_LOCAL |
ASSERT_EQ(m[6], HMM_DMIRROR_PROT_NONE); HMM_DMIRROR_PROT_WRITE);
ASSERT_EQ(m[6], HMM_DMIRROR_PROT_NONE);
} else {
ASSERT_EQ(m[5], HMM_DMIRROR_PROT_DEV_COHERENT_LOCAL |
HMM_DMIRROR_PROT_WRITE);
ASSERT_EQ(m[6], HMM_DMIRROR_PROT_DEV_COHERENT_REMOTE |
HMM_DMIRROR_PROT_WRITE);
}
hmm_buffer_free(buffer); hmm_buffer_free(buffer);
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment