Commit cf3e3e86 authored by Maarten Lankhorst's avatar Maarten Lankhorst

drm/i915: Use ttm mmap handling for ttm bo's.

Use the ttm handlers for servicing page faults, and vm_access.

We do our own validation of read-only access, otherwise use the
ttm handlers as much as possible.

Because the ttm handlers expect the vma_node at vma->base, we slightly
need to massage the mmap handlers to look at vma_node->driver_private
to fetch the bo, if it's NULL, we assume i915's normal mmap_offset uapi
is used.

This is the easiest way to achieve compatibility without changing ttm's
semantics.
Signed-off-by: default avatarMaarten Lankhorst <maarten.lankhorst@linux.intel.com>
Reviewed-by: default avatarThomas Hellström <thomas.hellstrom@linux.intel.com>
Signed-off-by: default avatarMaarten Lankhorst <maarten.lankhorst@linux.intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20210610070152.572423-5-thomas.hellstrom@linux.intel.com
parent f425821b
...@@ -19,6 +19,7 @@ ...@@ -19,6 +19,7 @@
#include "i915_gem_mman.h" #include "i915_gem_mman.h"
#include "i915_trace.h" #include "i915_trace.h"
#include "i915_user_extensions.h" #include "i915_user_extensions.h"
#include "i915_gem_ttm.h"
#include "i915_vma.h" #include "i915_vma.h"
static inline bool static inline bool
...@@ -623,6 +624,8 @@ mmap_offset_attach(struct drm_i915_gem_object *obj, ...@@ -623,6 +624,8 @@ mmap_offset_attach(struct drm_i915_gem_object *obj,
struct i915_mmap_offset *mmo; struct i915_mmap_offset *mmo;
int err; int err;
GEM_BUG_ON(obj->ops->mmap_offset || obj->ops->mmap_ops);
mmo = lookup_mmo(obj, mmap_type); mmo = lookup_mmo(obj, mmap_type);
if (mmo) if (mmo)
goto out; goto out;
...@@ -665,40 +668,47 @@ mmap_offset_attach(struct drm_i915_gem_object *obj, ...@@ -665,40 +668,47 @@ mmap_offset_attach(struct drm_i915_gem_object *obj,
} }
static int static int
__assign_mmap_offset(struct drm_file *file, __assign_mmap_offset(struct drm_i915_gem_object *obj,
u32 handle,
enum i915_mmap_type mmap_type, enum i915_mmap_type mmap_type,
u64 *offset) u64 *offset, struct drm_file *file)
{ {
struct drm_i915_gem_object *obj;
struct i915_mmap_offset *mmo; struct i915_mmap_offset *mmo;
int err;
obj = i915_gem_object_lookup(file, handle); if (i915_gem_object_never_mmap(obj))
if (!obj) return -ENODEV;
return -ENOENT;
if (i915_gem_object_never_mmap(obj)) { if (obj->ops->mmap_offset) {
err = -ENODEV; *offset = obj->ops->mmap_offset(obj);
goto out; return 0;
} }
if (mmap_type != I915_MMAP_TYPE_GTT && if (mmap_type != I915_MMAP_TYPE_GTT &&
!i915_gem_object_has_struct_page(obj) && !i915_gem_object_has_struct_page(obj) &&
!i915_gem_object_type_has(obj, I915_GEM_OBJECT_HAS_IOMEM)) { !i915_gem_object_type_has(obj, I915_GEM_OBJECT_HAS_IOMEM))
err = -ENODEV; return -ENODEV;
goto out;
}
mmo = mmap_offset_attach(obj, mmap_type, file); mmo = mmap_offset_attach(obj, mmap_type, file);
if (IS_ERR(mmo)) { if (IS_ERR(mmo))
err = PTR_ERR(mmo); return PTR_ERR(mmo);
goto out;
}
*offset = drm_vma_node_offset_addr(&mmo->vma_node); *offset = drm_vma_node_offset_addr(&mmo->vma_node);
err = 0; return 0;
out: }
static int
__assign_mmap_offset_handle(struct drm_file *file,
u32 handle,
enum i915_mmap_type mmap_type,
u64 *offset)
{
struct drm_i915_gem_object *obj;
int err;
obj = i915_gem_object_lookup(file, handle);
if (!obj)
return -ENOENT;
err = __assign_mmap_offset(obj, mmap_type, offset, file);
i915_gem_object_put(obj); i915_gem_object_put(obj);
return err; return err;
} }
...@@ -718,7 +728,7 @@ i915_gem_dumb_mmap_offset(struct drm_file *file, ...@@ -718,7 +728,7 @@ i915_gem_dumb_mmap_offset(struct drm_file *file,
else else
mmap_type = I915_MMAP_TYPE_GTT; mmap_type = I915_MMAP_TYPE_GTT;
return __assign_mmap_offset(file, handle, mmap_type, offset); return __assign_mmap_offset_handle(file, handle, mmap_type, offset);
} }
/** /**
...@@ -786,7 +796,7 @@ i915_gem_mmap_offset_ioctl(struct drm_device *dev, void *data, ...@@ -786,7 +796,7 @@ i915_gem_mmap_offset_ioctl(struct drm_device *dev, void *data,
return -EINVAL; return -EINVAL;
} }
return __assign_mmap_offset(file, args->handle, type, &args->offset); return __assign_mmap_offset_handle(file, args->handle, type, &args->offset);
} }
static void vm_open(struct vm_area_struct *vma) static void vm_open(struct vm_area_struct *vma)
...@@ -890,8 +900,18 @@ int i915_gem_mmap(struct file *filp, struct vm_area_struct *vma) ...@@ -890,8 +900,18 @@ int i915_gem_mmap(struct file *filp, struct vm_area_struct *vma)
* destroyed and will be invalid when the vma manager lock * destroyed and will be invalid when the vma manager lock
* is released. * is released.
*/ */
if (!node->driver_private) {
mmo = container_of(node, struct i915_mmap_offset, vma_node); mmo = container_of(node, struct i915_mmap_offset, vma_node);
obj = i915_gem_object_get_rcu(mmo->obj); obj = i915_gem_object_get_rcu(mmo->obj);
GEM_BUG_ON(obj && obj->ops->mmap_ops);
} else {
obj = i915_gem_object_get_rcu
(container_of(node, struct drm_i915_gem_object,
base.vma_node));
GEM_BUG_ON(obj && !obj->ops->mmap_ops);
}
} }
drm_vma_offset_unlock_lookup(dev->vma_offset_manager); drm_vma_offset_unlock_lookup(dev->vma_offset_manager);
rcu_read_unlock(); rcu_read_unlock();
...@@ -913,7 +933,9 @@ int i915_gem_mmap(struct file *filp, struct vm_area_struct *vma) ...@@ -913,7 +933,9 @@ int i915_gem_mmap(struct file *filp, struct vm_area_struct *vma)
} }
vma->vm_flags |= VM_PFNMAP | VM_DONTEXPAND | VM_DONTDUMP; vma->vm_flags |= VM_PFNMAP | VM_DONTEXPAND | VM_DONTDUMP;
vma->vm_private_data = mmo;
if (i915_gem_object_has_iomem(obj))
vma->vm_flags |= VM_IO;
/* /*
* We keep the ref on mmo->obj, not vm_file, but we require * We keep the ref on mmo->obj, not vm_file, but we require
...@@ -927,6 +949,15 @@ int i915_gem_mmap(struct file *filp, struct vm_area_struct *vma) ...@@ -927,6 +949,15 @@ int i915_gem_mmap(struct file *filp, struct vm_area_struct *vma)
/* Drop the initial creation reference, the vma is now holding one. */ /* Drop the initial creation reference, the vma is now holding one. */
fput(anon); fput(anon);
if (obj->ops->mmap_ops) {
vma->vm_page_prot = pgprot_decrypted(vm_get_page_prot(vma->vm_flags));
vma->vm_ops = obj->ops->mmap_ops;
vma->vm_private_data = node->driver_private;
return 0;
}
vma->vm_private_data = mmo;
switch (mmo->mmap_type) { switch (mmo->mmap_type) {
case I915_MMAP_TYPE_WC: case I915_MMAP_TYPE_WC:
vma->vm_page_prot = vma->vm_page_prot =
......
...@@ -342,14 +342,14 @@ struct scatterlist * ...@@ -342,14 +342,14 @@ struct scatterlist *
__i915_gem_object_get_sg(struct drm_i915_gem_object *obj, __i915_gem_object_get_sg(struct drm_i915_gem_object *obj,
struct i915_gem_object_page_iter *iter, struct i915_gem_object_page_iter *iter,
unsigned int n, unsigned int n,
unsigned int *offset, bool allow_alloc); unsigned int *offset, bool allow_alloc, bool dma);
static inline struct scatterlist * static inline struct scatterlist *
i915_gem_object_get_sg(struct drm_i915_gem_object *obj, i915_gem_object_get_sg(struct drm_i915_gem_object *obj,
unsigned int n, unsigned int n,
unsigned int *offset, bool allow_alloc) unsigned int *offset, bool allow_alloc)
{ {
return __i915_gem_object_get_sg(obj, &obj->mm.get_page, n, offset, allow_alloc); return __i915_gem_object_get_sg(obj, &obj->mm.get_page, n, offset, allow_alloc, false);
} }
static inline struct scatterlist * static inline struct scatterlist *
...@@ -357,7 +357,7 @@ i915_gem_object_get_sg_dma(struct drm_i915_gem_object *obj, ...@@ -357,7 +357,7 @@ i915_gem_object_get_sg_dma(struct drm_i915_gem_object *obj,
unsigned int n, unsigned int n,
unsigned int *offset, bool allow_alloc) unsigned int *offset, bool allow_alloc)
{ {
return __i915_gem_object_get_sg(obj, &obj->mm.get_dma_page, n, offset, allow_alloc); return __i915_gem_object_get_sg(obj, &obj->mm.get_dma_page, n, offset, allow_alloc, true);
} }
struct page * struct page *
......
...@@ -61,6 +61,7 @@ struct drm_i915_gem_object_ops { ...@@ -61,6 +61,7 @@ struct drm_i915_gem_object_ops {
const struct drm_i915_gem_pread *arg); const struct drm_i915_gem_pread *arg);
int (*pwrite)(struct drm_i915_gem_object *obj, int (*pwrite)(struct drm_i915_gem_object *obj,
const struct drm_i915_gem_pwrite *arg); const struct drm_i915_gem_pwrite *arg);
u64 (*mmap_offset)(struct drm_i915_gem_object *obj);
int (*dmabuf_export)(struct drm_i915_gem_object *obj); int (*dmabuf_export)(struct drm_i915_gem_object *obj);
...@@ -79,6 +80,7 @@ struct drm_i915_gem_object_ops { ...@@ -79,6 +80,7 @@ struct drm_i915_gem_object_ops {
void (*delayed_free)(struct drm_i915_gem_object *obj); void (*delayed_free)(struct drm_i915_gem_object *obj);
void (*release)(struct drm_i915_gem_object *obj); void (*release)(struct drm_i915_gem_object *obj);
const struct vm_operations_struct *mmap_ops;
const char *name; /* friendly name for debug, e.g. lockdep classes */ const char *name; /* friendly name for debug, e.g. lockdep classes */
}; };
...@@ -328,6 +330,7 @@ struct drm_i915_gem_object { ...@@ -328,6 +330,7 @@ struct drm_i915_gem_object {
struct { struct {
struct sg_table *cached_io_st; struct sg_table *cached_io_st;
struct i915_gem_object_page_iter get_io_page;
bool created:1; bool created:1;
} ttm; } ttm;
......
...@@ -467,9 +467,8 @@ __i915_gem_object_get_sg(struct drm_i915_gem_object *obj, ...@@ -467,9 +467,8 @@ __i915_gem_object_get_sg(struct drm_i915_gem_object *obj,
struct i915_gem_object_page_iter *iter, struct i915_gem_object_page_iter *iter,
unsigned int n, unsigned int n,
unsigned int *offset, unsigned int *offset,
bool allow_alloc) bool allow_alloc, bool dma)
{ {
const bool dma = iter == &obj->mm.get_dma_page;
struct scatterlist *sg; struct scatterlist *sg;
unsigned int idx, count; unsigned int idx, count;
......
...@@ -13,6 +13,7 @@ ...@@ -13,6 +13,7 @@
#include "gem/i915_gem_object.h" #include "gem/i915_gem_object.h"
#include "gem/i915_gem_region.h" #include "gem/i915_gem_region.h"
#include "gem/i915_gem_ttm.h" #include "gem/i915_gem_ttm.h"
#include "gem/i915_gem_mman.h"
#define I915_PL_LMEM0 TTM_PL_PRIV #define I915_PL_LMEM0 TTM_PL_PRIV
#define I915_PL_SYSTEM TTM_PL_SYSTEM #define I915_PL_SYSTEM TTM_PL_SYSTEM
...@@ -158,11 +159,20 @@ static int i915_ttm_move_notify(struct ttm_buffer_object *bo) ...@@ -158,11 +159,20 @@ static int i915_ttm_move_notify(struct ttm_buffer_object *bo)
static void i915_ttm_free_cached_io_st(struct drm_i915_gem_object *obj) static void i915_ttm_free_cached_io_st(struct drm_i915_gem_object *obj)
{ {
if (obj->ttm.cached_io_st) { struct radix_tree_iter iter;
void __rcu **slot;
if (!obj->ttm.cached_io_st)
return;
rcu_read_lock();
radix_tree_for_each_slot(slot, &obj->ttm.get_io_page.radix, &iter, 0)
radix_tree_delete(&obj->ttm.get_io_page.radix, iter.index);
rcu_read_unlock();
sg_free_table(obj->ttm.cached_io_st); sg_free_table(obj->ttm.cached_io_st);
kfree(obj->ttm.cached_io_st); kfree(obj->ttm.cached_io_st);
obj->ttm.cached_io_st = NULL; obj->ttm.cached_io_st = NULL;
}
} }
static void i915_ttm_purge(struct drm_i915_gem_object *obj) static void i915_ttm_purge(struct drm_i915_gem_object *obj)
...@@ -338,12 +348,41 @@ static int i915_ttm_move(struct ttm_buffer_object *bo, bool evict, ...@@ -338,12 +348,41 @@ static int i915_ttm_move(struct ttm_buffer_object *bo, bool evict,
ttm_bo_move_sync_cleanup(bo, dst_mem); ttm_bo_move_sync_cleanup(bo, dst_mem);
i915_ttm_free_cached_io_st(obj); i915_ttm_free_cached_io_st(obj);
if (!dst_man->use_tt) if (!dst_man->use_tt) {
obj->ttm.cached_io_st = dst_st; obj->ttm.cached_io_st = dst_st;
obj->ttm.get_io_page.sg_pos = dst_st->sgl;
obj->ttm.get_io_page.sg_idx = 0;
}
return 0; return 0;
} }
static int i915_ttm_io_mem_reserve(struct ttm_device *bdev, struct ttm_resource *mem)
{
if (mem->mem_type < I915_PL_LMEM0)
return 0;
mem->bus.caching = ttm_write_combined;
mem->bus.is_iomem = true;
return 0;
}
static unsigned long i915_ttm_io_mem_pfn(struct ttm_buffer_object *bo,
unsigned long page_offset)
{
struct drm_i915_gem_object *obj = i915_ttm_to_gem(bo);
unsigned long base = obj->mm.region->iomap.base - obj->mm.region->region.start;
struct scatterlist *sg;
unsigned int ofs;
GEM_WARN_ON(bo->ttm);
sg = __i915_gem_object_get_sg(obj, &obj->ttm.get_io_page, page_offset, &ofs, true, true);
return ((base + sg_dma_address(sg)) >> PAGE_SHIFT) + ofs;
}
static struct ttm_device_funcs i915_ttm_bo_driver = { static struct ttm_device_funcs i915_ttm_bo_driver = {
.ttm_tt_create = i915_ttm_tt_create, .ttm_tt_create = i915_ttm_tt_create,
.ttm_tt_unpopulate = i915_ttm_tt_unpopulate, .ttm_tt_unpopulate = i915_ttm_tt_unpopulate,
...@@ -353,6 +392,8 @@ static struct ttm_device_funcs i915_ttm_bo_driver = { ...@@ -353,6 +392,8 @@ static struct ttm_device_funcs i915_ttm_bo_driver = {
.move = i915_ttm_move, .move = i915_ttm_move,
.swap_notify = i915_ttm_swap_notify, .swap_notify = i915_ttm_swap_notify,
.delete_mem_notify = i915_ttm_delete_mem_notify, .delete_mem_notify = i915_ttm_delete_mem_notify,
.io_mem_reserve = i915_ttm_io_mem_reserve,
.io_mem_pfn = i915_ttm_io_mem_pfn,
}; };
/** /**
...@@ -460,7 +501,67 @@ static void i915_ttm_delayed_free(struct drm_i915_gem_object *obj) ...@@ -460,7 +501,67 @@ static void i915_ttm_delayed_free(struct drm_i915_gem_object *obj)
} }
} }
static const struct drm_i915_gem_object_ops i915_gem_ttm_obj_ops = { static vm_fault_t vm_fault_ttm(struct vm_fault *vmf)
{
struct vm_area_struct *area = vmf->vma;
struct drm_i915_gem_object *obj =
i915_ttm_to_gem(area->vm_private_data);
/* Sanity check that we allow writing into this object */
if (unlikely(i915_gem_object_is_readonly(obj) &&
area->vm_flags & VM_WRITE))
return VM_FAULT_SIGBUS;
return ttm_bo_vm_fault(vmf);
}
static int
vm_access_ttm(struct vm_area_struct *area, unsigned long addr,
void *buf, int len, int write)
{
struct drm_i915_gem_object *obj =
i915_ttm_to_gem(area->vm_private_data);
if (i915_gem_object_is_readonly(obj) && write)
return -EACCES;
return ttm_bo_vm_access(area, addr, buf, len, write);
}
static void ttm_vm_open(struct vm_area_struct *vma)
{
struct drm_i915_gem_object *obj =
i915_ttm_to_gem(vma->vm_private_data);
GEM_BUG_ON(!obj);
i915_gem_object_get(obj);
}
static void ttm_vm_close(struct vm_area_struct *vma)
{
struct drm_i915_gem_object *obj =
i915_ttm_to_gem(vma->vm_private_data);
GEM_BUG_ON(!obj);
i915_gem_object_put(obj);
}
static const struct vm_operations_struct vm_ops_ttm = {
.fault = vm_fault_ttm,
.access = vm_access_ttm,
.open = ttm_vm_open,
.close = ttm_vm_close,
};
static u64 i915_ttm_mmap_offset(struct drm_i915_gem_object *obj)
{
/* The ttm_bo must be allocated with I915_BO_ALLOC_USER */
GEM_BUG_ON(!drm_mm_node_allocated(&obj->base.vma_node.vm_node));
return drm_vma_node_offset_addr(&obj->base.vma_node);
}
const struct drm_i915_gem_object_ops i915_gem_ttm_obj_ops = {
.name = "i915_gem_object_ttm", .name = "i915_gem_object_ttm",
.flags = I915_GEM_OBJECT_HAS_IOMEM, .flags = I915_GEM_OBJECT_HAS_IOMEM,
...@@ -469,6 +570,8 @@ static const struct drm_i915_gem_object_ops i915_gem_ttm_obj_ops = { ...@@ -469,6 +570,8 @@ static const struct drm_i915_gem_object_ops i915_gem_ttm_obj_ops = {
.truncate = i915_ttm_purge, .truncate = i915_ttm_purge,
.adjust_lru = i915_ttm_adjust_lru, .adjust_lru = i915_ttm_adjust_lru,
.delayed_free = i915_ttm_delayed_free, .delayed_free = i915_ttm_delayed_free,
.mmap_offset = i915_ttm_mmap_offset,
.mmap_ops = &vm_ops_ttm,
}; };
void i915_ttm_bo_destroy(struct ttm_buffer_object *bo) void i915_ttm_bo_destroy(struct ttm_buffer_object *bo)
...@@ -476,6 +579,7 @@ void i915_ttm_bo_destroy(struct ttm_buffer_object *bo) ...@@ -476,6 +579,7 @@ void i915_ttm_bo_destroy(struct ttm_buffer_object *bo)
struct drm_i915_gem_object *obj = i915_ttm_to_gem(bo); struct drm_i915_gem_object *obj = i915_ttm_to_gem(bo);
i915_gem_object_release_memory_region(obj); i915_gem_object_release_memory_region(obj);
mutex_destroy(&obj->ttm.get_io_page.lock);
if (obj->ttm.created) if (obj->ttm.created)
call_rcu(&obj->rcu, __i915_gem_free_object_rcu); call_rcu(&obj->rcu, __i915_gem_free_object_rcu);
} }
...@@ -517,6 +621,8 @@ int __i915_gem_ttm_object_init(struct intel_memory_region *mem, ...@@ -517,6 +621,8 @@ int __i915_gem_ttm_object_init(struct intel_memory_region *mem,
i915_gem_object_make_unshrinkable(obj); i915_gem_object_make_unshrinkable(obj);
obj->read_domains = I915_GEM_DOMAIN_WC | I915_GEM_DOMAIN_GTT; obj->read_domains = I915_GEM_DOMAIN_WC | I915_GEM_DOMAIN_GTT;
i915_gem_object_set_cache_coherency(obj, I915_CACHE_NONE); i915_gem_object_set_cache_coherency(obj, I915_CACHE_NONE);
INIT_RADIX_TREE(&obj->ttm.get_io_page.radix, GFP_KERNEL | __GFP_NOWARN);
mutex_init(&obj->ttm.get_io_page.lock);
bo_type = (obj->flags & I915_BO_ALLOC_USER) ? ttm_bo_type_device : bo_type = (obj->flags & I915_BO_ALLOC_USER) ? ttm_bo_type_device :
ttm_bo_type_kernel; ttm_bo_type_kernel;
...@@ -528,6 +634,7 @@ int __i915_gem_ttm_object_init(struct intel_memory_region *mem, ...@@ -528,6 +634,7 @@ int __i915_gem_ttm_object_init(struct intel_memory_region *mem,
* Similarly, in delayed_destroy, we can't call ttm_bo_put() * Similarly, in delayed_destroy, we can't call ttm_bo_put()
* until successful initialization. * until successful initialization.
*/ */
obj->base.vma_node.driver_private = i915_gem_to_ttm(obj);
ret = ttm_bo_init(&i915->bdev, i915_gem_to_ttm(obj), size, ret = ttm_bo_init(&i915->bdev, i915_gem_to_ttm(obj), size,
bo_type, &i915_sys_placement, alignment, bo_type, &i915_sys_placement, alignment,
true, NULL, NULL, i915_ttm_bo_destroy); true, NULL, NULL, i915_ttm_bo_destroy);
......
...@@ -578,16 +578,17 @@ static bool assert_mmap_offset(struct drm_i915_private *i915, ...@@ -578,16 +578,17 @@ static bool assert_mmap_offset(struct drm_i915_private *i915,
int expected) int expected)
{ {
struct drm_i915_gem_object *obj; struct drm_i915_gem_object *obj;
struct i915_mmap_offset *mmo; u64 offset;
int ret;
obj = i915_gem_object_create_internal(i915, size); obj = i915_gem_object_create_internal(i915, size);
if (IS_ERR(obj)) if (IS_ERR(obj))
return false; return expected && expected == PTR_ERR(obj);
mmo = mmap_offset_attach(obj, I915_MMAP_OFFSET_GTT, NULL); ret = __assign_mmap_offset(obj, I915_MMAP_TYPE_GTT, &offset, NULL);
i915_gem_object_put(obj); i915_gem_object_put(obj);
return PTR_ERR_OR_ZERO(mmo) == expected; return ret == expected;
} }
static void disable_retire_worker(struct drm_i915_private *i915) static void disable_retire_worker(struct drm_i915_private *i915)
...@@ -622,8 +623,8 @@ static int igt_mmap_offset_exhaustion(void *arg) ...@@ -622,8 +623,8 @@ static int igt_mmap_offset_exhaustion(void *arg)
struct drm_mm *mm = &i915->drm.vma_offset_manager->vm_addr_space_mm; struct drm_mm *mm = &i915->drm.vma_offset_manager->vm_addr_space_mm;
struct drm_i915_gem_object *obj; struct drm_i915_gem_object *obj;
struct drm_mm_node *hole, *next; struct drm_mm_node *hole, *next;
struct i915_mmap_offset *mmo;
int loop, err = 0; int loop, err = 0;
u64 offset;
/* Disable background reaper */ /* Disable background reaper */
disable_retire_worker(i915); disable_retire_worker(i915);
...@@ -684,13 +685,13 @@ static int igt_mmap_offset_exhaustion(void *arg) ...@@ -684,13 +685,13 @@ static int igt_mmap_offset_exhaustion(void *arg)
obj = i915_gem_object_create_internal(i915, PAGE_SIZE); obj = i915_gem_object_create_internal(i915, PAGE_SIZE);
if (IS_ERR(obj)) { if (IS_ERR(obj)) {
err = PTR_ERR(obj); err = PTR_ERR(obj);
pr_err("Unable to create object for reclaimed hole\n");
goto out; goto out;
} }
mmo = mmap_offset_attach(obj, I915_MMAP_OFFSET_GTT, NULL); err = __assign_mmap_offset(obj, I915_MMAP_TYPE_GTT, &offset, NULL);
if (IS_ERR(mmo)) { if (err) {
pr_err("Unable to insert object into reclaimed hole\n"); pr_err("Unable to insert object into reclaimed hole\n");
err = PTR_ERR(mmo);
goto err_obj; goto err_obj;
} }
...@@ -865,10 +866,10 @@ static int __igt_mmap(struct drm_i915_private *i915, ...@@ -865,10 +866,10 @@ static int __igt_mmap(struct drm_i915_private *i915,
struct drm_i915_gem_object *obj, struct drm_i915_gem_object *obj,
enum i915_mmap_type type) enum i915_mmap_type type)
{ {
struct i915_mmap_offset *mmo;
struct vm_area_struct *area; struct vm_area_struct *area;
unsigned long addr; unsigned long addr;
int err, i; int err, i;
u64 offset;
if (!can_mmap(obj, type)) if (!can_mmap(obj, type))
return 0; return 0;
...@@ -879,11 +880,11 @@ static int __igt_mmap(struct drm_i915_private *i915, ...@@ -879,11 +880,11 @@ static int __igt_mmap(struct drm_i915_private *i915,
if (err) if (err)
return err; return err;
mmo = mmap_offset_attach(obj, type, NULL); err = __assign_mmap_offset(obj, type, &offset, NULL);
if (IS_ERR(mmo)) if (err)
return PTR_ERR(mmo); return err;
addr = igt_mmap_node(i915, &mmo->vma_node, 0, PROT_WRITE, MAP_SHARED); addr = igt_mmap_offset(i915, offset, obj->base.size, PROT_WRITE, MAP_SHARED);
if (IS_ERR_VALUE(addr)) if (IS_ERR_VALUE(addr))
return addr; return addr;
...@@ -897,13 +898,6 @@ static int __igt_mmap(struct drm_i915_private *i915, ...@@ -897,13 +898,6 @@ static int __igt_mmap(struct drm_i915_private *i915,
goto out_unmap; goto out_unmap;
} }
if (area->vm_private_data != mmo) {
pr_err("%s: vm_area_struct did not point back to our mmap_offset object!\n",
obj->mm.region->name);
err = -EINVAL;
goto out_unmap;
}
for (i = 0; i < obj->base.size / sizeof(u32); i++) { for (i = 0; i < obj->base.size / sizeof(u32); i++) {
u32 __user *ux = u64_to_user_ptr((u64)(addr + i * sizeof(*ux))); u32 __user *ux = u64_to_user_ptr((u64)(addr + i * sizeof(*ux)));
u32 x; u32 x;
...@@ -961,7 +955,7 @@ static int igt_mmap(void *arg) ...@@ -961,7 +955,7 @@ static int igt_mmap(void *arg)
struct drm_i915_gem_object *obj; struct drm_i915_gem_object *obj;
int err; int err;
obj = i915_gem_object_create_region(mr, sizes[i], 0); obj = i915_gem_object_create_region(mr, sizes[i], I915_BO_ALLOC_USER);
if (obj == ERR_PTR(-ENODEV)) if (obj == ERR_PTR(-ENODEV))
continue; continue;
...@@ -1004,12 +998,12 @@ static int __igt_mmap_access(struct drm_i915_private *i915, ...@@ -1004,12 +998,12 @@ static int __igt_mmap_access(struct drm_i915_private *i915,
struct drm_i915_gem_object *obj, struct drm_i915_gem_object *obj,
enum i915_mmap_type type) enum i915_mmap_type type)
{ {
struct i915_mmap_offset *mmo;
unsigned long __user *ptr; unsigned long __user *ptr;
unsigned long A, B; unsigned long A, B;
unsigned long x, y; unsigned long x, y;
unsigned long addr; unsigned long addr;
int err; int err;
u64 offset;
memset(&A, 0xAA, sizeof(A)); memset(&A, 0xAA, sizeof(A));
memset(&B, 0xBB, sizeof(B)); memset(&B, 0xBB, sizeof(B));
...@@ -1017,11 +1011,11 @@ static int __igt_mmap_access(struct drm_i915_private *i915, ...@@ -1017,11 +1011,11 @@ static int __igt_mmap_access(struct drm_i915_private *i915,
if (!can_mmap(obj, type) || !can_access(obj)) if (!can_mmap(obj, type) || !can_access(obj))
return 0; return 0;
mmo = mmap_offset_attach(obj, type, NULL); err = __assign_mmap_offset(obj, type, &offset, NULL);
if (IS_ERR(mmo)) if (err)
return PTR_ERR(mmo); return err;
addr = igt_mmap_node(i915, &mmo->vma_node, 0, PROT_WRITE, MAP_SHARED); addr = igt_mmap_offset(i915, offset, obj->base.size, PROT_WRITE, MAP_SHARED);
if (IS_ERR_VALUE(addr)) if (IS_ERR_VALUE(addr))
return addr; return addr;
ptr = (unsigned long __user *)addr; ptr = (unsigned long __user *)addr;
...@@ -1081,7 +1075,7 @@ static int igt_mmap_access(void *arg) ...@@ -1081,7 +1075,7 @@ static int igt_mmap_access(void *arg)
struct drm_i915_gem_object *obj; struct drm_i915_gem_object *obj;
int err; int err;
obj = i915_gem_object_create_region(mr, PAGE_SIZE, 0); obj = i915_gem_object_create_region(mr, PAGE_SIZE, I915_BO_ALLOC_USER);
if (obj == ERR_PTR(-ENODEV)) if (obj == ERR_PTR(-ENODEV))
continue; continue;
...@@ -1111,11 +1105,11 @@ static int __igt_mmap_gpu(struct drm_i915_private *i915, ...@@ -1111,11 +1105,11 @@ static int __igt_mmap_gpu(struct drm_i915_private *i915,
enum i915_mmap_type type) enum i915_mmap_type type)
{ {
struct intel_engine_cs *engine; struct intel_engine_cs *engine;
struct i915_mmap_offset *mmo;
unsigned long addr; unsigned long addr;
u32 __user *ux; u32 __user *ux;
u32 bbe; u32 bbe;
int err; int err;
u64 offset;
/* /*
* Verify that the mmap access into the backing store aligns with * Verify that the mmap access into the backing store aligns with
...@@ -1132,11 +1126,11 @@ static int __igt_mmap_gpu(struct drm_i915_private *i915, ...@@ -1132,11 +1126,11 @@ static int __igt_mmap_gpu(struct drm_i915_private *i915,
if (err) if (err)
return err; return err;
mmo = mmap_offset_attach(obj, type, NULL); err = __assign_mmap_offset(obj, type, &offset, NULL);
if (IS_ERR(mmo)) if (err)
return PTR_ERR(mmo); return err;
addr = igt_mmap_node(i915, &mmo->vma_node, 0, PROT_WRITE, MAP_SHARED); addr = igt_mmap_offset(i915, offset, obj->base.size, PROT_WRITE, MAP_SHARED);
if (IS_ERR_VALUE(addr)) if (IS_ERR_VALUE(addr))
return addr; return addr;
...@@ -1226,7 +1220,7 @@ static int igt_mmap_gpu(void *arg) ...@@ -1226,7 +1220,7 @@ static int igt_mmap_gpu(void *arg)
struct drm_i915_gem_object *obj; struct drm_i915_gem_object *obj;
int err; int err;
obj = i915_gem_object_create_region(mr, PAGE_SIZE, 0); obj = i915_gem_object_create_region(mr, PAGE_SIZE, I915_BO_ALLOC_USER);
if (obj == ERR_PTR(-ENODEV)) if (obj == ERR_PTR(-ENODEV))
continue; continue;
...@@ -1303,18 +1297,18 @@ static int __igt_mmap_revoke(struct drm_i915_private *i915, ...@@ -1303,18 +1297,18 @@ static int __igt_mmap_revoke(struct drm_i915_private *i915,
struct drm_i915_gem_object *obj, struct drm_i915_gem_object *obj,
enum i915_mmap_type type) enum i915_mmap_type type)
{ {
struct i915_mmap_offset *mmo;
unsigned long addr; unsigned long addr;
int err; int err;
u64 offset;
if (!can_mmap(obj, type)) if (!can_mmap(obj, type))
return 0; return 0;
mmo = mmap_offset_attach(obj, type, NULL); err = __assign_mmap_offset(obj, type, &offset, NULL);
if (IS_ERR(mmo)) if (err)
return PTR_ERR(mmo); return err;
addr = igt_mmap_node(i915, &mmo->vma_node, 0, PROT_WRITE, MAP_SHARED); addr = igt_mmap_offset(i915, offset, obj->base.size, PROT_WRITE, MAP_SHARED);
if (IS_ERR_VALUE(addr)) if (IS_ERR_VALUE(addr))
return addr; return addr;
...@@ -1350,11 +1344,21 @@ static int __igt_mmap_revoke(struct drm_i915_private *i915, ...@@ -1350,11 +1344,21 @@ static int __igt_mmap_revoke(struct drm_i915_private *i915,
} }
} }
if (!obj->ops->mmap_ops) {
err = check_absent(addr, obj->base.size); err = check_absent(addr, obj->base.size);
if (err) { if (err) {
pr_err("%s: was not absent\n", obj->mm.region->name); pr_err("%s: was not absent\n", obj->mm.region->name);
goto out_unmap; goto out_unmap;
} }
} else {
/* ttm allows access to evicted regions by design */
err = check_present(addr, obj->base.size);
if (err) {
pr_err("%s: was not present\n", obj->mm.region->name);
goto out_unmap;
}
}
out_unmap: out_unmap:
vm_munmap(addr, obj->base.size); vm_munmap(addr, obj->base.size);
...@@ -1371,7 +1375,7 @@ static int igt_mmap_revoke(void *arg) ...@@ -1371,7 +1375,7 @@ static int igt_mmap_revoke(void *arg)
struct drm_i915_gem_object *obj; struct drm_i915_gem_object *obj;
int err; int err;
obj = i915_gem_object_create_region(mr, PAGE_SIZE, 0); obj = i915_gem_object_create_region(mr, PAGE_SIZE, I915_BO_ALLOC_USER);
if (obj == ERR_PTR(-ENODEV)) if (obj == ERR_PTR(-ENODEV))
continue; continue;
......
...@@ -9,15 +9,28 @@ ...@@ -9,15 +9,28 @@
#include "i915_drv.h" #include "i915_drv.h"
#include "igt_mmap.h" #include "igt_mmap.h"
unsigned long igt_mmap_node(struct drm_i915_private *i915, unsigned long igt_mmap_offset(struct drm_i915_private *i915,
struct drm_vma_offset_node *node, u64 offset,
unsigned long addr, unsigned long size,
unsigned long prot, unsigned long prot,
unsigned long flags) unsigned long flags)
{ {
struct drm_vma_offset_node *node;
struct file *file; struct file *file;
unsigned long addr;
int err; int err;
/* no need to refcount, we own this object */
drm_vma_offset_lock_lookup(i915->drm.vma_offset_manager);
node = drm_vma_offset_exact_lookup_locked(i915->drm.vma_offset_manager,
offset / PAGE_SIZE, size / PAGE_SIZE);
drm_vma_offset_unlock_lookup(i915->drm.vma_offset_manager);
if (GEM_WARN_ON(!node)) {
pr_info("Failed to lookup %llx\n", offset);
return -ENOENT;
}
/* Pretend to open("/dev/dri/card0") */ /* Pretend to open("/dev/dri/card0") */
file = mock_drm_getfile(i915->drm.primary, O_RDWR); file = mock_drm_getfile(i915->drm.primary, O_RDWR);
if (IS_ERR(file)) if (IS_ERR(file))
...@@ -29,7 +42,7 @@ unsigned long igt_mmap_node(struct drm_i915_private *i915, ...@@ -29,7 +42,7 @@ unsigned long igt_mmap_node(struct drm_i915_private *i915,
goto out_file; goto out_file;
} }
addr = vm_mmap(file, addr, drm_vma_node_size(node) << PAGE_SHIFT, addr = vm_mmap(file, 0, drm_vma_node_size(node) << PAGE_SHIFT,
prot, flags, drm_vma_node_offset_addr(node)); prot, flags, drm_vma_node_offset_addr(node));
drm_vma_node_revoke(node, file->private_data); drm_vma_node_revoke(node, file->private_data);
......
...@@ -7,12 +7,14 @@ ...@@ -7,12 +7,14 @@
#ifndef IGT_MMAP_H #ifndef IGT_MMAP_H
#define IGT_MMAP_H #define IGT_MMAP_H
#include <linux/types.h>
struct drm_i915_private; struct drm_i915_private;
struct drm_vma_offset_node; struct drm_vma_offset_node;
unsigned long igt_mmap_node(struct drm_i915_private *i915, unsigned long igt_mmap_offset(struct drm_i915_private *i915,
struct drm_vma_offset_node *node, u64 offset,
unsigned long addr, unsigned long size,
unsigned long prot, unsigned long prot,
unsigned long flags); unsigned long flags);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment