Commit eecbb3cd authored by Rob Herring's avatar Rob Herring

drm/panfrost: Split panfrost_mmu_map SG list mapping to its own function

In preparation to create partial GPU mappings of BOs on page faults,
split out the SG list handling of panfrost_mmu_map().

Cc: Tomeu Vizoso <tomeu.vizoso@collabora.com>
Cc: Boris Brezillon <boris.brezillon@collabora.com>
Cc: Robin Murphy <robin.murphy@arm.com>
Reviewed: Steven Price <steven.price@arm.com>
Acked-by: default avatarAlyssa Rosenzweig <alyssa.rosenzweig@collabora.com>
Signed-off-by: default avatarRob Herring <robh@kernel.org>
Link: https://patchwork.freedesktop.org/patch/msgid/20190808222200.13176-5-robh@kernel.org
parent a5efb4c9
...@@ -145,27 +145,13 @@ static size_t get_pgsize(u64 addr, size_t size) ...@@ -145,27 +145,13 @@ static size_t get_pgsize(u64 addr, size_t size)
return SZ_2M; return SZ_2M;
} }
int panfrost_mmu_map(struct panfrost_gem_object *bo) static int mmu_map_sg(struct panfrost_device *pfdev, u64 iova,
int prot, struct sg_table *sgt)
{ {
struct drm_gem_object *obj = &bo->base.base;
struct panfrost_device *pfdev = to_panfrost_device(obj->dev);
struct io_pgtable_ops *ops = pfdev->mmu->pgtbl_ops;
u64 iova = bo->node.start << PAGE_SHIFT;
unsigned int count; unsigned int count;
struct scatterlist *sgl; struct scatterlist *sgl;
struct sg_table *sgt; struct io_pgtable_ops *ops = pfdev->mmu->pgtbl_ops;
int ret; u64 start_iova = iova;
if (WARN_ON(bo->is_mapped))
return 0;
sgt = drm_gem_shmem_get_pages_sgt(obj);
if (WARN_ON(IS_ERR(sgt)))
return PTR_ERR(sgt);
ret = pm_runtime_get_sync(pfdev->dev);
if (ret < 0)
return ret;
mutex_lock(&pfdev->mmu->lock); mutex_lock(&pfdev->mmu->lock);
...@@ -178,18 +164,42 @@ int panfrost_mmu_map(struct panfrost_gem_object *bo) ...@@ -178,18 +164,42 @@ int panfrost_mmu_map(struct panfrost_gem_object *bo)
while (len) { while (len) {
size_t pgsize = get_pgsize(iova | paddr, len); size_t pgsize = get_pgsize(iova | paddr, len);
ops->map(ops, iova, paddr, pgsize, IOMMU_WRITE | IOMMU_READ); ops->map(ops, iova, paddr, pgsize, prot);
iova += pgsize; iova += pgsize;
paddr += pgsize; paddr += pgsize;
len -= pgsize; len -= pgsize;
} }
} }
mmu_hw_do_operation(pfdev, 0, bo->node.start << PAGE_SHIFT, mmu_hw_do_operation(pfdev, 0, start_iova, iova - start_iova,
bo->node.size << PAGE_SHIFT, AS_COMMAND_FLUSH_PT); AS_COMMAND_FLUSH_PT);
mutex_unlock(&pfdev->mmu->lock); mutex_unlock(&pfdev->mmu->lock);
return 0;
}
int panfrost_mmu_map(struct panfrost_gem_object *bo)
{
struct drm_gem_object *obj = &bo->base.base;
struct panfrost_device *pfdev = to_panfrost_device(obj->dev);
struct sg_table *sgt;
int ret;
int prot = IOMMU_READ | IOMMU_WRITE;
if (WARN_ON(bo->is_mapped))
return 0;
sgt = drm_gem_shmem_get_pages_sgt(obj);
if (WARN_ON(IS_ERR(sgt)))
return PTR_ERR(sgt);
ret = pm_runtime_get_sync(pfdev->dev);
if (ret < 0)
return ret;
mmu_map_sg(pfdev, bo->node.start << PAGE_SHIFT, prot, sgt);
pm_runtime_mark_last_busy(pfdev->dev); pm_runtime_mark_last_busy(pfdev->dev);
pm_runtime_put_autosuspend(pfdev->dev); pm_runtime_put_autosuspend(pfdev->dev);
bo->is_mapped = true; bo->is_mapped = true;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment