Commit 2f8f0de8 authored by Ohad Sharabi's avatar Ohad Sharabi Committed by Greg Kroah-Hartman

habanalabs: change mmu_get_real_page_size to be ASIC-specific

This patch breaks the cumbersome implementation of "get real page size"
along with it's multiple inner conditions and implement each case
(according to the real complexity) inside an ASIC function.
Signed-off-by: default avatarOhad Sharabi <osharabi@habana.ai>
Reviewed-by: default avatarOded Gabbay <ogabbay@kernel.org>
Signed-off-by: default avatarOded Gabbay <ogabbay@kernel.org>
Signed-off-by: default avatarGreg Kroah-Hartman <gregkh@linuxfoundation.org>
parent 1359fcbe
...@@ -1450,6 +1450,8 @@ struct hl_asic_funcs { ...@@ -1450,6 +1450,8 @@ struct hl_asic_funcs {
void (*set_pci_memory_regions)(struct hl_device *hdev); void (*set_pci_memory_regions)(struct hl_device *hdev);
u32* (*get_stream_master_qid_arr)(void); u32* (*get_stream_master_qid_arr)(void);
bool (*is_valid_dram_page_size)(u32 page_size); bool (*is_valid_dram_page_size)(u32 page_size);
int (*mmu_get_real_page_size)(struct hl_device *hdev, struct hl_mmu_properties *mmu_prop,
u32 page_size, u32 *real_page_size, bool is_dram_addr);
}; };
...@@ -3109,6 +3111,8 @@ int hl_mmu_ctx_init(struct hl_ctx *ctx); ...@@ -3109,6 +3111,8 @@ int hl_mmu_ctx_init(struct hl_ctx *ctx);
void hl_mmu_ctx_fini(struct hl_ctx *ctx); void hl_mmu_ctx_fini(struct hl_ctx *ctx);
int hl_mmu_map_page(struct hl_ctx *ctx, u64 virt_addr, u64 phys_addr, int hl_mmu_map_page(struct hl_ctx *ctx, u64 virt_addr, u64 phys_addr,
u32 page_size, bool flush_pte); u32 page_size, bool flush_pte);
int hl_mmu_get_real_page_size(struct hl_device *hdev, struct hl_mmu_properties *mmu_prop,
u32 page_size, u32 *real_page_size, bool is_dram_addr);
int hl_mmu_unmap_page(struct hl_ctx *ctx, u64 virt_addr, u32 page_size, int hl_mmu_unmap_page(struct hl_ctx *ctx, u64 virt_addr, u32 page_size,
bool flush_pte); bool flush_pte);
int hl_mmu_map_contiguous(struct hl_ctx *ctx, u64 virt_addr, int hl_mmu_map_contiguous(struct hl_ctx *ctx, u64 virt_addr,
......
...@@ -9,6 +9,20 @@ ...@@ -9,6 +9,20 @@
#include "../habanalabs.h" #include "../habanalabs.h"
/**
* hl_mmu_get_funcs() - get MMU functions structure
* @hdev: habanalabs device structure.
* @pgt_residency: page table residency.
* @is_dram_addr: true if we need HMMU functions
*
* @return appropriate MMU functions structure
*/
static struct hl_mmu_funcs *hl_mmu_get_funcs(struct hl_device *hdev, int pgt_residency,
bool is_dram_addr)
{
return &hdev->mmu_func[pgt_residency];
}
bool hl_is_dram_va(struct hl_device *hdev, u64 virt_addr) bool hl_is_dram_va(struct hl_device *hdev, u64 virt_addr)
{ {
struct asic_fixed_properties *prop = &hdev->asic_prop; struct asic_fixed_properties *prop = &hdev->asic_prop;
...@@ -121,6 +135,53 @@ void hl_mmu_ctx_fini(struct hl_ctx *ctx) ...@@ -121,6 +135,53 @@ void hl_mmu_ctx_fini(struct hl_ctx *ctx)
mutex_destroy(&ctx->mmu_lock); mutex_destroy(&ctx->mmu_lock);
} }
/*
* hl_mmu_get_real_page_size - get real page size to use in map/unmap operation
*
* @hdev: pointer to device data.
* @mmu_prop: MMU properties.
* @page_size: page size
* @real_page_size: set here the actual page size to use for the operation
* @is_dram_addr: true if DRAM address, otherwise false.
*
* @return 0 on success, otherwise non 0 error code
*
* note that this is general implementation that can fit most MMU arch. but as this is used as an
* MMU function:
* 1. it shall not be called directly- only from mmu_func structure instance
* 2. each MMU may modify the implementation internally
*/
int hl_mmu_get_real_page_size(struct hl_device *hdev, struct hl_mmu_properties *mmu_prop,
u32 page_size, u32 *real_page_size, bool is_dram_addr)
{
/*
* The H/W handles mapping of specific page sizes. Hence if the page
* size is bigger, we break it to sub-pages and map them separately.
*/
if ((page_size % mmu_prop->page_size) == 0) {
*real_page_size = mmu_prop->page_size;
return 0;
}
dev_err(hdev->dev, "page size of %u is not %uKB aligned, can't map\n",
page_size, mmu_prop->page_size >> 10);
return -EFAULT;
}
static struct hl_mmu_properties *hl_mmu_get_prop(struct hl_device *hdev, u32 page_size,
bool is_dram_addr)
{
struct asic_fixed_properties *prop = &hdev->asic_prop;
if (is_dram_addr)
return &prop->dmmu;
else if ((page_size % prop->pmmu_huge.page_size) == 0)
return &prop->pmmu_huge;
return &prop->pmmu;
}
/* /*
* hl_mmu_unmap_page - unmaps a virtual addr * hl_mmu_unmap_page - unmaps a virtual addr
* *
...@@ -142,60 +203,35 @@ void hl_mmu_ctx_fini(struct hl_ctx *ctx) ...@@ -142,60 +203,35 @@ void hl_mmu_ctx_fini(struct hl_ctx *ctx)
* For optimization reasons PCI flush may be requested once after unmapping of * For optimization reasons PCI flush may be requested once after unmapping of
* large area. * large area.
*/ */
int hl_mmu_unmap_page(struct hl_ctx *ctx, u64 virt_addr, u32 page_size, int hl_mmu_unmap_page(struct hl_ctx *ctx, u64 virt_addr, u32 page_size, bool flush_pte)
bool flush_pte)
{ {
struct hl_device *hdev = ctx->hdev; struct hl_device *hdev = ctx->hdev;
struct asic_fixed_properties *prop = &hdev->asic_prop;
struct hl_mmu_properties *mmu_prop; struct hl_mmu_properties *mmu_prop;
u64 real_virt_addr; struct hl_mmu_funcs *mmu_funcs;
int i, pgt_residency, rc = 0;
u32 real_page_size, npages; u32 real_page_size, npages;
int i, rc = 0, pgt_residency; u64 real_virt_addr;
bool is_dram_addr; bool is_dram_addr;
if (!hdev->mmu_enable) if (!hdev->mmu_enable)
return 0; return 0;
is_dram_addr = hl_is_dram_va(hdev, virt_addr); is_dram_addr = hl_is_dram_va(hdev, virt_addr);
mmu_prop = hl_mmu_get_prop(hdev, page_size, is_dram_addr);
if (is_dram_addr)
mmu_prop = &prop->dmmu;
else if ((page_size % prop->pmmu_huge.page_size) == 0)
mmu_prop = &prop->pmmu_huge;
else
mmu_prop = &prop->pmmu;
pgt_residency = mmu_prop->host_resident ? MMU_HR_PGT : MMU_DR_PGT; pgt_residency = mmu_prop->host_resident ? MMU_HR_PGT : MMU_DR_PGT;
/* mmu_funcs = hl_mmu_get_funcs(hdev, pgt_residency, is_dram_addr);
* The H/W handles mapping of specific page sizes. Hence if the page
* size is bigger, we break it to sub-pages and unmap them separately.
*/
if ((page_size % mmu_prop->page_size) == 0) {
real_page_size = mmu_prop->page_size;
} else {
/*
* MMU page size may differ from DRAM page size.
* In such case work with the DRAM page size and let the MMU
* scrambling routine to handle this mismatch when
* calculating the address to remove from the MMU page table
*/
if (is_dram_addr && ((page_size % prop->dram_page_size) == 0)) {
real_page_size = prop->dram_page_size;
} else {
dev_err(hdev->dev,
"page size of %u is not %uKB aligned, can't unmap\n",
page_size, mmu_prop->page_size >> 10);
return -EFAULT; rc = hdev->asic_funcs->mmu_get_real_page_size(hdev, mmu_prop, page_size, &real_page_size,
} is_dram_addr);
} if (rc)
return rc;
npages = page_size / real_page_size; npages = page_size / real_page_size;
real_virt_addr = virt_addr; real_virt_addr = virt_addr;
for (i = 0 ; i < npages ; i++) { for (i = 0 ; i < npages ; i++) {
rc = hdev->mmu_func[pgt_residency].unmap(ctx, rc = mmu_funcs->unmap(ctx, real_virt_addr, is_dram_addr);
real_virt_addr, is_dram_addr);
if (rc) if (rc)
break; break;
...@@ -203,7 +239,7 @@ int hl_mmu_unmap_page(struct hl_ctx *ctx, u64 virt_addr, u32 page_size, ...@@ -203,7 +239,7 @@ int hl_mmu_unmap_page(struct hl_ctx *ctx, u64 virt_addr, u32 page_size,
} }
if (flush_pte) if (flush_pte)
hdev->mmu_func[pgt_residency].flush(ctx); mmu_funcs->flush(ctx);
return rc; return rc;
} }
...@@ -230,15 +266,15 @@ int hl_mmu_unmap_page(struct hl_ctx *ctx, u64 virt_addr, u32 page_size, ...@@ -230,15 +266,15 @@ int hl_mmu_unmap_page(struct hl_ctx *ctx, u64 virt_addr, u32 page_size,
* For optimization reasons PCI flush may be requested once after mapping of * For optimization reasons PCI flush may be requested once after mapping of
* large area. * large area.
*/ */
int hl_mmu_map_page(struct hl_ctx *ctx, u64 virt_addr, u64 phys_addr, int hl_mmu_map_page(struct hl_ctx *ctx, u64 virt_addr, u64 phys_addr, u32 page_size,
u32 page_size, bool flush_pte) bool flush_pte)
{ {
int i, rc, pgt_residency, mapped_cnt = 0;
struct hl_device *hdev = ctx->hdev; struct hl_device *hdev = ctx->hdev;
struct asic_fixed_properties *prop = &hdev->asic_prop;
struct hl_mmu_properties *mmu_prop; struct hl_mmu_properties *mmu_prop;
u64 real_virt_addr, real_phys_addr; u64 real_virt_addr, real_phys_addr;
struct hl_mmu_funcs *mmu_funcs;
u32 real_page_size, npages; u32 real_page_size, npages;
int i, rc, pgt_residency, mapped_cnt = 0;
bool is_dram_addr; bool is_dram_addr;
...@@ -246,40 +282,15 @@ int hl_mmu_map_page(struct hl_ctx *ctx, u64 virt_addr, u64 phys_addr, ...@@ -246,40 +282,15 @@ int hl_mmu_map_page(struct hl_ctx *ctx, u64 virt_addr, u64 phys_addr,
return 0; return 0;
is_dram_addr = hl_is_dram_va(hdev, virt_addr); is_dram_addr = hl_is_dram_va(hdev, virt_addr);
mmu_prop = hl_mmu_get_prop(hdev, page_size, is_dram_addr);
if (is_dram_addr)
mmu_prop = &prop->dmmu;
else if ((page_size % prop->pmmu_huge.page_size) == 0)
mmu_prop = &prop->pmmu_huge;
else
mmu_prop = &prop->pmmu;
pgt_residency = mmu_prop->host_resident ? MMU_HR_PGT : MMU_DR_PGT; pgt_residency = mmu_prop->host_resident ? MMU_HR_PGT : MMU_DR_PGT;
mmu_funcs = hl_mmu_get_funcs(hdev, pgt_residency, is_dram_addr);
/* rc = hdev->asic_funcs->mmu_get_real_page_size(hdev, mmu_prop, page_size, &real_page_size,
* The H/W handles mapping of specific page sizes. Hence if the page is_dram_addr);
* size is bigger, we break it to sub-pages and map them separately. if (rc)
*/ return rc;
if ((page_size % mmu_prop->page_size) == 0) {
real_page_size = mmu_prop->page_size;
} else if (is_dram_addr && ((page_size % prop->dram_page_size) == 0) &&
(prop->dram_page_size < mmu_prop->page_size)) {
/*
* MMU page size may differ from DRAM page size.
* In such case work with the DRAM page size and let the MMU
* scrambling routine handle this mismatch when calculating
* the address to place in the MMU page table. (in that case
* also make sure that the dram_page_size smaller than the
* mmu page size)
*/
real_page_size = prop->dram_page_size;
} else {
dev_err(hdev->dev,
"page size of %u is not %uKB aligned, can't map\n",
page_size, mmu_prop->page_size >> 10);
return -EFAULT;
}
/* /*
* Verify that the phys and virt addresses are aligned with the * Verify that the phys and virt addresses are aligned with the
...@@ -302,9 +313,8 @@ int hl_mmu_map_page(struct hl_ctx *ctx, u64 virt_addr, u64 phys_addr, ...@@ -302,9 +313,8 @@ int hl_mmu_map_page(struct hl_ctx *ctx, u64 virt_addr, u64 phys_addr,
real_phys_addr = phys_addr; real_phys_addr = phys_addr;
for (i = 0 ; i < npages ; i++) { for (i = 0 ; i < npages ; i++) {
rc = hdev->mmu_func[pgt_residency].map(ctx, rc = mmu_funcs->map(ctx, real_virt_addr, real_phys_addr, real_page_size,
real_virt_addr, real_phys_addr, is_dram_addr);
real_page_size, is_dram_addr);
if (rc) if (rc)
goto err; goto err;
...@@ -314,22 +324,21 @@ int hl_mmu_map_page(struct hl_ctx *ctx, u64 virt_addr, u64 phys_addr, ...@@ -314,22 +324,21 @@ int hl_mmu_map_page(struct hl_ctx *ctx, u64 virt_addr, u64 phys_addr,
} }
if (flush_pte) if (flush_pte)
hdev->mmu_func[pgt_residency].flush(ctx); mmu_funcs->flush(ctx);
return 0; return 0;
err: err:
real_virt_addr = virt_addr; real_virt_addr = virt_addr;
for (i = 0 ; i < mapped_cnt ; i++) { for (i = 0 ; i < mapped_cnt ; i++) {
if (hdev->mmu_func[pgt_residency].unmap(ctx, if (mmu_funcs->unmap(ctx, real_virt_addr, is_dram_addr))
real_virt_addr, is_dram_addr))
dev_warn_ratelimited(hdev->dev, dev_warn_ratelimited(hdev->dev,
"failed to unmap va: 0x%llx\n", real_virt_addr); "failed to unmap va: 0x%llx\n", real_virt_addr);
real_virt_addr += real_page_size; real_virt_addr += real_page_size;
} }
hdev->mmu_func[pgt_residency].flush(ctx); mmu_funcs->flush(ctx);
return rc; return rc;
} }
...@@ -508,7 +517,7 @@ static void hl_mmu_pa_page_with_offset(struct hl_ctx *ctx, u64 virt_addr, ...@@ -508,7 +517,7 @@ static void hl_mmu_pa_page_with_offset(struct hl_ctx *ctx, u64 virt_addr,
/* /*
* Bit arithmetics cannot be used for non power of two page * Bit arithmetics cannot be used for non power of two page
* sizes. In addition, since bit arithmetics is not used, * sizes. In addition, since bit arithmetics is not used,
* we cannot ignore dram base. All that shall be considerd. * we cannot ignore dram base. All that shall be considered.
*/ */
dram_page_size = prop->dram_page_size; dram_page_size = prop->dram_page_size;
...@@ -557,40 +566,39 @@ int hl_mmu_get_tlb_info(struct hl_ctx *ctx, u64 virt_addr, ...@@ -557,40 +566,39 @@ int hl_mmu_get_tlb_info(struct hl_ctx *ctx, u64 virt_addr,
struct hl_mmu_hop_info *hops) struct hl_mmu_hop_info *hops)
{ {
struct hl_device *hdev = ctx->hdev; struct hl_device *hdev = ctx->hdev;
struct asic_fixed_properties *prop = &hdev->asic_prop; struct asic_fixed_properties *prop;
struct hl_mmu_properties *mmu_prop; struct hl_mmu_properties *mmu_prop;
int rc; struct hl_mmu_funcs *mmu_funcs;
int pgt_residency, rc;
bool is_dram_addr; bool is_dram_addr;
if (!hdev->mmu_enable) if (!hdev->mmu_enable)
return -EOPNOTSUPP; return -EOPNOTSUPP;
prop = &hdev->asic_prop;
hops->scrambled_vaddr = virt_addr; /* assume no scrambling */ hops->scrambled_vaddr = virt_addr; /* assume no scrambling */
is_dram_addr = hl_mem_area_inside_range(virt_addr, prop->dmmu.page_size, is_dram_addr = hl_mem_area_inside_range(virt_addr, prop->dmmu.page_size,
prop->dmmu.start_addr, prop->dmmu.start_addr,
prop->dmmu.end_addr); prop->dmmu.end_addr);
/* host-residency is the same in PMMU and HPMMU, use one of them */ /* host-residency is the same in PMMU and PMMU huge, no need to distinguish here */
mmu_prop = is_dram_addr ? &prop->dmmu : &prop->pmmu; mmu_prop = is_dram_addr ? &prop->dmmu : &prop->pmmu;
pgt_residency = mmu_prop->host_resident ? MMU_HR_PGT : MMU_DR_PGT;
mmu_funcs = hl_mmu_get_funcs(hdev, pgt_residency, is_dram_addr);
mutex_lock(&ctx->mmu_lock); mutex_lock(&ctx->mmu_lock);
rc = mmu_funcs->get_tlb_info(ctx, virt_addr, hops);
if (mmu_prop->host_resident)
rc = hdev->mmu_func[MMU_HR_PGT].get_tlb_info(ctx,
virt_addr, hops);
else
rc = hdev->mmu_func[MMU_DR_PGT].get_tlb_info(ctx,
virt_addr, hops);
mutex_unlock(&ctx->mmu_lock); mutex_unlock(&ctx->mmu_lock);
if (rc)
return rc;
/* add page offset to physical address */ /* add page offset to physical address */
if (hops->unscrambled_paddr) if (hops->unscrambled_paddr)
hl_mmu_pa_page_with_offset(ctx, virt_addr, hops, hl_mmu_pa_page_with_offset(ctx, virt_addr, hops, &hops->unscrambled_paddr);
&hops->unscrambled_paddr);
return rc; return 0;
} }
int hl_mmu_if_set_funcs(struct hl_device *hdev) int hl_mmu_if_set_funcs(struct hl_device *hdev)
......
...@@ -9487,7 +9487,8 @@ static const struct hl_asic_funcs gaudi_funcs = { ...@@ -9487,7 +9487,8 @@ static const struct hl_asic_funcs gaudi_funcs = {
.get_sob_addr = gaudi_get_sob_addr, .get_sob_addr = gaudi_get_sob_addr,
.set_pci_memory_regions = gaudi_set_pci_memory_regions, .set_pci_memory_regions = gaudi_set_pci_memory_regions,
.get_stream_master_qid_arr = gaudi_get_stream_master_qid_arr, .get_stream_master_qid_arr = gaudi_get_stream_master_qid_arr,
.is_valid_dram_page_size = NULL .is_valid_dram_page_size = NULL,
.mmu_get_real_page_size = hl_mmu_get_real_page_size,
}; };
/** /**
......
...@@ -5765,7 +5765,8 @@ static const struct hl_asic_funcs goya_funcs = { ...@@ -5765,7 +5765,8 @@ static const struct hl_asic_funcs goya_funcs = {
.get_sob_addr = &goya_get_sob_addr, .get_sob_addr = &goya_get_sob_addr,
.set_pci_memory_regions = goya_set_pci_memory_regions, .set_pci_memory_regions = goya_set_pci_memory_regions,
.get_stream_master_qid_arr = goya_get_stream_master_qid_arr, .get_stream_master_qid_arr = goya_get_stream_master_qid_arr,
.is_valid_dram_page_size = NULL .is_valid_dram_page_size = NULL,
.mmu_get_real_page_size = hl_mmu_get_real_page_size,
}; };
/* /*
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment