Commit bfa00489 authored by Marek Szyprowski's avatar Marek Szyprowski Committed by Joerg Roedel

iommu/exynos: Rename variables to reflect their purpose

This patch renames some variables to make the code easier to understand.
'domain' is replaced by 'iommu_domain' (more generic entity) and really
meaningless 'priv' by 'domain' to reflect its purpose.
Signed-off-by: default avatarMarek Szyprowski <m.szyprowski@samsung.com>
Tested-by: default avatarJavier Martinez Canillas <javier.martinez@collabora.co.uk>
Signed-off-by: default avatarJoerg Roedel <jroedel@suse.de>
parent 469acebe
...@@ -435,8 +435,8 @@ static void __sysmmu_enable_nocount(struct sysmmu_drvdata *data) ...@@ -435,8 +435,8 @@ static void __sysmmu_enable_nocount(struct sysmmu_drvdata *data)
clk_disable(data->clk_master); clk_disable(data->clk_master);
} }
static int __sysmmu_enable(struct sysmmu_drvdata *data, static int __sysmmu_enable(struct sysmmu_drvdata *data, phys_addr_t pgtable,
phys_addr_t pgtable, struct iommu_domain *domain) struct iommu_domain *iommu_domain)
{ {
int ret = 0; int ret = 0;
unsigned long flags; unsigned long flags;
...@@ -444,7 +444,7 @@ static int __sysmmu_enable(struct sysmmu_drvdata *data, ...@@ -444,7 +444,7 @@ static int __sysmmu_enable(struct sysmmu_drvdata *data,
spin_lock_irqsave(&data->lock, flags); spin_lock_irqsave(&data->lock, flags);
if (set_sysmmu_active(data)) { if (set_sysmmu_active(data)) {
data->pgtable = pgtable; data->pgtable = pgtable;
data->domain = domain; data->domain = iommu_domain;
__sysmmu_enable_nocount(data); __sysmmu_enable_nocount(data);
...@@ -609,91 +609,91 @@ static inline void pgtable_flush(void *vastart, void *vaend) ...@@ -609,91 +609,91 @@ static inline void pgtable_flush(void *vastart, void *vaend)
static struct iommu_domain *exynos_iommu_domain_alloc(unsigned type) static struct iommu_domain *exynos_iommu_domain_alloc(unsigned type)
{ {
struct exynos_iommu_domain *exynos_domain; struct exynos_iommu_domain *domain;
int i; int i;
if (type != IOMMU_DOMAIN_UNMANAGED) if (type != IOMMU_DOMAIN_UNMANAGED)
return NULL; return NULL;
exynos_domain = kzalloc(sizeof(*exynos_domain), GFP_KERNEL); domain = kzalloc(sizeof(*domain), GFP_KERNEL);
if (!exynos_domain) if (!domain)
return NULL; return NULL;
exynos_domain->pgtable = (sysmmu_pte_t *)__get_free_pages(GFP_KERNEL, 2); domain->pgtable = (sysmmu_pte_t *)__get_free_pages(GFP_KERNEL, 2);
if (!exynos_domain->pgtable) if (!domain->pgtable)
goto err_pgtable; goto err_pgtable;
exynos_domain->lv2entcnt = (short *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, 1); domain->lv2entcnt = (short *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, 1);
if (!exynos_domain->lv2entcnt) if (!domain->lv2entcnt)
goto err_counter; goto err_counter;
/* Workaround for System MMU v3.3 to prevent caching 1MiB mapping */ /* Workaround for System MMU v3.3 to prevent caching 1MiB mapping */
for (i = 0; i < NUM_LV1ENTRIES; i += 8) { for (i = 0; i < NUM_LV1ENTRIES; i += 8) {
exynos_domain->pgtable[i + 0] = ZERO_LV2LINK; domain->pgtable[i + 0] = ZERO_LV2LINK;
exynos_domain->pgtable[i + 1] = ZERO_LV2LINK; domain->pgtable[i + 1] = ZERO_LV2LINK;
exynos_domain->pgtable[i + 2] = ZERO_LV2LINK; domain->pgtable[i + 2] = ZERO_LV2LINK;
exynos_domain->pgtable[i + 3] = ZERO_LV2LINK; domain->pgtable[i + 3] = ZERO_LV2LINK;
exynos_domain->pgtable[i + 4] = ZERO_LV2LINK; domain->pgtable[i + 4] = ZERO_LV2LINK;
exynos_domain->pgtable[i + 5] = ZERO_LV2LINK; domain->pgtable[i + 5] = ZERO_LV2LINK;
exynos_domain->pgtable[i + 6] = ZERO_LV2LINK; domain->pgtable[i + 6] = ZERO_LV2LINK;
exynos_domain->pgtable[i + 7] = ZERO_LV2LINK; domain->pgtable[i + 7] = ZERO_LV2LINK;
} }
pgtable_flush(exynos_domain->pgtable, exynos_domain->pgtable + NUM_LV1ENTRIES); pgtable_flush(domain->pgtable, domain->pgtable + NUM_LV1ENTRIES);
spin_lock_init(&exynos_domain->lock); spin_lock_init(&domain->lock);
spin_lock_init(&exynos_domain->pgtablelock); spin_lock_init(&domain->pgtablelock);
INIT_LIST_HEAD(&exynos_domain->clients); INIT_LIST_HEAD(&domain->clients);
exynos_domain->domain.geometry.aperture_start = 0; domain->domain.geometry.aperture_start = 0;
exynos_domain->domain.geometry.aperture_end = ~0UL; domain->domain.geometry.aperture_end = ~0UL;
exynos_domain->domain.geometry.force_aperture = true; domain->domain.geometry.force_aperture = true;
return &exynos_domain->domain; return &domain->domain;
err_counter: err_counter:
free_pages((unsigned long)exynos_domain->pgtable, 2); free_pages((unsigned long)domain->pgtable, 2);
err_pgtable: err_pgtable:
kfree(exynos_domain); kfree(domain);
return NULL; return NULL;
} }
static void exynos_iommu_domain_free(struct iommu_domain *domain) static void exynos_iommu_domain_free(struct iommu_domain *iommu_domain)
{ {
struct exynos_iommu_domain *priv = to_exynos_domain(domain); struct exynos_iommu_domain *domain = to_exynos_domain(iommu_domain);
struct sysmmu_drvdata *data, *next; struct sysmmu_drvdata *data, *next;
unsigned long flags; unsigned long flags;
int i; int i;
WARN_ON(!list_empty(&priv->clients)); WARN_ON(!list_empty(&domain->clients));
spin_lock_irqsave(&priv->lock, flags); spin_lock_irqsave(&domain->lock, flags);
list_for_each_entry_safe(data, next, &priv->clients, domain_node) { list_for_each_entry_safe(data, next, &domain->clients, domain_node) {
if (__sysmmu_disable(data)) if (__sysmmu_disable(data))
data->master = NULL; data->master = NULL;
list_del_init(&data->domain_node); list_del_init(&data->domain_node);
} }
spin_unlock_irqrestore(&priv->lock, flags); spin_unlock_irqrestore(&domain->lock, flags);
for (i = 0; i < NUM_LV1ENTRIES; i++) for (i = 0; i < NUM_LV1ENTRIES; i++)
if (lv1ent_page(priv->pgtable + i)) if (lv1ent_page(domain->pgtable + i))
kmem_cache_free(lv2table_kmem_cache, kmem_cache_free(lv2table_kmem_cache,
phys_to_virt(lv2table_base(priv->pgtable + i))); phys_to_virt(lv2table_base(domain->pgtable + i)));
free_pages((unsigned long)priv->pgtable, 2); free_pages((unsigned long)domain->pgtable, 2);
free_pages((unsigned long)priv->lv2entcnt, 1); free_pages((unsigned long)domain->lv2entcnt, 1);
kfree(priv); kfree(domain);
} }
static int exynos_iommu_attach_device(struct iommu_domain *domain, static int exynos_iommu_attach_device(struct iommu_domain *iommu_domain,
struct device *dev) struct device *dev)
{ {
struct exynos_iommu_owner *owner = dev->archdata.iommu; struct exynos_iommu_owner *owner = dev->archdata.iommu;
struct exynos_iommu_domain *priv = to_exynos_domain(domain); struct exynos_iommu_domain *domain = to_exynos_domain(iommu_domain);
struct sysmmu_drvdata *data; struct sysmmu_drvdata *data;
phys_addr_t pagetable = virt_to_phys(priv->pgtable); phys_addr_t pagetable = virt_to_phys(domain->pgtable);
unsigned long flags; unsigned long flags;
int ret = -ENODEV; int ret = -ENODEV;
...@@ -702,13 +702,13 @@ static int exynos_iommu_attach_device(struct iommu_domain *domain, ...@@ -702,13 +702,13 @@ static int exynos_iommu_attach_device(struct iommu_domain *domain,
data = dev_get_drvdata(owner->sysmmu); data = dev_get_drvdata(owner->sysmmu);
if (data) { if (data) {
ret = __sysmmu_enable(data, pagetable, domain); ret = __sysmmu_enable(data, pagetable, iommu_domain);
if (ret >= 0) { if (ret >= 0) {
data->master = dev; data->master = dev;
spin_lock_irqsave(&priv->lock, flags); spin_lock_irqsave(&domain->lock, flags);
list_add_tail(&data->domain_node, &priv->clients); list_add_tail(&data->domain_node, &domain->clients);
spin_unlock_irqrestore(&priv->lock, flags); spin_unlock_irqrestore(&domain->lock, flags);
} }
} }
...@@ -724,11 +724,11 @@ static int exynos_iommu_attach_device(struct iommu_domain *domain, ...@@ -724,11 +724,11 @@ static int exynos_iommu_attach_device(struct iommu_domain *domain,
return ret; return ret;
} }
static void exynos_iommu_detach_device(struct iommu_domain *domain, static void exynos_iommu_detach_device(struct iommu_domain *iommu_domain,
struct device *dev) struct device *dev)
{ {
struct exynos_iommu_domain *priv = to_exynos_domain(domain); struct exynos_iommu_domain *domain = to_exynos_domain(iommu_domain);
phys_addr_t pagetable = virt_to_phys(priv->pgtable); phys_addr_t pagetable = virt_to_phys(domain->pgtable);
struct sysmmu_drvdata *data; struct sysmmu_drvdata *data;
unsigned long flags; unsigned long flags;
bool found = false; bool found = false;
...@@ -736,8 +736,8 @@ static void exynos_iommu_detach_device(struct iommu_domain *domain, ...@@ -736,8 +736,8 @@ static void exynos_iommu_detach_device(struct iommu_domain *domain,
if (!has_sysmmu(dev)) if (!has_sysmmu(dev))
return; return;
spin_lock_irqsave(&priv->lock, flags); spin_lock_irqsave(&domain->lock, flags);
list_for_each_entry(data, &priv->clients, domain_node) { list_for_each_entry(data, &domain->clients, domain_node) {
if (data->master == dev) { if (data->master == dev) {
if (__sysmmu_disable(data)) { if (__sysmmu_disable(data)) {
data->master = NULL; data->master = NULL;
...@@ -747,7 +747,7 @@ static void exynos_iommu_detach_device(struct iommu_domain *domain, ...@@ -747,7 +747,7 @@ static void exynos_iommu_detach_device(struct iommu_domain *domain,
break; break;
} }
} }
spin_unlock_irqrestore(&priv->lock, flags); spin_unlock_irqrestore(&domain->lock, flags);
if (found) if (found)
dev_dbg(dev, "%s: Detached IOMMU with pgtable %pa\n", dev_dbg(dev, "%s: Detached IOMMU with pgtable %pa\n",
...@@ -756,7 +756,7 @@ static void exynos_iommu_detach_device(struct iommu_domain *domain, ...@@ -756,7 +756,7 @@ static void exynos_iommu_detach_device(struct iommu_domain *domain,
dev_err(dev, "%s: No IOMMU is attached\n", __func__); dev_err(dev, "%s: No IOMMU is attached\n", __func__);
} }
static sysmmu_pte_t *alloc_lv2entry(struct exynos_iommu_domain *priv, static sysmmu_pte_t *alloc_lv2entry(struct exynos_iommu_domain *domain,
sysmmu_pte_t *sent, sysmmu_iova_t iova, short *pgcounter) sysmmu_pte_t *sent, sysmmu_iova_t iova, short *pgcounter)
{ {
if (lv1ent_section(sent)) { if (lv1ent_section(sent)) {
...@@ -799,17 +799,17 @@ static sysmmu_pte_t *alloc_lv2entry(struct exynos_iommu_domain *priv, ...@@ -799,17 +799,17 @@ static sysmmu_pte_t *alloc_lv2entry(struct exynos_iommu_domain *priv,
if (need_flush_flpd_cache) { if (need_flush_flpd_cache) {
struct sysmmu_drvdata *data; struct sysmmu_drvdata *data;
spin_lock(&priv->lock); spin_lock(&domain->lock);
list_for_each_entry(data, &priv->clients, domain_node) list_for_each_entry(data, &domain->clients, domain_node)
sysmmu_tlb_invalidate_flpdcache(data, iova); sysmmu_tlb_invalidate_flpdcache(data, iova);
spin_unlock(&priv->lock); spin_unlock(&domain->lock);
} }
} }
return page_entry(sent, iova); return page_entry(sent, iova);
} }
static int lv1set_section(struct exynos_iommu_domain *priv, static int lv1set_section(struct exynos_iommu_domain *domain,
sysmmu_pte_t *sent, sysmmu_iova_t iova, sysmmu_pte_t *sent, sysmmu_iova_t iova,
phys_addr_t paddr, short *pgcnt) phys_addr_t paddr, short *pgcnt)
{ {
...@@ -834,17 +834,17 @@ static int lv1set_section(struct exynos_iommu_domain *priv, ...@@ -834,17 +834,17 @@ static int lv1set_section(struct exynos_iommu_domain *priv,
pgtable_flush(sent, sent + 1); pgtable_flush(sent, sent + 1);
spin_lock(&priv->lock); spin_lock(&domain->lock);
if (lv1ent_page_zero(sent)) { if (lv1ent_page_zero(sent)) {
struct sysmmu_drvdata *data; struct sysmmu_drvdata *data;
/* /*
* Flushing FLPD cache in System MMU v3.3 that may cache a FLPD * Flushing FLPD cache in System MMU v3.3 that may cache a FLPD
* entry by speculative prefetch of SLPD which has no mapping. * entry by speculative prefetch of SLPD which has no mapping.
*/ */
list_for_each_entry(data, &priv->clients, domain_node) list_for_each_entry(data, &domain->clients, domain_node)
sysmmu_tlb_invalidate_flpdcache(data, iova); sysmmu_tlb_invalidate_flpdcache(data, iova);
} }
spin_unlock(&priv->lock); spin_unlock(&domain->lock);
return 0; return 0;
} }
...@@ -904,74 +904,75 @@ static int lv2set_page(sysmmu_pte_t *pent, phys_addr_t paddr, size_t size, ...@@ -904,74 +904,75 @@ static int lv2set_page(sysmmu_pte_t *pent, phys_addr_t paddr, size_t size,
* than or equal to 128KiB. * than or equal to 128KiB.
* - Start address of an I/O virtual region must be aligned by 128KiB. * - Start address of an I/O virtual region must be aligned by 128KiB.
*/ */
static int exynos_iommu_map(struct iommu_domain *domain, unsigned long l_iova, static int exynos_iommu_map(struct iommu_domain *iommu_domain,
phys_addr_t paddr, size_t size, int prot) unsigned long l_iova, phys_addr_t paddr, size_t size,
int prot)
{ {
struct exynos_iommu_domain *priv = to_exynos_domain(domain); struct exynos_iommu_domain *domain = to_exynos_domain(iommu_domain);
sysmmu_pte_t *entry; sysmmu_pte_t *entry;
sysmmu_iova_t iova = (sysmmu_iova_t)l_iova; sysmmu_iova_t iova = (sysmmu_iova_t)l_iova;
unsigned long flags; unsigned long flags;
int ret = -ENOMEM; int ret = -ENOMEM;
BUG_ON(priv->pgtable == NULL); BUG_ON(domain->pgtable == NULL);
spin_lock_irqsave(&priv->pgtablelock, flags); spin_lock_irqsave(&domain->pgtablelock, flags);
entry = section_entry(priv->pgtable, iova); entry = section_entry(domain->pgtable, iova);
if (size == SECT_SIZE) { if (size == SECT_SIZE) {
ret = lv1set_section(priv, entry, iova, paddr, ret = lv1set_section(domain, entry, iova, paddr,
&priv->lv2entcnt[lv1ent_offset(iova)]); &domain->lv2entcnt[lv1ent_offset(iova)]);
} else { } else {
sysmmu_pte_t *pent; sysmmu_pte_t *pent;
pent = alloc_lv2entry(priv, entry, iova, pent = alloc_lv2entry(domain, entry, iova,
&priv->lv2entcnt[lv1ent_offset(iova)]); &domain->lv2entcnt[lv1ent_offset(iova)]);
if (IS_ERR(pent)) if (IS_ERR(pent))
ret = PTR_ERR(pent); ret = PTR_ERR(pent);
else else
ret = lv2set_page(pent, paddr, size, ret = lv2set_page(pent, paddr, size,
&priv->lv2entcnt[lv1ent_offset(iova)]); &domain->lv2entcnt[lv1ent_offset(iova)]);
} }
if (ret) if (ret)
pr_err("%s: Failed(%d) to map %#zx bytes @ %#x\n", pr_err("%s: Failed(%d) to map %#zx bytes @ %#x\n",
__func__, ret, size, iova); __func__, ret, size, iova);
spin_unlock_irqrestore(&priv->pgtablelock, flags); spin_unlock_irqrestore(&domain->pgtablelock, flags);
return ret; return ret;
} }
static void exynos_iommu_tlb_invalidate_entry(struct exynos_iommu_domain *priv, static void exynos_iommu_tlb_invalidate_entry(struct exynos_iommu_domain *domain,
sysmmu_iova_t iova, size_t size) sysmmu_iova_t iova, size_t size)
{ {
struct sysmmu_drvdata *data; struct sysmmu_drvdata *data;
unsigned long flags; unsigned long flags;
spin_lock_irqsave(&priv->lock, flags); spin_lock_irqsave(&domain->lock, flags);
list_for_each_entry(data, &priv->clients, domain_node) list_for_each_entry(data, &domain->clients, domain_node)
sysmmu_tlb_invalidate_entry(data, iova, size); sysmmu_tlb_invalidate_entry(data, iova, size);
spin_unlock_irqrestore(&priv->lock, flags); spin_unlock_irqrestore(&domain->lock, flags);
} }
static size_t exynos_iommu_unmap(struct iommu_domain *domain, static size_t exynos_iommu_unmap(struct iommu_domain *iommu_domain,
unsigned long l_iova, size_t size) unsigned long l_iova, size_t size)
{ {
struct exynos_iommu_domain *priv = to_exynos_domain(domain); struct exynos_iommu_domain *domain = to_exynos_domain(iommu_domain);
sysmmu_iova_t iova = (sysmmu_iova_t)l_iova; sysmmu_iova_t iova = (sysmmu_iova_t)l_iova;
sysmmu_pte_t *ent; sysmmu_pte_t *ent;
size_t err_pgsize; size_t err_pgsize;
unsigned long flags; unsigned long flags;
BUG_ON(priv->pgtable == NULL); BUG_ON(domain->pgtable == NULL);
spin_lock_irqsave(&priv->pgtablelock, flags); spin_lock_irqsave(&domain->pgtablelock, flags);
ent = section_entry(priv->pgtable, iova); ent = section_entry(domain->pgtable, iova);
if (lv1ent_section(ent)) { if (lv1ent_section(ent)) {
if (WARN_ON(size < SECT_SIZE)) { if (WARN_ON(size < SECT_SIZE)) {
...@@ -1005,7 +1006,7 @@ static size_t exynos_iommu_unmap(struct iommu_domain *domain, ...@@ -1005,7 +1006,7 @@ static size_t exynos_iommu_unmap(struct iommu_domain *domain,
*ent = 0; *ent = 0;
size = SPAGE_SIZE; size = SPAGE_SIZE;
pgtable_flush(ent, ent + 1); pgtable_flush(ent, ent + 1);
priv->lv2entcnt[lv1ent_offset(iova)] += 1; domain->lv2entcnt[lv1ent_offset(iova)] += 1;
goto done; goto done;
} }
...@@ -1019,15 +1020,15 @@ static size_t exynos_iommu_unmap(struct iommu_domain *domain, ...@@ -1019,15 +1020,15 @@ static size_t exynos_iommu_unmap(struct iommu_domain *domain,
pgtable_flush(ent, ent + SPAGES_PER_LPAGE); pgtable_flush(ent, ent + SPAGES_PER_LPAGE);
size = LPAGE_SIZE; size = LPAGE_SIZE;
priv->lv2entcnt[lv1ent_offset(iova)] += SPAGES_PER_LPAGE; domain->lv2entcnt[lv1ent_offset(iova)] += SPAGES_PER_LPAGE;
done: done:
spin_unlock_irqrestore(&priv->pgtablelock, flags); spin_unlock_irqrestore(&domain->pgtablelock, flags);
exynos_iommu_tlb_invalidate_entry(priv, iova, size); exynos_iommu_tlb_invalidate_entry(domain, iova, size);
return size; return size;
err: err:
spin_unlock_irqrestore(&priv->pgtablelock, flags); spin_unlock_irqrestore(&domain->pgtablelock, flags);
pr_err("%s: Failed: size(%#zx) @ %#x is smaller than page size %#zx\n", pr_err("%s: Failed: size(%#zx) @ %#x is smaller than page size %#zx\n",
__func__, size, iova, err_pgsize); __func__, size, iova, err_pgsize);
...@@ -1035,17 +1036,17 @@ static size_t exynos_iommu_unmap(struct iommu_domain *domain, ...@@ -1035,17 +1036,17 @@ static size_t exynos_iommu_unmap(struct iommu_domain *domain,
return 0; return 0;
} }
static phys_addr_t exynos_iommu_iova_to_phys(struct iommu_domain *domain, static phys_addr_t exynos_iommu_iova_to_phys(struct iommu_domain *iommu_domain,
dma_addr_t iova) dma_addr_t iova)
{ {
struct exynos_iommu_domain *priv = to_exynos_domain(domain); struct exynos_iommu_domain *domain = to_exynos_domain(iommu_domain);
sysmmu_pte_t *entry; sysmmu_pte_t *entry;
unsigned long flags; unsigned long flags;
phys_addr_t phys = 0; phys_addr_t phys = 0;
spin_lock_irqsave(&priv->pgtablelock, flags); spin_lock_irqsave(&domain->pgtablelock, flags);
entry = section_entry(priv->pgtable, iova); entry = section_entry(domain->pgtable, iova);
if (lv1ent_section(entry)) { if (lv1ent_section(entry)) {
phys = section_phys(entry) + section_offs(iova); phys = section_phys(entry) + section_offs(iova);
...@@ -1058,7 +1059,7 @@ static phys_addr_t exynos_iommu_iova_to_phys(struct iommu_domain *domain, ...@@ -1058,7 +1059,7 @@ static phys_addr_t exynos_iommu_iova_to_phys(struct iommu_domain *domain,
phys = spage_phys(entry) + spage_offs(iova); phys = spage_phys(entry) + spage_offs(iova);
} }
spin_unlock_irqrestore(&priv->pgtablelock, flags); spin_unlock_irqrestore(&domain->pgtablelock, flags);
return phys; return phys;
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment