Commit 2552d3a2 authored by Jason Gunthorpe's avatar Jason Gunthorpe Committed by Joerg Roedel

iommu/intel: Add a gfp parameter to alloc_pgtable_page()

This is eventually called by iommufd through intel_iommu_map_pages() and
it should not be forced to atomic. Push the GFP_ATOMIC to all callers.
Reviewed-by: default avatarKevin Tian <kevin.tian@intel.com>
Reviewed-by: default avatarLu Baolu <baolu.lu@linux.intel.com>
Signed-off-by: default avatarJason Gunthorpe <jgg@nvidia.com>
Link: https://lore.kernel.org/r/6-v3-76b587fe28df+6e3-iommu_map_gfp_jgg@nvidia.comSigned-off-by: default avatarJoerg Roedel <jroedel@suse.de>
parent e787a38e
...@@ -362,12 +362,12 @@ static int __init intel_iommu_setup(char *str) ...@@ -362,12 +362,12 @@ static int __init intel_iommu_setup(char *str)
} }
__setup("intel_iommu=", intel_iommu_setup); __setup("intel_iommu=", intel_iommu_setup);
void *alloc_pgtable_page(int node) void *alloc_pgtable_page(int node, gfp_t gfp)
{ {
struct page *page; struct page *page;
void *vaddr = NULL; void *vaddr = NULL;
page = alloc_pages_node(node, GFP_ATOMIC | __GFP_ZERO, 0); page = alloc_pages_node(node, gfp | __GFP_ZERO, 0);
if (page) if (page)
vaddr = page_address(page); vaddr = page_address(page);
return vaddr; return vaddr;
...@@ -612,7 +612,7 @@ struct context_entry *iommu_context_addr(struct intel_iommu *iommu, u8 bus, ...@@ -612,7 +612,7 @@ struct context_entry *iommu_context_addr(struct intel_iommu *iommu, u8 bus,
if (!alloc) if (!alloc)
return NULL; return NULL;
context = alloc_pgtable_page(iommu->node); context = alloc_pgtable_page(iommu->node, GFP_ATOMIC);
if (!context) if (!context)
return NULL; return NULL;
...@@ -935,7 +935,7 @@ static struct dma_pte *pfn_to_dma_pte(struct dmar_domain *domain, ...@@ -935,7 +935,7 @@ static struct dma_pte *pfn_to_dma_pte(struct dmar_domain *domain,
if (!dma_pte_present(pte)) { if (!dma_pte_present(pte)) {
uint64_t pteval; uint64_t pteval;
tmp_page = alloc_pgtable_page(domain->nid); tmp_page = alloc_pgtable_page(domain->nid, GFP_ATOMIC);
if (!tmp_page) if (!tmp_page)
return NULL; return NULL;
...@@ -1186,7 +1186,7 @@ static int iommu_alloc_root_entry(struct intel_iommu *iommu) ...@@ -1186,7 +1186,7 @@ static int iommu_alloc_root_entry(struct intel_iommu *iommu)
{ {
struct root_entry *root; struct root_entry *root;
root = (struct root_entry *)alloc_pgtable_page(iommu->node); root = (struct root_entry *)alloc_pgtable_page(iommu->node, GFP_ATOMIC);
if (!root) { if (!root) {
pr_err("Allocating root entry for %s failed\n", pr_err("Allocating root entry for %s failed\n",
iommu->name); iommu->name);
...@@ -2676,7 +2676,7 @@ static int copy_context_table(struct intel_iommu *iommu, ...@@ -2676,7 +2676,7 @@ static int copy_context_table(struct intel_iommu *iommu,
if (!old_ce) if (!old_ce)
goto out; goto out;
new_ce = alloc_pgtable_page(iommu->node); new_ce = alloc_pgtable_page(iommu->node, GFP_ATOMIC);
if (!new_ce) if (!new_ce)
goto out_unmap; goto out_unmap;
...@@ -4136,7 +4136,7 @@ static int md_domain_init(struct dmar_domain *domain, int guest_width) ...@@ -4136,7 +4136,7 @@ static int md_domain_init(struct dmar_domain *domain, int guest_width)
domain->max_addr = 0; domain->max_addr = 0;
/* always allocate the top pgd */ /* always allocate the top pgd */
domain->pgd = alloc_pgtable_page(domain->nid); domain->pgd = alloc_pgtable_page(domain->nid, GFP_ATOMIC);
if (!domain->pgd) if (!domain->pgd)
return -ENOMEM; return -ENOMEM;
domain_flush_cache(domain, domain->pgd, PAGE_SIZE); domain_flush_cache(domain, domain->pgd, PAGE_SIZE);
......
...@@ -737,7 +737,7 @@ int qi_submit_sync(struct intel_iommu *iommu, struct qi_desc *desc, ...@@ -737,7 +737,7 @@ int qi_submit_sync(struct intel_iommu *iommu, struct qi_desc *desc,
extern int dmar_ir_support(void); extern int dmar_ir_support(void);
void *alloc_pgtable_page(int node); void *alloc_pgtable_page(int node, gfp_t gfp);
void free_pgtable_page(void *vaddr); void free_pgtable_page(void *vaddr);
void iommu_flush_write_buffer(struct intel_iommu *iommu); void iommu_flush_write_buffer(struct intel_iommu *iommu);
struct intel_iommu *device_to_iommu(struct device *dev, u8 *bus, u8 *devfn); struct intel_iommu *device_to_iommu(struct device *dev, u8 *bus, u8 *devfn);
......
...@@ -200,7 +200,7 @@ static struct pasid_entry *intel_pasid_get_entry(struct device *dev, u32 pasid) ...@@ -200,7 +200,7 @@ static struct pasid_entry *intel_pasid_get_entry(struct device *dev, u32 pasid)
retry: retry:
entries = get_pasid_table_from_pde(&dir[dir_index]); entries = get_pasid_table_from_pde(&dir[dir_index]);
if (!entries) { if (!entries) {
entries = alloc_pgtable_page(info->iommu->node); entries = alloc_pgtable_page(info->iommu->node, GFP_ATOMIC);
if (!entries) if (!entries)
return NULL; return NULL;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment