Commit 4dcfa600 authored by Russell King's avatar Russell King

ARM: DMA-API: better handing of DMA masks for coherent allocations

We need to start treating DMA masks as something which is specific to
the bus that the device resides on, otherwise we're going to hit all
sorts of nasty issues with LPAE and 32-bit DMA controllers in >32-bit
systems, where memory is offset from PFN 0.

In order to start doing this, we convert the DMA mask to a PFN using
the device specific dma_to_pfn() macro.  This is the reverse of the
pfn_to_dma() macro which is used to get the DMA address for the device.

This gives us a PFN mask, which we can then check against the PFN
limit of the DMA zone.
Signed-off-by: default avatarRussell King <rmk+kernel@arm.linux.org.uk>
parent 42536b9f
...@@ -159,7 +159,7 @@ EXPORT_SYMBOL(arm_coherent_dma_ops); ...@@ -159,7 +159,7 @@ EXPORT_SYMBOL(arm_coherent_dma_ops);
static u64 get_coherent_dma_mask(struct device *dev) static u64 get_coherent_dma_mask(struct device *dev)
{ {
u64 mask = (u64)arm_dma_limit; u64 mask = (u64)DMA_BIT_MASK(32);
if (dev) { if (dev) {
mask = dev->coherent_dma_mask; mask = dev->coherent_dma_mask;
...@@ -173,10 +173,30 @@ static u64 get_coherent_dma_mask(struct device *dev) ...@@ -173,10 +173,30 @@ static u64 get_coherent_dma_mask(struct device *dev)
return 0; return 0;
} }
if ((~mask) & (u64)arm_dma_limit) { /*
dev_warn(dev, "coherent DMA mask %#llx is smaller " * If the mask allows for more memory than we can address,
"than system GFP_DMA mask %#llx\n", * and we actually have that much memory, then fail the
mask, (u64)arm_dma_limit); * allocation.
*/
if (sizeof(mask) != sizeof(dma_addr_t) &&
mask > (dma_addr_t)~0 &&
dma_to_pfn(dev, ~0) > arm_dma_pfn_limit) {
dev_warn(dev, "Coherent DMA mask %#llx is larger than dma_addr_t allows\n",
mask);
dev_warn(dev, "Driver did not use or check the return value from dma_set_coherent_mask()?\n");
return 0;
}
/*
* Now check that the mask, when translated to a PFN,
* fits within the allowable addresses which we can
* allocate.
*/
if (dma_to_pfn(dev, mask) < arm_dma_pfn_limit) {
dev_warn(dev, "Coherent DMA mask %#llx (pfn %#lx-%#lx) covers a smaller range of system memory than the DMA zone pfn 0x0-%#lx\n",
mask,
dma_to_pfn(dev, 0), dma_to_pfn(dev, mask) + 1,
arm_dma_pfn_limit + 1);
return 0; return 0;
} }
} }
...@@ -1007,8 +1027,27 @@ void arm_dma_sync_sg_for_device(struct device *dev, struct scatterlist *sg, ...@@ -1007,8 +1027,27 @@ void arm_dma_sync_sg_for_device(struct device *dev, struct scatterlist *sg,
*/ */
int dma_supported(struct device *dev, u64 mask) int dma_supported(struct device *dev, u64 mask)
{ {
if (mask < (u64)arm_dma_limit) unsigned long limit;
/*
* If the mask allows for more memory than we can address,
* and we actually have that much memory, then we must
* indicate that DMA to this device is not supported.
*/
if (sizeof(mask) != sizeof(dma_addr_t) &&
mask > (dma_addr_t)~0 &&
dma_to_pfn(dev, ~0) > arm_dma_pfn_limit)
return 0;
/*
* Translate the device's DMA mask to a PFN limit. This
* PFN number includes the page which we can DMA to.
*/
limit = dma_to_pfn(dev, mask);
if (limit < arm_dma_pfn_limit)
return 0; return 0;
return 1; return 1;
} }
EXPORT_SYMBOL(dma_supported); EXPORT_SYMBOL(dma_supported);
......
...@@ -218,6 +218,7 @@ EXPORT_SYMBOL(arm_dma_zone_size); ...@@ -218,6 +218,7 @@ EXPORT_SYMBOL(arm_dma_zone_size);
* so a successful GFP_DMA allocation will always satisfy this. * so a successful GFP_DMA allocation will always satisfy this.
*/ */
phys_addr_t arm_dma_limit; phys_addr_t arm_dma_limit;
unsigned long arm_dma_pfn_limit;
static void __init arm_adjust_dma_zone(unsigned long *size, unsigned long *hole, static void __init arm_adjust_dma_zone(unsigned long *size, unsigned long *hole,
unsigned long dma_size) unsigned long dma_size)
...@@ -240,6 +241,7 @@ void __init setup_dma_zone(const struct machine_desc *mdesc) ...@@ -240,6 +241,7 @@ void __init setup_dma_zone(const struct machine_desc *mdesc)
arm_dma_limit = PHYS_OFFSET + arm_dma_zone_size - 1; arm_dma_limit = PHYS_OFFSET + arm_dma_zone_size - 1;
} else } else
arm_dma_limit = 0xffffffff; arm_dma_limit = 0xffffffff;
arm_dma_pfn_limit = arm_dma_limit >> PAGE_SHIFT;
#endif #endif
} }
......
...@@ -81,8 +81,10 @@ extern __init void add_static_vm_early(struct static_vm *svm); ...@@ -81,8 +81,10 @@ extern __init void add_static_vm_early(struct static_vm *svm);
#ifdef CONFIG_ZONE_DMA #ifdef CONFIG_ZONE_DMA
extern phys_addr_t arm_dma_limit; extern phys_addr_t arm_dma_limit;
extern unsigned long arm_dma_pfn_limit;
#else #else
#define arm_dma_limit ((phys_addr_t)~0) #define arm_dma_limit ((phys_addr_t)~0)
#define arm_dma_pfn_limit (~0ul >> PAGE_SHIFT)
#endif #endif
extern phys_addr_t arm_lowmem_limit; extern phys_addr_t arm_lowmem_limit;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment