Commit 1d990882 authored by FUJITA Tomonori's avatar FUJITA Tomonori Committed by Ingo Molnar

x86: restore old GART alloc_coherent behavior

Currently, GART alloc_coherent tries to allocate pages with GFP_DMA32
for a device having dma_masks > 24bit < 32bits. If GART gets an
address that a device can't access to, GART try to map the address to
a virtual I/O address that the device can access to.

But Andi pointed out, "The GART is somewhere in the 4GB range so you
cannot use it to map anything < 4GB. Also GART is pretty small."

http://lkml.org/lkml/2008/9/12/43

That is, it's possible that GART doesn't have virtual I/O address
space that a device can access to. The above behavior doesn't work for
a device having dma_masks > 24bit < 32bits.

This patch restores old GART alloc_coherent behavior (before the
alloc_coherent rewrite).
Signed-off-by: default avatarFUJITA Tomonori <fujita.tomonori@lab.ntt.co.jp>
Signed-off-by: default avatarIngo Molnar <mingo@elte.hu>
parent ecef533e
...@@ -487,31 +487,28 @@ static void * ...@@ -487,31 +487,28 @@ static void *
gart_alloc_coherent(struct device *dev, size_t size, dma_addr_t *dma_addr, gart_alloc_coherent(struct device *dev, size_t size, dma_addr_t *dma_addr,
gfp_t flag) gfp_t flag)
{ {
void *vaddr;
dma_addr_t paddr; dma_addr_t paddr;
unsigned long align_mask; unsigned long align_mask;
u64 dma_mask = dma_alloc_coherent_mask(dev, flag); struct page *page;
vaddr = (void *)__get_free_pages(flag | __GFP_ZERO, get_order(size)); if (force_iommu && !(flag & GFP_DMA)) {
if (!vaddr) flag &= ~(__GFP_DMA | __GFP_HIGHMEM | __GFP_DMA32);
return NULL; page = alloc_pages(flag | __GFP_ZERO, get_order(size));
if (!page)
paddr = virt_to_phys(vaddr); return NULL;
if (is_buffer_dma_capable(dma_mask, paddr, size)) {
*dma_addr = paddr; align_mask = (1UL << get_order(size)) - 1;
return vaddr; paddr = dma_map_area(dev, page_to_phys(page), size,
} DMA_BIDIRECTIONAL, align_mask);
align_mask = (1UL << get_order(size)) - 1; flush_gart();
if (paddr != bad_dma_address) {
*dma_addr = dma_map_area(dev, paddr, size, DMA_BIDIRECTIONAL, *dma_addr = paddr;
align_mask); return page_address(page);
flush_gart(); }
__free_pages(page, get_order(size));
if (*dma_addr != bad_dma_address) } else
return vaddr; return dma_generic_alloc_coherent(dev, size, dma_addr, flag);
free_pages((unsigned long)vaddr, get_order(size));
return NULL; return NULL;
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment