swiotlb.c 25.7 KB
Newer Older
Linus Torvalds's avatar
Linus Torvalds committed
1 2 3
/*
 * Dynamic DMA mapping support.
 *
Jan Beulich's avatar
Jan Beulich committed
4
 * This implementation is a fallback for platforms that do not support
Linus Torvalds's avatar
Linus Torvalds committed
5 6 7 8 9 10 11 12 13
 * I/O TLBs (aka DMA address translation hardware).
 * Copyright (C) 2000 Asit Mallick <Asit.K.Mallick@intel.com>
 * Copyright (C) 2000 Goutham Rao <goutham.rao@intel.com>
 * Copyright (C) 2000, 2003 Hewlett-Packard Co
 *	David Mosberger-Tang <davidm@hpl.hp.com>
 *
 * 03/05/07 davidm	Switch from PCI-DMA to generic device DMA API.
 * 00/12/13 davidm	Rename to swiotlb.c and add mark_clean() to avoid
 *			unnecessary i-cache flushing.
14 15 16
 * 04/07/.. ak		Better overflow handling. Assorted fixes.
 * 05/09/10 linville	Add support for syncing ranges, support syncing for
 *			DMA_BIDIRECTIONAL mappings, miscellaneous cleanup.
17
 * 08/12/11 beckyb	Add highmem support
Linus Torvalds's avatar
Linus Torvalds committed
18 19 20
 */

#include <linux/cache.h>
21
#include <linux/dma-mapping.h>
Linus Torvalds's avatar
Linus Torvalds committed
22 23 24 25
#include <linux/mm.h>
#include <linux/module.h>
#include <linux/spinlock.h>
#include <linux/string.h>
26
#include <linux/swiotlb.h>
27
#include <linux/pfn.h>
Linus Torvalds's avatar
Linus Torvalds committed
28 29
#include <linux/types.h>
#include <linux/ctype.h>
30
#include <linux/highmem.h>
31
#include <linux/gfp.h>
Linus Torvalds's avatar
Linus Torvalds committed
32 33 34

#include <asm/io.h>
#include <asm/dma.h>
35
#include <asm/scatterlist.h>
Linus Torvalds's avatar
Linus Torvalds committed
36 37 38

#include <linux/init.h>
#include <linux/bootmem.h>
39
#include <linux/iommu-helper.h>
Linus Torvalds's avatar
Linus Torvalds committed
40 41 42 43

#define OFFSET(val,align) ((unsigned long)	\
	                   ( (val) & ( (align) - 1)))

44 45 46 47 48 49 50 51 52
#define SLABS_PER_PAGE (1 << (PAGE_SHIFT - IO_TLB_SHIFT))

/*
 * Minimum IO TLB size to bother booting with.  Systems with mainly
 * 64bit capable cards will only lightly use the swiotlb.  If we can't
 * allocate a contiguous 1MB, we're probably in trouble anyway.
 */
#define IO_TLB_MIN_SLABS ((1<<20) >> IO_TLB_SHIFT)

Linus Torvalds's avatar
Linus Torvalds committed
53 54 55
int swiotlb_force;

/*
56 57
 * Used to do a quick range check in swiotlb_tbl_unmap_single and
 * swiotlb_tbl_sync_single_*, to see if the memory was in fact allocated by this
Linus Torvalds's avatar
Linus Torvalds committed
58 59 60 61 62
 * API.
 */
static char *io_tlb_start, *io_tlb_end;

/*
63
 * The number of IO TLB blocks (in groups of 64) between io_tlb_start and
Linus Torvalds's avatar
Linus Torvalds committed
64 65 66 67 68 69 70 71 72
 * io_tlb_end.  This is command line adjustable via setup_io_tlb_npages.
 */
static unsigned long io_tlb_nslabs;

/*
 * When the IOMMU overflows we return a fallback buffer. This sets the size.
 */
static unsigned long io_tlb_overflow = 32*1024;

73
static void *io_tlb_overflow_buffer;
Linus Torvalds's avatar
Linus Torvalds committed
74 75 76 77 78 79 80 81 82 83 84 85

/*
 * This is a free list describing the number of free entries available from
 * each index
 */
static unsigned int *io_tlb_list;
static unsigned int io_tlb_index;

/*
 * We need to save away the original address corresponding to a mapped entry
 * for the sync operations.
 */
86
static phys_addr_t *io_tlb_orig_addr;
Linus Torvalds's avatar
Linus Torvalds committed
87 88 89 90 91 92

/*
 * Protect the above data structures in the map and unmap calls
 */
static DEFINE_SPINLOCK(io_tlb_lock);

93 94
static int late_alloc;

Linus Torvalds's avatar
Linus Torvalds committed
95 96 97 98
static int __init
setup_io_tlb_npages(char *str)
{
	if (isdigit(*str)) {
99
		io_tlb_nslabs = simple_strtoul(str, &str, 0);
Linus Torvalds's avatar
Linus Torvalds committed
100 101 102 103 104
		/* avoid tail segment of size < IO_TLB_SEGSIZE */
		io_tlb_nslabs = ALIGN(io_tlb_nslabs, IO_TLB_SEGSIZE);
	}
	if (*str == ',')
		++str;
105
	if (!strcmp(str, "force"))
Linus Torvalds's avatar
Linus Torvalds committed
106
		swiotlb_force = 1;
107

Linus Torvalds's avatar
Linus Torvalds committed
108 109 110 111 112
	return 1;
}
__setup("swiotlb=", setup_io_tlb_npages);
/* make io_tlb_overflow tunable too? */

113
unsigned long swiotlb_nr_tbl(void)
114 115 116
{
	return io_tlb_nslabs;
}
117
EXPORT_SYMBOL_GPL(swiotlb_nr_tbl);
118
/* Note that this doesn't work with highmem page */
119 120
static dma_addr_t swiotlb_virt_to_bus(struct device *hwdev,
				      volatile void *address)
121
{
122
	return phys_to_dma(hwdev, virt_to_phys(address));
123 124
}

125
void swiotlb_print_info(void)
126
{
127
	unsigned long bytes = io_tlb_nslabs << IO_TLB_SHIFT;
128 129 130 131 132 133 134
	phys_addr_t pstart, pend;

	pstart = virt_to_phys(io_tlb_start);
	pend = virt_to_phys(io_tlb_end);

	printk(KERN_INFO "Placing %luMB software IO TLB between %p - %p\n",
	       bytes >> 20, io_tlb_start, io_tlb_end);
135 136 137
	printk(KERN_INFO "software IO TLB at phys %#llx - %#llx\n",
	       (unsigned long long)pstart,
	       (unsigned long long)pend);
138 139
}

140
void __init swiotlb_init_with_tbl(char *tlb, unsigned long nslabs, int verbose)
Linus Torvalds's avatar
Linus Torvalds committed
141
{
Jan Beulich's avatar
Jan Beulich committed
142
	unsigned long i, bytes;
Linus Torvalds's avatar
Linus Torvalds committed
143

144
	bytes = nslabs << IO_TLB_SHIFT;
Linus Torvalds's avatar
Linus Torvalds committed
145

146 147
	io_tlb_nslabs = nslabs;
	io_tlb_start = tlb;
Jan Beulich's avatar
Jan Beulich committed
148
	io_tlb_end = io_tlb_start + bytes;
Linus Torvalds's avatar
Linus Torvalds committed
149 150 151 152 153 154

	/*
	 * Allocate and initialize the free list array.  This array is used
	 * to find contiguous free memory regions of size up to IO_TLB_SEGSIZE
	 * between io_tlb_start and io_tlb_end.
	 */
155
	io_tlb_list = alloc_bootmem_pages(PAGE_ALIGN(io_tlb_nslabs * sizeof(int)));
156
	for (i = 0; i < io_tlb_nslabs; i++)
Linus Torvalds's avatar
Linus Torvalds committed
157 158
 		io_tlb_list[i] = IO_TLB_SEGSIZE - OFFSET(i, IO_TLB_SEGSIZE);
	io_tlb_index = 0;
159
	io_tlb_orig_addr = alloc_bootmem_pages(PAGE_ALIGN(io_tlb_nslabs * sizeof(phys_addr_t)));
Linus Torvalds's avatar
Linus Torvalds committed
160 161 162 163

	/*
	 * Get the overflow emergency buffer
	 */
164
	io_tlb_overflow_buffer = alloc_bootmem_low_pages(PAGE_ALIGN(io_tlb_overflow));
Jan Beulich's avatar
Jan Beulich committed
165 166
	if (!io_tlb_overflow_buffer)
		panic("Cannot allocate SWIOTLB overflow buffer!\n");
167 168
	if (verbose)
		swiotlb_print_info();
Linus Torvalds's avatar
Linus Torvalds committed
169 170
}

171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189
/*
 * Statically reserve bounce buffer space and initialize bounce buffer data
 * structures for the software IO TLB used to implement the DMA API.
 */
void __init
swiotlb_init_with_default_size(size_t default_size, int verbose)
{
	unsigned long bytes;

	if (!io_tlb_nslabs) {
		io_tlb_nslabs = (default_size >> IO_TLB_SHIFT);
		io_tlb_nslabs = ALIGN(io_tlb_nslabs, IO_TLB_SEGSIZE);
	}

	bytes = io_tlb_nslabs << IO_TLB_SHIFT;

	/*
	 * Get IO TLB memory from the low pages
	 */
190
	io_tlb_start = alloc_bootmem_low_pages(PAGE_ALIGN(bytes));
191 192 193 194 195 196
	if (!io_tlb_start)
		panic("Cannot allocate SWIOTLB buffer");

	swiotlb_init_with_tbl(io_tlb_start, io_tlb_nslabs, verbose);
}

Jan Beulich's avatar
Jan Beulich committed
197
void __init
198
swiotlb_init(int verbose)
Linus Torvalds's avatar
Linus Torvalds committed
199
{
200
	swiotlb_init_with_default_size(64 * (1<<20), verbose);	/* default to 64MB */
Linus Torvalds's avatar
Linus Torvalds committed
201 202
}

203 204 205 206 207 208
/*
 * Systems with larger DMA zones (those that don't support ISA) can
 * initialize the swiotlb later using the slab allocator if needed.
 * This should be just like above, but with some error catching.
 */
int
Jan Beulich's avatar
Jan Beulich committed
209
swiotlb_late_init_with_default_size(size_t default_size)
210
{
Jan Beulich's avatar
Jan Beulich committed
211
	unsigned long i, bytes, req_nslabs = io_tlb_nslabs;
212 213 214 215 216 217 218 219 220 221
	unsigned int order;

	if (!io_tlb_nslabs) {
		io_tlb_nslabs = (default_size >> IO_TLB_SHIFT);
		io_tlb_nslabs = ALIGN(io_tlb_nslabs, IO_TLB_SEGSIZE);
	}

	/*
	 * Get IO TLB memory from the low pages
	 */
Jan Beulich's avatar
Jan Beulich committed
222
	order = get_order(io_tlb_nslabs << IO_TLB_SHIFT);
223
	io_tlb_nslabs = SLABS_PER_PAGE << order;
Jan Beulich's avatar
Jan Beulich committed
224
	bytes = io_tlb_nslabs << IO_TLB_SHIFT;
225 226

	while ((SLABS_PER_PAGE << order) > IO_TLB_MIN_SLABS) {
227 228
		io_tlb_start = (void *)__get_free_pages(GFP_DMA | __GFP_NOWARN,
							order);
229 230 231 232 233 234 235 236
		if (io_tlb_start)
			break;
		order--;
	}

	if (!io_tlb_start)
		goto cleanup1;

Jan Beulich's avatar
Jan Beulich committed
237
	if (order != get_order(bytes)) {
238 239 240
		printk(KERN_WARNING "Warning: only able to allocate %ld MB "
		       "for software IO TLB\n", (PAGE_SIZE << order) >> 20);
		io_tlb_nslabs = SLABS_PER_PAGE << order;
Jan Beulich's avatar
Jan Beulich committed
241
		bytes = io_tlb_nslabs << IO_TLB_SHIFT;
242
	}
Jan Beulich's avatar
Jan Beulich committed
243 244
	io_tlb_end = io_tlb_start + bytes;
	memset(io_tlb_start, 0, bytes);
245 246 247 248 249 250 251 252 253 254 255 256 257 258 259

	/*
	 * Allocate and initialize the free list array.  This array is used
	 * to find contiguous free memory regions of size up to IO_TLB_SEGSIZE
	 * between io_tlb_start and io_tlb_end.
	 */
	io_tlb_list = (unsigned int *)__get_free_pages(GFP_KERNEL,
	                              get_order(io_tlb_nslabs * sizeof(int)));
	if (!io_tlb_list)
		goto cleanup2;

	for (i = 0; i < io_tlb_nslabs; i++)
 		io_tlb_list[i] = IO_TLB_SEGSIZE - OFFSET(i, IO_TLB_SEGSIZE);
	io_tlb_index = 0;

260 261 262 263
	io_tlb_orig_addr = (phys_addr_t *)
		__get_free_pages(GFP_KERNEL,
				 get_order(io_tlb_nslabs *
					   sizeof(phys_addr_t)));
264 265 266
	if (!io_tlb_orig_addr)
		goto cleanup3;

267
	memset(io_tlb_orig_addr, 0, io_tlb_nslabs * sizeof(phys_addr_t));
268 269 270 271 272 273 274 275 276

	/*
	 * Get the overflow emergency buffer
	 */
	io_tlb_overflow_buffer = (void *)__get_free_pages(GFP_DMA,
	                                          get_order(io_tlb_overflow));
	if (!io_tlb_overflow_buffer)
		goto cleanup4;

277
	swiotlb_print_info();
278

279 280
	late_alloc = 1;

281 282 283
	return 0;

cleanup4:
284 285
	free_pages((unsigned long)io_tlb_orig_addr,
		   get_order(io_tlb_nslabs * sizeof(phys_addr_t)));
286 287
	io_tlb_orig_addr = NULL;
cleanup3:
288 289
	free_pages((unsigned long)io_tlb_list, get_order(io_tlb_nslabs *
	                                                 sizeof(int)));
290 291
	io_tlb_list = NULL;
cleanup2:
Jan Beulich's avatar
Jan Beulich committed
292
	io_tlb_end = NULL;
293 294 295 296 297 298 299
	free_pages((unsigned long)io_tlb_start, order);
	io_tlb_start = NULL;
cleanup1:
	io_tlb_nslabs = req_nslabs;
	return -ENOMEM;
}

300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315
void __init swiotlb_free(void)
{
	if (!io_tlb_overflow_buffer)
		return;

	if (late_alloc) {
		free_pages((unsigned long)io_tlb_overflow_buffer,
			   get_order(io_tlb_overflow));
		free_pages((unsigned long)io_tlb_orig_addr,
			   get_order(io_tlb_nslabs * sizeof(phys_addr_t)));
		free_pages((unsigned long)io_tlb_list, get_order(io_tlb_nslabs *
								 sizeof(int)));
		free_pages((unsigned long)io_tlb_start,
			   get_order(io_tlb_nslabs << IO_TLB_SHIFT));
	} else {
		free_bootmem_late(__pa(io_tlb_overflow_buffer),
316
				  PAGE_ALIGN(io_tlb_overflow));
317
		free_bootmem_late(__pa(io_tlb_orig_addr),
318
				  PAGE_ALIGN(io_tlb_nslabs * sizeof(phys_addr_t)));
319
		free_bootmem_late(__pa(io_tlb_list),
320
				  PAGE_ALIGN(io_tlb_nslabs * sizeof(int)));
321
		free_bootmem_late(__pa(io_tlb_start),
322
				  PAGE_ALIGN(io_tlb_nslabs << IO_TLB_SHIFT));
323
	}
324
	io_tlb_nslabs = 0;
325 326
}

327
static int is_swiotlb_buffer(phys_addr_t paddr)
328
{
329 330
	return paddr >= virt_to_phys(io_tlb_start) &&
		paddr < virt_to_phys(io_tlb_end);
331 332
}

333 334 335
/*
 * Bounce: copy the swiotlb buffer back to the original dma location
 */
336 337
void swiotlb_bounce(phys_addr_t phys, char *dma_addr, size_t size,
		    enum dma_data_direction dir)
338 339 340 341 342 343 344 345 346 347 348
{
	unsigned long pfn = PFN_DOWN(phys);

	if (PageHighMem(pfn_to_page(pfn))) {
		/* The buffer does not have a mapping.  Map it in and copy */
		unsigned int offset = phys & ~PAGE_MASK;
		char *buffer;
		unsigned int sz = 0;
		unsigned long flags;

		while (size) {
Becky Bruce's avatar
Becky Bruce committed
349
			sz = min_t(size_t, PAGE_SIZE - offset, size);
350 351 352 353 354 355

			local_irq_save(flags);
			buffer = kmap_atomic(pfn_to_page(pfn),
					     KM_BOUNCE_READ);
			if (dir == DMA_TO_DEVICE)
				memcpy(dma_addr, buffer + offset, sz);
356
			else
357 358
				memcpy(buffer + offset, dma_addr, sz);
			kunmap_atomic(buffer, KM_BOUNCE_READ);
359
			local_irq_restore(flags);
360 361 362 363 364

			size -= sz;
			pfn++;
			dma_addr += sz;
			offset = 0;
365 366 367
		}
	} else {
		if (dir == DMA_TO_DEVICE)
368
			memcpy(dma_addr, phys_to_virt(phys), size);
369
		else
370
			memcpy(phys_to_virt(phys), dma_addr, size);
371
	}
372
}
373
EXPORT_SYMBOL_GPL(swiotlb_bounce);
374

375
void *swiotlb_tbl_map_single(struct device *hwdev, dma_addr_t tbl_dma_addr,
376 377
			     phys_addr_t phys, size_t size,
			     enum dma_data_direction dir)
Linus Torvalds's avatar
Linus Torvalds committed
378 379 380 381 382
{
	unsigned long flags;
	char *dma_addr;
	unsigned int nslots, stride, index, wrap;
	int i;
383 384 385 386 387 388
	unsigned long mask;
	unsigned long offset_slots;
	unsigned long max_slots;

	mask = dma_get_seg_boundary(hwdev);

389 390 391
	tbl_dma_addr &= mask;

	offset_slots = ALIGN(tbl_dma_addr, 1 << IO_TLB_SHIFT) >> IO_TLB_SHIFT;
392 393 394 395

	/*
 	 * Carefully handle integer overflow which can occur when mask == ~0UL.
 	 */
396 397 398
	max_slots = mask + 1
		    ? ALIGN(mask + 1, 1 << IO_TLB_SHIFT) >> IO_TLB_SHIFT
		    : 1UL << (BITS_PER_LONG - IO_TLB_SHIFT);
Linus Torvalds's avatar
Linus Torvalds committed
399 400 401 402 403 404 405 406 407 408 409

	/*
	 * For mappings greater than a page, we limit the stride (and
	 * hence alignment) to a page size.
	 */
	nslots = ALIGN(size, 1 << IO_TLB_SHIFT) >> IO_TLB_SHIFT;
	if (size > PAGE_SIZE)
		stride = (1 << (PAGE_SHIFT - IO_TLB_SHIFT));
	else
		stride = 1;

410
	BUG_ON(!nslots);
Linus Torvalds's avatar
Linus Torvalds committed
411 412 413 414 415 416

	/*
	 * Find suitable number of IO TLB entries size that will fit this
	 * request and allocate a buffer from that IO TLB pool.
	 */
	spin_lock_irqsave(&io_tlb_lock, flags);
Andrew Morton's avatar
Andrew Morton committed
417 418 419 420 421 422
	index = ALIGN(io_tlb_index, stride);
	if (index >= io_tlb_nslabs)
		index = 0;
	wrap = index;

	do {
423 424
		while (iommu_is_span_boundary(index, nslots, offset_slots,
					      max_slots)) {
425 426 427
			index += stride;
			if (index >= io_tlb_nslabs)
				index = 0;
Andrew Morton's avatar
Andrew Morton committed
428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444
			if (index == wrap)
				goto not_found;
		}

		/*
		 * If we find a slot that indicates we have 'nslots' number of
		 * contiguous buffers, we allocate the buffers from that slot
		 * and mark the entries as '0' indicating unavailable.
		 */
		if (io_tlb_list[index] >= nslots) {
			int count = 0;

			for (i = index; i < (int) (index + nslots); i++)
				io_tlb_list[i] = 0;
			for (i = index - 1; (OFFSET(i, IO_TLB_SEGSIZE) != IO_TLB_SEGSIZE - 1) && io_tlb_list[i]; i--)
				io_tlb_list[i] = ++count;
			dma_addr = io_tlb_start + (index << IO_TLB_SHIFT);
Linus Torvalds's avatar
Linus Torvalds committed
445

Andrew Morton's avatar
Andrew Morton committed
446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463
			/*
			 * Update the indices to avoid searching in the next
			 * round.
			 */
			io_tlb_index = ((index + nslots) < io_tlb_nslabs
					? (index + nslots) : 0);

			goto found;
		}
		index += stride;
		if (index >= io_tlb_nslabs)
			index = 0;
	} while (index != wrap);

not_found:
	spin_unlock_irqrestore(&io_tlb_lock, flags);
	return NULL;
found:
Linus Torvalds's avatar
Linus Torvalds committed
464 465 466 467 468 469 470
	spin_unlock_irqrestore(&io_tlb_lock, flags);

	/*
	 * Save away the mapping from the original address to the DMA address.
	 * This is needed when we sync the memory.  Then we sync the buffer if
	 * needed.
	 */
471 472
	for (i = 0; i < nslots; i++)
		io_tlb_orig_addr[index+i] = phys + (i << IO_TLB_SHIFT);
Linus Torvalds's avatar
Linus Torvalds committed
473
	if (dir == DMA_TO_DEVICE || dir == DMA_BIDIRECTIONAL)
474
		swiotlb_bounce(phys, dma_addr, size, DMA_TO_DEVICE);
Linus Torvalds's avatar
Linus Torvalds committed
475 476 477

	return dma_addr;
}
478
EXPORT_SYMBOL_GPL(swiotlb_tbl_map_single);
Linus Torvalds's avatar
Linus Torvalds committed
479

480 481 482 483 484
/*
 * Allocates bounce buffer and returns its kernel virtual address.
 */

static void *
485 486
map_single(struct device *hwdev, phys_addr_t phys, size_t size,
	   enum dma_data_direction dir)
487 488 489 490 491 492
{
	dma_addr_t start_dma_addr = swiotlb_virt_to_bus(hwdev, io_tlb_start);

	return swiotlb_tbl_map_single(hwdev, start_dma_addr, phys, size, dir);
}

Linus Torvalds's avatar
Linus Torvalds committed
493 494 495
/*
 * dma_addr is the kernel virtual address of the bounce buffer to unmap.
 */
496
void
497
swiotlb_tbl_unmap_single(struct device *hwdev, char *dma_addr, size_t size,
498
			enum dma_data_direction dir)
Linus Torvalds's avatar
Linus Torvalds committed
499 500 501 502
{
	unsigned long flags;
	int i, count, nslots = ALIGN(size, 1 << IO_TLB_SHIFT) >> IO_TLB_SHIFT;
	int index = (dma_addr - io_tlb_start) >> IO_TLB_SHIFT;
503
	phys_addr_t phys = io_tlb_orig_addr[index];
Linus Torvalds's avatar
Linus Torvalds committed
504 505 506 507

	/*
	 * First, sync the memory before unmapping the entry
	 */
508
	if (phys && ((dir == DMA_FROM_DEVICE) || (dir == DMA_BIDIRECTIONAL)))
509
		swiotlb_bounce(phys, dma_addr, size, DMA_FROM_DEVICE);
Linus Torvalds's avatar
Linus Torvalds committed
510 511 512

	/*
	 * Return the buffer to the free list by setting the corresponding
513
	 * entries to indicate the number of contiguous entries available.
Linus Torvalds's avatar
Linus Torvalds committed
514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535
	 * While returning the entries to the free list, we merge the entries
	 * with slots below and above the pool being returned.
	 */
	spin_lock_irqsave(&io_tlb_lock, flags);
	{
		count = ((index + nslots) < ALIGN(index + 1, IO_TLB_SEGSIZE) ?
			 io_tlb_list[index + nslots] : 0);
		/*
		 * Step 1: return the slots to the free list, merging the
		 * slots with superceeding slots
		 */
		for (i = index + nslots - 1; i >= index; i--)
			io_tlb_list[i] = ++count;
		/*
		 * Step 2: merge the returned slots with the preceding slots,
		 * if available (non zero)
		 */
		for (i = index - 1; (OFFSET(i, IO_TLB_SEGSIZE) != IO_TLB_SEGSIZE -1) && io_tlb_list[i]; i--)
			io_tlb_list[i] = ++count;
	}
	spin_unlock_irqrestore(&io_tlb_lock, flags);
}
536
EXPORT_SYMBOL_GPL(swiotlb_tbl_unmap_single);
Linus Torvalds's avatar
Linus Torvalds committed
537

538
void
539
swiotlb_tbl_sync_single(struct device *hwdev, char *dma_addr, size_t size,
540 541
			enum dma_data_direction dir,
			enum dma_sync_target target)
Linus Torvalds's avatar
Linus Torvalds committed
542
{
543 544 545 546
	int index = (dma_addr - io_tlb_start) >> IO_TLB_SHIFT;
	phys_addr_t phys = io_tlb_orig_addr[index];

	phys += ((unsigned long)dma_addr & ((1 << IO_TLB_SHIFT) - 1));
547

548 549 550
	switch (target) {
	case SYNC_FOR_CPU:
		if (likely(dir == DMA_FROM_DEVICE || dir == DMA_BIDIRECTIONAL))
551
			swiotlb_bounce(phys, dma_addr, size, DMA_FROM_DEVICE);
552 553
		else
			BUG_ON(dir != DMA_TO_DEVICE);
554 555 556
		break;
	case SYNC_FOR_DEVICE:
		if (likely(dir == DMA_TO_DEVICE || dir == DMA_BIDIRECTIONAL))
557
			swiotlb_bounce(phys, dma_addr, size, DMA_TO_DEVICE);
558 559
		else
			BUG_ON(dir != DMA_FROM_DEVICE);
560 561
		break;
	default:
Linus Torvalds's avatar
Linus Torvalds committed
562
		BUG();
563
	}
Linus Torvalds's avatar
Linus Torvalds committed
564
}
565
EXPORT_SYMBOL_GPL(swiotlb_tbl_sync_single);
Linus Torvalds's avatar
Linus Torvalds committed
566 567 568

void *
swiotlb_alloc_coherent(struct device *hwdev, size_t size,
Al Viro's avatar
Al Viro committed
569
		       dma_addr_t *dma_handle, gfp_t flags)
Linus Torvalds's avatar
Linus Torvalds committed
570
{
Jan Beulich's avatar
Jan Beulich committed
571
	dma_addr_t dev_addr;
Linus Torvalds's avatar
Linus Torvalds committed
572 573
	void *ret;
	int order = get_order(size);
574
	u64 dma_mask = DMA_BIT_MASK(32);
575 576 577

	if (hwdev && hwdev->coherent_dma_mask)
		dma_mask = hwdev->coherent_dma_mask;
Linus Torvalds's avatar
Linus Torvalds committed
578

579
	ret = (void *)__get_free_pages(flags, order);
580
	if (ret && swiotlb_virt_to_bus(hwdev, ret) + size - 1 > dma_mask) {
Linus Torvalds's avatar
Linus Torvalds committed
581 582 583 584 585 586 587 588
		/*
		 * The allocated memory isn't reachable by the device.
		 */
		free_pages((unsigned long) ret, order);
		ret = NULL;
	}
	if (!ret) {
		/*
589 590
		 * We are either out of memory or the device can't DMA to
		 * GFP_DMA memory; fall back on map_single(), which
Becky Bruce's avatar
Becky Bruce committed
591
		 * will grab memory from the lowest available address range.
Linus Torvalds's avatar
Linus Torvalds committed
592
		 */
593
		ret = map_single(hwdev, 0, size, DMA_FROM_DEVICE);
594
		if (!ret)
Linus Torvalds's avatar
Linus Torvalds committed
595 596 597 598
			return NULL;
	}

	memset(ret, 0, size);
599
	dev_addr = swiotlb_virt_to_bus(hwdev, ret);
Linus Torvalds's avatar
Linus Torvalds committed
600 601

	/* Confirm address can be DMA'd by device */
602
	if (dev_addr + size - 1 > dma_mask) {
Jan Beulich's avatar
Jan Beulich committed
603
		printk("hwdev DMA mask = 0x%016Lx, dev_addr = 0x%016Lx\n",
604
		       (unsigned long long)dma_mask,
Jan Beulich's avatar
Jan Beulich committed
605
		       (unsigned long long)dev_addr);
606 607

		/* DMA_TO_DEVICE to avoid memcpy in unmap_single */
608
		swiotlb_tbl_unmap_single(hwdev, ret, size, DMA_TO_DEVICE);
609
		return NULL;
Linus Torvalds's avatar
Linus Torvalds committed
610 611 612 613
	}
	*dma_handle = dev_addr;
	return ret;
}
614
EXPORT_SYMBOL(swiotlb_alloc_coherent);
Linus Torvalds's avatar
Linus Torvalds committed
615 616 617

void
swiotlb_free_coherent(struct device *hwdev, size_t size, void *vaddr,
618
		      dma_addr_t dev_addr)
Linus Torvalds's avatar
Linus Torvalds committed
619
{
620
	phys_addr_t paddr = dma_to_phys(hwdev, dev_addr);
621

622
	WARN_ON(irqs_disabled());
623 624
	if (!is_swiotlb_buffer(paddr))
		free_pages((unsigned long)vaddr, get_order(size));
Linus Torvalds's avatar
Linus Torvalds committed
625
	else
626 627
		/* DMA_TO_DEVICE to avoid memcpy in swiotlb_tbl_unmap_single */
		swiotlb_tbl_unmap_single(hwdev, vaddr, size, DMA_TO_DEVICE);
Linus Torvalds's avatar
Linus Torvalds committed
628
}
629
EXPORT_SYMBOL(swiotlb_free_coherent);
Linus Torvalds's avatar
Linus Torvalds committed
630 631

static void
632 633
swiotlb_full(struct device *dev, size_t size, enum dma_data_direction dir,
	     int do_panic)
Linus Torvalds's avatar
Linus Torvalds committed
634 635 636 637
{
	/*
	 * Ran out of IOMMU space for this operation. This is very bad.
	 * Unfortunately the drivers cannot handle this operation properly.
638
	 * unless they check for dma_mapping_error (most don't)
Linus Torvalds's avatar
Linus Torvalds committed
639 640 641
	 * When the mapping is small enough return a static buffer to limit
	 * the damage, or panic when the transfer is too big.
	 */
Jan Beulich's avatar
Jan Beulich committed
642
	printk(KERN_ERR "DMA: Out of SW-IOMMU space for %zu bytes at "
643
	       "device %s\n", size, dev ? dev_name(dev) : "?");
Linus Torvalds's avatar
Linus Torvalds committed
644

645 646 647 648 649 650 651 652 653
	if (size <= io_tlb_overflow || !do_panic)
		return;

	if (dir == DMA_BIDIRECTIONAL)
		panic("DMA: Random memory could be DMA accessed\n");
	if (dir == DMA_FROM_DEVICE)
		panic("DMA: Random memory could be DMA written\n");
	if (dir == DMA_TO_DEVICE)
		panic("DMA: Random memory could be DMA read\n");
Linus Torvalds's avatar
Linus Torvalds committed
654 655 656 657
}

/*
 * Map a single buffer of the indicated size for DMA in streaming mode.  The
658
 * physical address to use is returned.
Linus Torvalds's avatar
Linus Torvalds committed
659 660
 *
 * Once the device is given the dma address, the device owns this memory until
Becky Bruce's avatar
Becky Bruce committed
661
 * either swiotlb_unmap_page or swiotlb_dma_sync_single is performed.
Linus Torvalds's avatar
Linus Torvalds committed
662
 */
663 664 665 666
dma_addr_t swiotlb_map_page(struct device *dev, struct page *page,
			    unsigned long offset, size_t size,
			    enum dma_data_direction dir,
			    struct dma_attrs *attrs)
Linus Torvalds's avatar
Linus Torvalds committed
667
{
668
	phys_addr_t phys = page_to_phys(page) + offset;
669
	dma_addr_t dev_addr = phys_to_dma(dev, phys);
Linus Torvalds's avatar
Linus Torvalds committed
670 671
	void *map;

672
	BUG_ON(dir == DMA_NONE);
Linus Torvalds's avatar
Linus Torvalds committed
673
	/*
Becky Bruce's avatar
Becky Bruce committed
674
	 * If the address happens to be in the device's DMA window,
Linus Torvalds's avatar
Linus Torvalds committed
675 676 677
	 * we can safely return the device addr and not worry about bounce
	 * buffering it.
	 */
FUJITA Tomonori's avatar
FUJITA Tomonori committed
678
	if (dma_capable(dev, dev_addr, size) && !swiotlb_force)
Linus Torvalds's avatar
Linus Torvalds committed
679 680 681 682 683
		return dev_addr;

	/*
	 * Oh well, have to allocate and map a bounce buffer.
	 */
684
	map = map_single(dev, phys, size, dir);
Linus Torvalds's avatar
Linus Torvalds committed
685
	if (!map) {
686
		swiotlb_full(dev, size, dir, 1);
Linus Torvalds's avatar
Linus Torvalds committed
687 688 689
		map = io_tlb_overflow_buffer;
	}

690
	dev_addr = swiotlb_virt_to_bus(dev, map);
Linus Torvalds's avatar
Linus Torvalds committed
691 692 693 694

	/*
	 * Ensure that the address returned is DMA'ble
	 */
FUJITA Tomonori's avatar
FUJITA Tomonori committed
695 696 697 698
	if (!dma_capable(dev, dev_addr, size)) {
		swiotlb_tbl_unmap_single(dev, map, size, dir);
		dev_addr = swiotlb_virt_to_bus(dev, io_tlb_overflow_buffer);
	}
Linus Torvalds's avatar
Linus Torvalds committed
699 700 701

	return dev_addr;
}
702
EXPORT_SYMBOL_GPL(swiotlb_map_page);
Linus Torvalds's avatar
Linus Torvalds committed
703 704 705

/*
 * Unmap a single streaming mode DMA translation.  The dma_addr and size must
Becky Bruce's avatar
Becky Bruce committed
706
 * match what was provided for in a previous swiotlb_map_page call.  All
Linus Torvalds's avatar
Linus Torvalds committed
707 708 709 710 711
 * other usages are undefined.
 *
 * After this call, reads by the cpu to the buffer are guaranteed to see
 * whatever the device wrote there.
 */
712
static void unmap_single(struct device *hwdev, dma_addr_t dev_addr,
713
			 size_t size, enum dma_data_direction dir)
Linus Torvalds's avatar
Linus Torvalds committed
714
{
715
	phys_addr_t paddr = dma_to_phys(hwdev, dev_addr);
Linus Torvalds's avatar
Linus Torvalds committed
716

717
	BUG_ON(dir == DMA_NONE);
718

719
	if (is_swiotlb_buffer(paddr)) {
720
		swiotlb_tbl_unmap_single(hwdev, phys_to_virt(paddr), size, dir);
721 722 723 724 725 726
		return;
	}

	if (dir != DMA_FROM_DEVICE)
		return;

727 728 729 730 731 732 733
	/*
	 * phys_to_virt doesn't work with hihgmem page but we could
	 * call dma_mark_clean() with hihgmem page here. However, we
	 * are fine since dma_mark_clean() is null on POWERPC. We can
	 * make dma_mark_clean() take a physical address if necessary.
	 */
	dma_mark_clean(phys_to_virt(paddr), size);
734 735 736 737 738 739 740
}

void swiotlb_unmap_page(struct device *hwdev, dma_addr_t dev_addr,
			size_t size, enum dma_data_direction dir,
			struct dma_attrs *attrs)
{
	unmap_single(hwdev, dev_addr, size, dir);
Linus Torvalds's avatar
Linus Torvalds committed
741
}
742
EXPORT_SYMBOL_GPL(swiotlb_unmap_page);
743

Linus Torvalds's avatar
Linus Torvalds committed
744 745 746 747
/*
 * Make physical memory consistent for a single streaming mode DMA translation
 * after a transfer.
 *
Becky Bruce's avatar
Becky Bruce committed
748
 * If you perform a swiotlb_map_page() but wish to interrogate the buffer
749 750
 * using the cpu, yet do not wish to teardown the dma mapping, you must
 * call this function before doing so.  At the next point you give the dma
Linus Torvalds's avatar
Linus Torvalds committed
751 752 753
 * address back to the card, you must first perform a
 * swiotlb_dma_sync_for_device, and then the device again owns the buffer
 */
Andrew Morton's avatar
Andrew Morton committed
754
static void
755
swiotlb_sync_single(struct device *hwdev, dma_addr_t dev_addr,
756 757
		    size_t size, enum dma_data_direction dir,
		    enum dma_sync_target target)
Linus Torvalds's avatar
Linus Torvalds committed
758
{
759
	phys_addr_t paddr = dma_to_phys(hwdev, dev_addr);
Linus Torvalds's avatar
Linus Torvalds committed
760

761
	BUG_ON(dir == DMA_NONE);
762

763
	if (is_swiotlb_buffer(paddr)) {
764 765
		swiotlb_tbl_sync_single(hwdev, phys_to_virt(paddr), size, dir,
				       target);
766 767 768 769 770 771
		return;
	}

	if (dir != DMA_FROM_DEVICE)
		return;

772
	dma_mark_clean(phys_to_virt(paddr), size);
Linus Torvalds's avatar
Linus Torvalds committed
773 774
}

775 776
void
swiotlb_sync_single_for_cpu(struct device *hwdev, dma_addr_t dev_addr,
777
			    size_t size, enum dma_data_direction dir)
778
{
779
	swiotlb_sync_single(hwdev, dev_addr, size, dir, SYNC_FOR_CPU);
780
}
781
EXPORT_SYMBOL(swiotlb_sync_single_for_cpu);
782

Linus Torvalds's avatar
Linus Torvalds committed
783 784
void
swiotlb_sync_single_for_device(struct device *hwdev, dma_addr_t dev_addr,
785
			       size_t size, enum dma_data_direction dir)
Linus Torvalds's avatar
Linus Torvalds committed
786
{
787
	swiotlb_sync_single(hwdev, dev_addr, size, dir, SYNC_FOR_DEVICE);
Linus Torvalds's avatar
Linus Torvalds committed
788
}
789
EXPORT_SYMBOL(swiotlb_sync_single_for_device);
Linus Torvalds's avatar
Linus Torvalds committed
790 791 792

/*
 * Map a set of buffers described by scatterlist in streaming mode for DMA.
Becky Bruce's avatar
Becky Bruce committed
793
 * This is the scatter-gather version of the above swiotlb_map_page
Linus Torvalds's avatar
Linus Torvalds committed
794 795 796 797 798 799 800 801 802 803
 * interface.  Here the scatter gather list elements are each tagged with the
 * appropriate dma address and length.  They are obtained via
 * sg_dma_{address,length}(SG).
 *
 * NOTE: An implementation may be able to use a smaller number of
 *       DMA address/length pairs than there are SG table elements.
 *       (for example via virtual mapping capabilities)
 *       The routine returns the number of addr/length pairs actually
 *       used, at most nents.
 *
Becky Bruce's avatar
Becky Bruce committed
804
 * Device ownership issues as mentioned above for swiotlb_map_page are the
Linus Torvalds's avatar
Linus Torvalds committed
805 806 807
 * same here.
 */
int
808
swiotlb_map_sg_attrs(struct device *hwdev, struct scatterlist *sgl, int nelems,
809
		     enum dma_data_direction dir, struct dma_attrs *attrs)
Linus Torvalds's avatar
Linus Torvalds committed
810
{
Jens Axboe's avatar
Jens Axboe committed
811
	struct scatterlist *sg;
Linus Torvalds's avatar
Linus Torvalds committed
812 813
	int i;

814
	BUG_ON(dir == DMA_NONE);
Linus Torvalds's avatar
Linus Torvalds committed
815

Jens Axboe's avatar
Jens Axboe committed
816
	for_each_sg(sgl, sg, nelems, i) {
Ian Campbell's avatar
Ian Campbell committed
817
		phys_addr_t paddr = sg_phys(sg);
818
		dma_addr_t dev_addr = phys_to_dma(hwdev, paddr);
819

820
		if (swiotlb_force ||
FUJITA Tomonori's avatar
FUJITA Tomonori committed
821
		    !dma_capable(hwdev, dev_addr, sg->length)) {
822 823
			void *map = map_single(hwdev, sg_phys(sg),
					       sg->length, dir);
824
			if (!map) {
Linus Torvalds's avatar
Linus Torvalds committed
825 826 827
				/* Don't panic here, we expect map_sg users
				   to do proper error handling. */
				swiotlb_full(hwdev, sg->length, dir, 0);
828 829
				swiotlb_unmap_sg_attrs(hwdev, sgl, i, dir,
						       attrs);
Jens Axboe's avatar
Jens Axboe committed
830
				sgl[0].dma_length = 0;
Linus Torvalds's avatar
Linus Torvalds committed
831 832
				return 0;
			}
833
			sg->dma_address = swiotlb_virt_to_bus(hwdev, map);
Linus Torvalds's avatar
Linus Torvalds committed
834 835 836 837 838 839
		} else
			sg->dma_address = dev_addr;
		sg->dma_length = sg->length;
	}
	return nelems;
}
840 841 842 843
EXPORT_SYMBOL(swiotlb_map_sg_attrs);

int
swiotlb_map_sg(struct device *hwdev, struct scatterlist *sgl, int nelems,
844
	       enum dma_data_direction dir)
845 846 847
{
	return swiotlb_map_sg_attrs(hwdev, sgl, nelems, dir, NULL);
}
848
EXPORT_SYMBOL(swiotlb_map_sg);
Linus Torvalds's avatar
Linus Torvalds committed
849 850 851

/*
 * Unmap a set of streaming mode DMA translations.  Again, cpu read rules
Becky Bruce's avatar
Becky Bruce committed
852
 * concerning calls here are the same as for swiotlb_unmap_page() above.
Linus Torvalds's avatar
Linus Torvalds committed
853 854
 */
void
855
swiotlb_unmap_sg_attrs(struct device *hwdev, struct scatterlist *sgl,
856
		       int nelems, enum dma_data_direction dir, struct dma_attrs *attrs)
Linus Torvalds's avatar
Linus Torvalds committed
857
{
Jens Axboe's avatar
Jens Axboe committed
858
	struct scatterlist *sg;
Linus Torvalds's avatar
Linus Torvalds committed
859 860
	int i;

861
	BUG_ON(dir == DMA_NONE);
Linus Torvalds's avatar
Linus Torvalds committed
862

863 864 865
	for_each_sg(sgl, sg, nelems, i)
		unmap_single(hwdev, sg->dma_address, sg->dma_length, dir);

Linus Torvalds's avatar
Linus Torvalds committed
866
}
867 868 869 870
EXPORT_SYMBOL(swiotlb_unmap_sg_attrs);

void
swiotlb_unmap_sg(struct device *hwdev, struct scatterlist *sgl, int nelems,
871
		 enum dma_data_direction dir)
872 873 874
{
	return swiotlb_unmap_sg_attrs(hwdev, sgl, nelems, dir, NULL);
}
875
EXPORT_SYMBOL(swiotlb_unmap_sg);
Linus Torvalds's avatar
Linus Torvalds committed
876 877 878 879 880 881 882 883

/*
 * Make physical memory consistent for a set of streaming mode DMA translations
 * after a transfer.
 *
 * The same as swiotlb_sync_single_* but for a scatter-gather list, same rules
 * and usage.
 */
Andrew Morton's avatar
Andrew Morton committed
884
static void
Jens Axboe's avatar
Jens Axboe committed
885
swiotlb_sync_sg(struct device *hwdev, struct scatterlist *sgl,
886 887
		int nelems, enum dma_data_direction dir,
		enum dma_sync_target target)
Linus Torvalds's avatar
Linus Torvalds committed
888
{
Jens Axboe's avatar
Jens Axboe committed
889
	struct scatterlist *sg;
Linus Torvalds's avatar
Linus Torvalds committed
890 891
	int i;

892 893
	for_each_sg(sgl, sg, nelems, i)
		swiotlb_sync_single(hwdev, sg->dma_address,
894
				    sg->dma_length, dir, target);
Linus Torvalds's avatar
Linus Torvalds committed
895 896
}

897 898
void
swiotlb_sync_sg_for_cpu(struct device *hwdev, struct scatterlist *sg,
899
			int nelems, enum dma_data_direction dir)
900
{
901
	swiotlb_sync_sg(hwdev, sg, nelems, dir, SYNC_FOR_CPU);
902
}
903
EXPORT_SYMBOL(swiotlb_sync_sg_for_cpu);
904

Linus Torvalds's avatar
Linus Torvalds committed
905 906
void
swiotlb_sync_sg_for_device(struct device *hwdev, struct scatterlist *sg,
907
			   int nelems, enum dma_data_direction dir)
Linus Torvalds's avatar
Linus Torvalds committed
908
{
909
	swiotlb_sync_sg(hwdev, sg, nelems, dir, SYNC_FOR_DEVICE);
Linus Torvalds's avatar
Linus Torvalds committed
910
}
911
EXPORT_SYMBOL(swiotlb_sync_sg_for_device);
Linus Torvalds's avatar
Linus Torvalds committed
912 913

int
914
swiotlb_dma_mapping_error(struct device *hwdev, dma_addr_t dma_addr)
Linus Torvalds's avatar
Linus Torvalds committed
915
{
916
	return (dma_addr == swiotlb_virt_to_bus(hwdev, io_tlb_overflow_buffer));
Linus Torvalds's avatar
Linus Torvalds committed
917
}
918
EXPORT_SYMBOL(swiotlb_dma_mapping_error);
Linus Torvalds's avatar
Linus Torvalds committed
919 920

/*
921
 * Return whether the given device DMA address mask can be supported
Linus Torvalds's avatar
Linus Torvalds committed
922
 * properly.  For example, if your device can only drive the low 24-bits
923
 * during bus mastering, then you would pass 0x00ffffff as the mask to
Linus Torvalds's avatar
Linus Torvalds committed
924 925 926
 * this function.
 */
int
Jan Beulich's avatar
Jan Beulich committed
927
swiotlb_dma_supported(struct device *hwdev, u64 mask)
Linus Torvalds's avatar
Linus Torvalds committed
928
{
929
	return swiotlb_virt_to_bus(hwdev, io_tlb_end - 1) <= mask;
Linus Torvalds's avatar
Linus Torvalds committed
930 931
}
EXPORT_SYMBOL(swiotlb_dma_supported);