Commit 5ceda740 authored by Christoph Hellwig's avatar Christoph Hellwig

dma-direct: rename and cleanup __phys_to_dma

The __phys_to_dma vs phys_to_dma distinction isn't exactly obvious.  Try
to improve the situation by renaming __phys_to_dma to
phys_to_dma_unencryped, and not forcing architectures that want to
override phys_to_dma to actually provide __phys_to_dma.
Signed-off-by: default avatarChristoph Hellwig <hch@lst.de>
Reviewed-by: default avatarRobin Murphy <robin.murphy@arm.com>
parent 7bc5c428
...@@ -2,7 +2,7 @@ ...@@ -2,7 +2,7 @@
#ifndef ASM_ARM_DMA_DIRECT_H #ifndef ASM_ARM_DMA_DIRECT_H
#define ASM_ARM_DMA_DIRECT_H 1 #define ASM_ARM_DMA_DIRECT_H 1
static inline dma_addr_t __phys_to_dma(struct device *dev, phys_addr_t paddr) static inline dma_addr_t phys_to_dma(struct device *dev, phys_addr_t paddr)
{ {
unsigned int offset = paddr & ~PAGE_MASK; unsigned int offset = paddr & ~PAGE_MASK;
return pfn_to_dma(dev, __phys_to_pfn(paddr)) + offset; return pfn_to_dma(dev, __phys_to_pfn(paddr)) + offset;
......
...@@ -40,7 +40,7 @@ static struct bmips_dma_range *bmips_dma_ranges; ...@@ -40,7 +40,7 @@ static struct bmips_dma_range *bmips_dma_ranges;
#define FLUSH_RAC 0x100 #define FLUSH_RAC 0x100
dma_addr_t __phys_to_dma(struct device *dev, phys_addr_t pa) dma_addr_t phys_to_dma(struct device *dev, phys_addr_t pa)
{ {
struct bmips_dma_range *r; struct bmips_dma_range *r;
......
...@@ -168,7 +168,7 @@ void __init octeon_pci_dma_init(void) ...@@ -168,7 +168,7 @@ void __init octeon_pci_dma_init(void)
} }
#endif /* CONFIG_PCI */ #endif /* CONFIG_PCI */
dma_addr_t __phys_to_dma(struct device *dev, phys_addr_t paddr) dma_addr_t phys_to_dma(struct device *dev, phys_addr_t paddr)
{ {
#ifdef CONFIG_PCI #ifdef CONFIG_PCI
if (dev && dev_is_pci(dev)) if (dev && dev_is_pci(dev))
......
...@@ -2,7 +2,7 @@ ...@@ -2,7 +2,7 @@
#ifndef _MIPS_DMA_DIRECT_H #ifndef _MIPS_DMA_DIRECT_H
#define _MIPS_DMA_DIRECT_H 1 #define _MIPS_DMA_DIRECT_H 1
dma_addr_t __phys_to_dma(struct device *dev, phys_addr_t paddr); dma_addr_t phys_to_dma(struct device *dev, phys_addr_t paddr);
phys_addr_t dma_to_phys(struct device *dev, dma_addr_t daddr); phys_addr_t dma_to_phys(struct device *dev, dma_addr_t daddr);
#endif /* _MIPS_DMA_DIRECT_H */ #endif /* _MIPS_DMA_DIRECT_H */
// SPDX-License-Identifier: GPL-2.0 // SPDX-License-Identifier: GPL-2.0
#include <linux/dma-direct.h> #include <linux/dma-direct.h>
dma_addr_t __phys_to_dma(struct device *dev, phys_addr_t paddr) dma_addr_t phys_to_dma(struct device *dev, phys_addr_t paddr)
{ {
return paddr | 0x80000000; return paddr | 0x80000000;
} }
......
// SPDX-License-Identifier: GPL-2.0 // SPDX-License-Identifier: GPL-2.0
#include <linux/dma-direct.h> #include <linux/dma-direct.h>
dma_addr_t __phys_to_dma(struct device *dev, phys_addr_t paddr) dma_addr_t phys_to_dma(struct device *dev, phys_addr_t paddr)
{ {
return paddr | 0x80000000; return paddr | 0x80000000;
} }
......
...@@ -4,7 +4,7 @@ ...@@ -4,7 +4,7 @@
#include <linux/swiotlb.h> #include <linux/swiotlb.h>
#include <boot_param.h> #include <boot_param.h>
dma_addr_t __phys_to_dma(struct device *dev, phys_addr_t paddr) dma_addr_t phys_to_dma(struct device *dev, phys_addr_t paddr)
{ {
/* We extract 2bit node id (bit 44~47, only bit 44~45 used now) from /* We extract 2bit node id (bit 44~47, only bit 44~45 used now) from
* Loongson-3's 48bit address space and embed it into 40bit */ * Loongson-3's 48bit address space and embed it into 40bit */
......
...@@ -170,7 +170,7 @@ static inline dma_addr_t ar2315_dev_offset(struct device *dev) ...@@ -170,7 +170,7 @@ static inline dma_addr_t ar2315_dev_offset(struct device *dev)
return 0; return 0;
} }
dma_addr_t __phys_to_dma(struct device *dev, phys_addr_t paddr) dma_addr_t phys_to_dma(struct device *dev, phys_addr_t paddr)
{ {
return paddr + ar2315_dev_offset(dev); return paddr + ar2315_dev_offset(dev);
} }
......
...@@ -25,7 +25,7 @@ ...@@ -25,7 +25,7 @@
/* /*
* Common phys<->dma mapping for platforms using pci xtalk bridge * Common phys<->dma mapping for platforms using pci xtalk bridge
*/ */
dma_addr_t __phys_to_dma(struct device *dev, phys_addr_t paddr) dma_addr_t phys_to_dma(struct device *dev, phys_addr_t paddr)
{ {
struct pci_dev *pdev = to_pci_dev(dev); struct pci_dev *pdev = to_pci_dev(dev);
struct bridge_controller *bc = BRIDGE_CONTROLLER(pdev->bus); struct bridge_controller *bc = BRIDGE_CONTROLLER(pdev->bus);
......
...@@ -18,7 +18,7 @@ ...@@ -18,7 +18,7 @@
#define RAM_OFFSET_MASK 0x3fffffffUL #define RAM_OFFSET_MASK 0x3fffffffUL
dma_addr_t __phys_to_dma(struct device *dev, phys_addr_t paddr) dma_addr_t phys_to_dma(struct device *dev, phys_addr_t paddr)
{ {
dma_addr_t dma_addr = paddr & RAM_OFFSET_MASK; dma_addr_t dma_addr = paddr & RAM_OFFSET_MASK;
......
...@@ -2,7 +2,7 @@ ...@@ -2,7 +2,7 @@
#ifndef ASM_POWERPC_DMA_DIRECT_H #ifndef ASM_POWERPC_DMA_DIRECT_H
#define ASM_POWERPC_DMA_DIRECT_H 1 #define ASM_POWERPC_DMA_DIRECT_H 1
static inline dma_addr_t __phys_to_dma(struct device *dev, phys_addr_t paddr) static inline dma_addr_t phys_to_dma(struct device *dev, phys_addr_t paddr)
{ {
return paddr + dev->archdata.dma_offset; return paddr + dev->archdata.dma_offset;
} }
......
...@@ -3736,7 +3736,7 @@ bounce_map_single(struct device *dev, phys_addr_t paddr, size_t size, ...@@ -3736,7 +3736,7 @@ bounce_map_single(struct device *dev, phys_addr_t paddr, size_t size,
*/ */
if (!IS_ALIGNED(paddr | size, VTD_PAGE_SIZE)) { if (!IS_ALIGNED(paddr | size, VTD_PAGE_SIZE)) {
tlb_addr = swiotlb_tbl_map_single(dev, tlb_addr = swiotlb_tbl_map_single(dev,
__phys_to_dma(dev, io_tlb_start), phys_to_dma_unencrypted(dev, io_tlb_start),
paddr, size, aligned_size, dir, attrs); paddr, size, aligned_size, dir, attrs);
if (tlb_addr == DMA_MAPPING_ERROR) { if (tlb_addr == DMA_MAPPING_ERROR) {
goto swiotlb_error; goto swiotlb_error;
......
...@@ -16,14 +16,29 @@ extern unsigned int zone_dma_bits; ...@@ -16,14 +16,29 @@ extern unsigned int zone_dma_bits;
#ifdef CONFIG_ARCH_HAS_PHYS_TO_DMA #ifdef CONFIG_ARCH_HAS_PHYS_TO_DMA
#include <asm/dma-direct.h> #include <asm/dma-direct.h>
#ifndef phys_to_dma_unencrypted
#define phys_to_dma_unencrypted phys_to_dma
#endif
#else #else
static inline dma_addr_t __phys_to_dma(struct device *dev, phys_addr_t paddr) static inline dma_addr_t phys_to_dma_unencrypted(struct device *dev,
phys_addr_t paddr)
{ {
dma_addr_t dev_addr = (dma_addr_t)paddr; dma_addr_t dev_addr = (dma_addr_t)paddr;
return dev_addr - ((dma_addr_t)dev->dma_pfn_offset << PAGE_SHIFT); return dev_addr - ((dma_addr_t)dev->dma_pfn_offset << PAGE_SHIFT);
} }
/*
* If memory encryption is supported, phys_to_dma will set the memory encryption
* bit in the DMA address, and dma_to_phys will clear it.
* phys_to_dma_unencrypted is for use on special unencrypted memory like swiotlb
* buffers.
*/
static inline dma_addr_t phys_to_dma(struct device *dev, phys_addr_t paddr)
{
return __sme_set(phys_to_dma_unencrypted(dev, paddr));
}
static inline phys_addr_t dma_to_phys(struct device *dev, dma_addr_t dev_addr) static inline phys_addr_t dma_to_phys(struct device *dev, dma_addr_t dev_addr)
{ {
phys_addr_t paddr = (phys_addr_t)dev_addr + phys_addr_t paddr = (phys_addr_t)dev_addr +
...@@ -42,17 +57,6 @@ static inline bool force_dma_unencrypted(struct device *dev) ...@@ -42,17 +57,6 @@ static inline bool force_dma_unencrypted(struct device *dev)
} }
#endif /* CONFIG_ARCH_HAS_FORCE_DMA_UNENCRYPTED */ #endif /* CONFIG_ARCH_HAS_FORCE_DMA_UNENCRYPTED */
/*
* If memory encryption is supported, phys_to_dma will set the memory encryption
* bit in the DMA address, and dma_to_phys will clear it. The raw __phys_to_dma
* version should only be used on non-encrypted memory for
* special occasions like DMA coherent buffers.
*/
static inline dma_addr_t phys_to_dma(struct device *dev, phys_addr_t paddr)
{
return __sme_set(__phys_to_dma(dev, paddr));
}
static inline bool dma_capable(struct device *dev, dma_addr_t addr, size_t size, static inline bool dma_capable(struct device *dev, dma_addr_t addr, size_t size,
bool is_ram) bool is_ram)
{ {
......
...@@ -25,7 +25,7 @@ static inline dma_addr_t phys_to_dma_direct(struct device *dev, ...@@ -25,7 +25,7 @@ static inline dma_addr_t phys_to_dma_direct(struct device *dev,
phys_addr_t phys) phys_addr_t phys)
{ {
if (force_dma_unencrypted(dev)) if (force_dma_unencrypted(dev))
return __phys_to_dma(dev, phys); return phys_to_dma_unencrypted(dev, phys);
return phys_to_dma(dev, phys); return phys_to_dma(dev, phys);
} }
...@@ -438,13 +438,13 @@ int dma_direct_supported(struct device *dev, u64 mask) ...@@ -438,13 +438,13 @@ int dma_direct_supported(struct device *dev, u64 mask)
return 1; return 1;
/* /*
* This check needs to be against the actual bit mask value, so * This check needs to be against the actual bit mask value, so use
* use __phys_to_dma() here so that the SME encryption mask isn't * phys_to_dma_unencrypted() here so that the SME encryption mask isn't
* part of the check. * part of the check.
*/ */
if (IS_ENABLED(CONFIG_ZONE_DMA)) if (IS_ENABLED(CONFIG_ZONE_DMA))
min_mask = min_t(u64, min_mask, DMA_BIT_MASK(zone_dma_bits)); min_mask = min_t(u64, min_mask, DMA_BIT_MASK(zone_dma_bits));
return mask >= __phys_to_dma(dev, min_mask); return mask >= phys_to_dma_unencrypted(dev, min_mask);
} }
size_t dma_direct_max_mapping_size(struct device *dev) size_t dma_direct_max_mapping_size(struct device *dev)
......
...@@ -670,13 +670,13 @@ dma_addr_t swiotlb_map(struct device *dev, phys_addr_t paddr, size_t size, ...@@ -670,13 +670,13 @@ dma_addr_t swiotlb_map(struct device *dev, phys_addr_t paddr, size_t size,
swiotlb_force); swiotlb_force);
swiotlb_addr = swiotlb_tbl_map_single(dev, swiotlb_addr = swiotlb_tbl_map_single(dev,
__phys_to_dma(dev, io_tlb_start), phys_to_dma_unencrypted(dev, io_tlb_start),
paddr, size, size, dir, attrs); paddr, size, size, dir, attrs);
if (swiotlb_addr == (phys_addr_t)DMA_MAPPING_ERROR) if (swiotlb_addr == (phys_addr_t)DMA_MAPPING_ERROR)
return DMA_MAPPING_ERROR; return DMA_MAPPING_ERROR;
/* Ensure that the address returned is DMA'ble */ /* Ensure that the address returned is DMA'ble */
dma_addr = __phys_to_dma(dev, swiotlb_addr); dma_addr = phys_to_dma_unencrypted(dev, swiotlb_addr);
if (unlikely(!dma_capable(dev, dma_addr, size, true))) { if (unlikely(!dma_capable(dev, dma_addr, size, true))) {
swiotlb_tbl_unmap_single(dev, swiotlb_addr, size, size, dir, swiotlb_tbl_unmap_single(dev, swiotlb_addr, size, size, dir,
attrs | DMA_ATTR_SKIP_CPU_SYNC); attrs | DMA_ATTR_SKIP_CPU_SYNC);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment