Commit 17647b1d authored by Richard Russon's avatar Richard Russon

Merge flatcap.org:/home/flatcap/backup/bk/ntfs-2.6

into flatcap.org:/home/flatcap/backup/bk/ntfs-2.6-devel
parents d3762ba0 59f9f96e
...@@ -68,36 +68,6 @@ EXPORT_SYMBOL(alpha_using_srm); ...@@ -68,36 +68,6 @@ EXPORT_SYMBOL(alpha_using_srm);
#endif /* CONFIG_ALPHA_GENERIC */ #endif /* CONFIG_ALPHA_GENERIC */
/* platform dependent support */ /* platform dependent support */
EXPORT_SYMBOL(_inb);
EXPORT_SYMBOL(_inw);
EXPORT_SYMBOL(_inl);
EXPORT_SYMBOL(_outb);
EXPORT_SYMBOL(_outw);
EXPORT_SYMBOL(_outl);
EXPORT_SYMBOL(_readb);
EXPORT_SYMBOL(_readw);
EXPORT_SYMBOL(_readl);
EXPORT_SYMBOL(_writeb);
EXPORT_SYMBOL(_writew);
EXPORT_SYMBOL(_writel);
EXPORT_SYMBOL(___raw_readb);
EXPORT_SYMBOL(___raw_readw);
EXPORT_SYMBOL(___raw_readl);
EXPORT_SYMBOL(___raw_readq);
EXPORT_SYMBOL(___raw_writeb);
EXPORT_SYMBOL(___raw_writew);
EXPORT_SYMBOL(___raw_writel);
EXPORT_SYMBOL(___raw_writeq);
EXPORT_SYMBOL(_memcpy_fromio);
EXPORT_SYMBOL(_memcpy_toio);
EXPORT_SYMBOL(_memset_c_io);
EXPORT_SYMBOL(scr_memcpyw);
EXPORT_SYMBOL(insb);
EXPORT_SYMBOL(insw);
EXPORT_SYMBOL(insl);
EXPORT_SYMBOL(outsb);
EXPORT_SYMBOL(outsw);
EXPORT_SYMBOL(outsl);
EXPORT_SYMBOL(strcat); EXPORT_SYMBOL(strcat);
EXPORT_SYMBOL(strcmp); EXPORT_SYMBOL(strcmp);
EXPORT_SYMBOL(strcpy); EXPORT_SYMBOL(strcpy);
......
...@@ -282,7 +282,7 @@ void ...@@ -282,7 +282,7 @@ void
cia_pci_tbi_try2(struct pci_controller *hose, cia_pci_tbi_try2(struct pci_controller *hose,
dma_addr_t start, dma_addr_t end) dma_addr_t start, dma_addr_t end)
{ {
unsigned long bus_addr; void __iomem *bus_addr;
int ctrl; int ctrl;
/* Put the chip into PCI loopback mode. */ /* Put the chip into PCI loopback mode. */
...@@ -351,7 +351,7 @@ verify_tb_operation(void) ...@@ -351,7 +351,7 @@ verify_tb_operation(void)
struct pci_iommu_arena *arena = pci_isa_hose->sg_isa; struct pci_iommu_arena *arena = pci_isa_hose->sg_isa;
int ctrl, addr0, tag0, pte0, data0; int ctrl, addr0, tag0, pte0, data0;
int temp, use_tbia_try2 = 0; int temp, use_tbia_try2 = 0;
unsigned long bus_addr; void __iomem *bus_addr;
/* pyxis -- tbia is broken */ /* pyxis -- tbia is broken */
if (pci_isa_hose->dense_io_base) if (pci_isa_hose->dense_io_base)
......
...@@ -310,7 +310,7 @@ irongate_init_arch(void) ...@@ -310,7 +310,7 @@ irongate_init_arch(void)
#define GET_GATT_OFF(addr) ((addr & 0x003ff000) >> 12) #define GET_GATT_OFF(addr) ((addr & 0x003ff000) >> 12)
#define GET_GATT(addr) (gatt_pages[GET_PAGE_DIR_IDX(addr)]) #define GET_GATT(addr) (gatt_pages[GET_PAGE_DIR_IDX(addr)])
unsigned long void __iomem *
irongate_ioremap(unsigned long addr, unsigned long size) irongate_ioremap(unsigned long addr, unsigned long size)
{ {
struct vm_struct *area; struct vm_struct *area;
...@@ -320,7 +320,7 @@ irongate_ioremap(unsigned long addr, unsigned long size) ...@@ -320,7 +320,7 @@ irongate_ioremap(unsigned long addr, unsigned long size)
unsigned long gart_bus_addr; unsigned long gart_bus_addr;
if (!alpha_agpgart_size) if (!alpha_agpgart_size)
return addr + IRONGATE_MEM; return (void __iomem *)(addr + IRONGATE_MEM);
gart_bus_addr = (unsigned long)IRONGATE0->bar0 & gart_bus_addr = (unsigned long)IRONGATE0->bar0 &
PCI_BASE_ADDRESS_MEM_MASK; PCI_BASE_ADDRESS_MEM_MASK;
...@@ -339,7 +339,7 @@ irongate_ioremap(unsigned long addr, unsigned long size) ...@@ -339,7 +339,7 @@ irongate_ioremap(unsigned long addr, unsigned long size)
/* /*
* Not found - assume legacy ioremap * Not found - assume legacy ioremap
*/ */
return addr + IRONGATE_MEM; return (void __iomem *)(addr + IRONGATE_MEM);
} while(0); } while(0);
mmio_regs = (u32 *)(((unsigned long)IRONGATE0->bar1 & mmio_regs = (u32 *)(((unsigned long)IRONGATE0->bar1 &
...@@ -353,7 +353,7 @@ irongate_ioremap(unsigned long addr, unsigned long size) ...@@ -353,7 +353,7 @@ irongate_ioremap(unsigned long addr, unsigned long size)
if (addr & ~PAGE_MASK) { if (addr & ~PAGE_MASK) {
printk("AGP ioremap failed... addr not page aligned (0x%lx)\n", printk("AGP ioremap failed... addr not page aligned (0x%lx)\n",
addr); addr);
return addr + IRONGATE_MEM; return (void __iomem *)(addr + IRONGATE_MEM);
} }
last = addr + size - 1; last = addr + size - 1;
size = PAGE_ALIGN(last) - addr; size = PAGE_ALIGN(last) - addr;
...@@ -378,7 +378,7 @@ irongate_ioremap(unsigned long addr, unsigned long size) ...@@ -378,7 +378,7 @@ irongate_ioremap(unsigned long addr, unsigned long size)
* Map it * Map it
*/ */
area = get_vm_area(size, VM_IOREMAP); area = get_vm_area(size, VM_IOREMAP);
if (!area) return (unsigned long)NULL; if (!area) return NULL;
for(baddr = addr, vaddr = (unsigned long)area->addr; for(baddr = addr, vaddr = (unsigned long)area->addr;
baddr <= last; baddr <= last;
...@@ -391,7 +391,7 @@ irongate_ioremap(unsigned long addr, unsigned long size) ...@@ -391,7 +391,7 @@ irongate_ioremap(unsigned long addr, unsigned long size)
pte, PAGE_SIZE, 0)) { pte, PAGE_SIZE, 0)) {
printk("AGP ioremap: FAILED to map...\n"); printk("AGP ioremap: FAILED to map...\n");
vfree(area->addr); vfree(area->addr);
return (unsigned long)NULL; return NULL;
} }
} }
...@@ -402,13 +402,15 @@ irongate_ioremap(unsigned long addr, unsigned long size) ...@@ -402,13 +402,15 @@ irongate_ioremap(unsigned long addr, unsigned long size)
printk("irongate_ioremap(0x%lx, 0x%lx) returning 0x%lx\n", printk("irongate_ioremap(0x%lx, 0x%lx) returning 0x%lx\n",
addr, size, vaddr); addr, size, vaddr);
#endif #endif
return vaddr; return (void __iomem *)vaddr;
} }
void void
irongate_iounmap(unsigned long addr) irongate_iounmap(volatile void __iomem *xaddr)
{ {
unsigned long addr = (unsigned long) xaddr;
if (((long)addr >> 41) == -2) if (((long)addr >> 41) == -2)
return; /* kseg map, nothing to do */ return; /* kseg map, nothing to do */
if (addr) return vfree((void *)(PAGE_MASK & addr)); if (addr)
return vfree((void *)(PAGE_MASK & addr));
} }
...@@ -610,11 +610,84 @@ marvel_pci_tbi(struct pci_controller *hose, dma_addr_t start, dma_addr_t end) ...@@ -610,11 +610,84 @@ marvel_pci_tbi(struct pci_controller *hose, dma_addr_t start, dma_addr_t end)
csrs->POx_SG_TBIA.csr; csrs->POx_SG_TBIA.csr;
} }
/*
* RTC Support
*/
struct marvel_rtc_access_info {
unsigned long function;
unsigned long index;
unsigned long data;
};
static void
__marvel_access_rtc(void *info)
{
struct marvel_rtc_access_info *rtc_access = info;
register unsigned long __r0 __asm__("$0");
register unsigned long __r16 __asm__("$16") = rtc_access->function;
register unsigned long __r17 __asm__("$17") = rtc_access->index;
register unsigned long __r18 __asm__("$18") = rtc_access->data;
__asm__ __volatile__(
"call_pal %4 # cserve rtc"
: "=r"(__r16), "=r"(__r17), "=r"(__r18), "=r"(__r0)
: "i"(PAL_cserve), "0"(__r16), "1"(__r17), "2"(__r18)
: "$1", "$22", "$23", "$24", "$25");
rtc_access->data = __r0;
}
static u8
__marvel_rtc_io(u8 b, unsigned long addr, int write)
{
static u8 index = 0;
struct marvel_rtc_access_info rtc_access;
u8 ret = 0;
switch(addr) {
case 0x70: /* RTC_PORT(0) */
if (write) index = b;
ret = index;
break;
case 0x71: /* RTC_PORT(1) */
rtc_access.index = index;
rtc_access.data = BCD_TO_BIN(b);
rtc_access.function = 0x48 + !write; /* GET/PUT_TOY */
#ifdef CONFIG_SMP
if (smp_processor_id() != boot_cpuid)
smp_call_function_on_cpu(__marvel_access_rtc,
&rtc_access, 1, 1,
cpumask_of_cpu(boot_cpuid));
else
__marvel_access_rtc(&rtc_access);
#else
__marvel_access_rtc(&rtc_access);
#endif
ret = BIN_TO_BCD(rtc_access.data);
break;
default:
printk(KERN_WARNING "Illegal RTC port %lx\n", addr);
break;
}
return ret;
}
/* /*
* IO map support. * IO map support.
*/ */
unsigned long
#define __marvel_is_mem_vga(a) (((a) >= 0xa0000) && ((a) <= 0xc0000))
void __iomem *
marvel_ioremap(unsigned long addr, unsigned long size) marvel_ioremap(unsigned long addr, unsigned long size)
{ {
struct pci_controller *hose; struct pci_controller *hose;
...@@ -633,8 +706,6 @@ marvel_ioremap(unsigned long addr, unsigned long size) ...@@ -633,8 +706,6 @@ marvel_ioremap(unsigned long addr, unsigned long size)
} }
#endif #endif
if (!marvel_is_ioaddr(addr)) return 0UL;
/* /*
* Find the hose. * Find the hose.
*/ */
...@@ -643,7 +714,7 @@ marvel_ioremap(unsigned long addr, unsigned long size) ...@@ -643,7 +714,7 @@ marvel_ioremap(unsigned long addr, unsigned long size)
break; break;
} }
if (!hose) if (!hose)
return 0UL; return NULL;
/* /*
* We have the hose - calculate the bus limits. * We have the hose - calculate the bus limits.
...@@ -655,15 +726,17 @@ marvel_ioremap(unsigned long addr, unsigned long size) ...@@ -655,15 +726,17 @@ marvel_ioremap(unsigned long addr, unsigned long size)
* Is it direct-mapped? * Is it direct-mapped?
*/ */
if ((baddr >= __direct_map_base) && if ((baddr >= __direct_map_base) &&
((baddr + size - 1) < __direct_map_base + __direct_map_size)) ((baddr + size - 1) < __direct_map_base + __direct_map_size)) {
return IDENT_ADDR | (baddr - __direct_map_base); addr = IDENT_ADDR | (baddr - __direct_map_base);
return (void __iomem *) addr;
}
/* /*
* Check the scatter-gather arena. * Check the scatter-gather arena.
*/ */
if (hose->sg_pci && if (hose->sg_pci &&
baddr >= (unsigned long)hose->sg_pci->dma_base && baddr >= (unsigned long)hose->sg_pci->dma_base &&
last < (unsigned long)hose->sg_pci->dma_base + hose->sg_pci->size){ last < (unsigned long)hose->sg_pci->dma_base + hose->sg_pci->size) {
/* /*
* Adjust the limits (mappings must be page aligned) * Adjust the limits (mappings must be page aligned)
...@@ -677,7 +750,9 @@ marvel_ioremap(unsigned long addr, unsigned long size) ...@@ -677,7 +750,9 @@ marvel_ioremap(unsigned long addr, unsigned long size)
* Map it. * Map it.
*/ */
area = get_vm_area(size, VM_IOREMAP); area = get_vm_area(size, VM_IOREMAP);
if (!area) return (unsigned long)NULL; if (!area)
return NULL;
ptes = hose->sg_pci->ptes; ptes = hose->sg_pci->ptes;
for (vaddr = (unsigned long)area->addr; for (vaddr = (unsigned long)area->addr;
baddr <= last; baddr <= last;
...@@ -686,7 +761,7 @@ marvel_ioremap(unsigned long addr, unsigned long size) ...@@ -686,7 +761,7 @@ marvel_ioremap(unsigned long addr, unsigned long size)
if (!(pfn & 1)) { if (!(pfn & 1)) {
printk("ioremap failed... pte not valid...\n"); printk("ioremap failed... pte not valid...\n");
vfree(area->addr); vfree(area->addr);
return 0UL; return NULL;
} }
pfn >>= 1; /* make it a true pfn */ pfn >>= 1; /* make it a true pfn */
...@@ -695,7 +770,7 @@ marvel_ioremap(unsigned long addr, unsigned long size) ...@@ -695,7 +770,7 @@ marvel_ioremap(unsigned long addr, unsigned long size)
PAGE_SIZE, 0)) { PAGE_SIZE, 0)) {
printk("FAILED to map...\n"); printk("FAILED to map...\n");
vfree(area->addr); vfree(area->addr);
return 0UL; return NULL;
} }
} }
...@@ -703,101 +778,81 @@ marvel_ioremap(unsigned long addr, unsigned long size) ...@@ -703,101 +778,81 @@ marvel_ioremap(unsigned long addr, unsigned long size)
vaddr = (unsigned long)area->addr + (addr & ~PAGE_MASK); vaddr = (unsigned long)area->addr + (addr & ~PAGE_MASK);
return vaddr; return (void __iomem *) vaddr;
} }
/* return NULL;
* Not found - assume legacy ioremap.
*/
return addr;
} }
void void
marvel_iounmap(unsigned long addr) marvel_iounmap(volatile void __iomem *xaddr)
{ {
if (((long)addr >> 41) == -2) unsigned long addr = (unsigned long) xaddr;
return; /* kseg map, nothing to do */ if (addr >= VMALLOC_START)
if (addr)
vfree((void *)(PAGE_MASK & addr)); vfree((void *)(PAGE_MASK & addr));
} }
#ifndef CONFIG_ALPHA_GENERIC int
EXPORT_SYMBOL(marvel_ioremap); marvel_is_mmio(const volatile void __iomem *xaddr)
EXPORT_SYMBOL(marvel_iounmap);
#endif
/*
* RTC Support
*/
struct marvel_rtc_access_info {
unsigned long function;
unsigned long index;
unsigned long data;
};
static void
__marvel_access_rtc(void *info)
{ {
struct marvel_rtc_access_info *rtc_access = info; unsigned long addr = (unsigned long) xaddr;
register unsigned long __r0 __asm__("$0");
register unsigned long __r16 __asm__("$16") = rtc_access->function;
register unsigned long __r17 __asm__("$17") = rtc_access->index;
register unsigned long __r18 __asm__("$18") = rtc_access->data;
__asm__ __volatile__(
"call_pal %4 # cserve rtc"
: "=r"(__r16), "=r"(__r17), "=r"(__r18), "=r"(__r0)
: "i"(PAL_cserve), "0"(__r16), "1"(__r17), "2"(__r18)
: "$1", "$22", "$23", "$24", "$25");
rtc_access->data = __r0; if (addr >= VMALLOC_START)
return 1;
else
return (addr & 0xFF000000UL) == 0;
} }
u8 #define __marvel_is_port_vga(a) \
__marvel_rtc_io(int write, u8 b, unsigned long addr) (((a) >= 0x3b0) && ((a) < 0x3e0) && ((a) != 0x3b3) && ((a) != 0x3d3))
{ #define __marvel_is_port_kbd(a) (((a) == 0x60) || ((a) == 0x64))
struct marvel_rtc_access_info rtc_access = {0, }; #define __marvel_is_port_rtc(a) (((a) == 0x70) || ((a) == 0x71))
static u8 index = 0;
u8 ret = 0;
switch(addr) {
case 0x70: /* RTC_PORT(0) */
if (write) index = b;
ret = index;
break;
case 0x71: /* RTC_PORT(1) */ void __iomem *marvel_ioportmap (unsigned long addr)
rtc_access.index = index; {
rtc_access.data = BCD_TO_BIN(b); if (__marvel_is_port_rtc (addr) || __marvel_is_port_kbd(addr))
rtc_access.function = 0x49; /* GET_TOY */ ;
if (write) rtc_access.function = 0x48; /* PUT_TOY */ #ifdef CONFIG_VGA_HOSE
else if (__marvel_is_port_vga (addr) && pci_vga_hose)
#ifdef CONFIG_SMP addr += pci_vga_hose->io_space->start;
if (smp_processor_id() != boot_cpuid)
smp_call_function_on_cpu(__marvel_access_rtc,
&rtc_access,
1, /* retry */
1, /* wait */
1UL << boot_cpuid);
else
__marvel_access_rtc(&rtc_access);
#else
__marvel_access_rtc(&rtc_access);
#endif #endif
ret = BIN_TO_BCD(rtc_access.data); else
return NULL;
break; return (void __iomem *)addr;
}
default: unsigned int
printk(KERN_WARNING "Illegal RTC port %lx\n", addr); marvel_ioread8(void __iomem *xaddr)
break; {
} unsigned long addr = (unsigned long) xaddr;
if (__marvel_is_port_kbd(addr))
return 0;
else if (__marvel_is_port_rtc(addr))
return __marvel_rtc_io(0, addr, 0);
else
return __kernel_ldbu(*(vucp)addr);
}
return ret; void
marvel_iowrite8(u8 b, void __iomem *xaddr)
{
unsigned long addr = (unsigned long) xaddr;
if (__marvel_is_port_kbd(addr))
return;
else if (__marvel_is_port_rtc(addr))
__marvel_rtc_io(b, addr, 1);
else
__kernel_stb(b, *(vucp)addr);
} }
#ifndef CONFIG_ALPHA_GENERIC
EXPORT_SYMBOL(marvel_ioremap);
EXPORT_SYMBOL(marvel_iounmap);
EXPORT_SYMBOL(marvel_is_mmio);
EXPORT_SYMBOL(marvel_ioportmap);
EXPORT_SYMBOL(marvel_ioread8);
EXPORT_SYMBOL(marvel_iowrite8);
#endif
/* /*
* NUMA Support * NUMA Support
......
...@@ -461,7 +461,8 @@ titan_kill_arch(int mode) ...@@ -461,7 +461,8 @@ titan_kill_arch(int mode)
/* /*
* IO map support. * IO map support.
*/ */
unsigned long
void __iomem *
titan_ioremap(unsigned long addr, unsigned long size) titan_ioremap(unsigned long addr, unsigned long size)
{ {
int h = (addr & TITAN_HOSE_MASK) >> TITAN_HOSE_SHIFT; int h = (addr & TITAN_HOSE_MASK) >> TITAN_HOSE_SHIFT;
...@@ -487,15 +488,19 @@ titan_ioremap(unsigned long addr, unsigned long size) ...@@ -487,15 +488,19 @@ titan_ioremap(unsigned long addr, unsigned long size)
* Find the hose. * Find the hose.
*/ */
for (hose = hose_head; hose; hose = hose->next) for (hose = hose_head; hose; hose = hose->next)
if (hose->index == h) break; if (hose->index == h)
if (!hose) return (unsigned long)NULL; break;
if (!hose)
return NULL;
/* /*
* Is it direct-mapped? * Is it direct-mapped?
*/ */
if ((baddr >= __direct_map_base) && if ((baddr >= __direct_map_base) &&
((baddr + size - 1) < __direct_map_base + __direct_map_size)) ((baddr + size - 1) < __direct_map_base + __direct_map_size)) {
return addr - __direct_map_base + TITAN_MEM_BIAS; vaddr = addr - __direct_map_base + TITAN_MEM_BIAS;
return (void __iomem *) vaddr;
}
/* /*
* Check the scatter-gather arena. * Check the scatter-gather arena.
...@@ -516,7 +521,9 @@ titan_ioremap(unsigned long addr, unsigned long size) ...@@ -516,7 +521,9 @@ titan_ioremap(unsigned long addr, unsigned long size)
* Map it * Map it
*/ */
area = get_vm_area(size, VM_IOREMAP); area = get_vm_area(size, VM_IOREMAP);
if (!area) return (unsigned long)NULL; if (!area)
return NULL;
ptes = hose->sg_pci->ptes; ptes = hose->sg_pci->ptes;
for (vaddr = (unsigned long)area->addr; for (vaddr = (unsigned long)area->addr;
baddr <= last; baddr <= last;
...@@ -525,7 +532,7 @@ titan_ioremap(unsigned long addr, unsigned long size) ...@@ -525,7 +532,7 @@ titan_ioremap(unsigned long addr, unsigned long size)
if (!(pfn & 1)) { if (!(pfn & 1)) {
printk("ioremap failed... pte not valid...\n"); printk("ioremap failed... pte not valid...\n");
vfree(area->addr); vfree(area->addr);
return (unsigned long)NULL; return NULL;
} }
pfn >>= 1; /* make it a true pfn */ pfn >>= 1; /* make it a true pfn */
...@@ -534,35 +541,42 @@ titan_ioremap(unsigned long addr, unsigned long size) ...@@ -534,35 +541,42 @@ titan_ioremap(unsigned long addr, unsigned long size)
PAGE_SIZE, 0)) { PAGE_SIZE, 0)) {
printk("FAILED to map...\n"); printk("FAILED to map...\n");
vfree(area->addr); vfree(area->addr);
return (unsigned long)NULL; return NULL;
} }
} }
flush_tlb_all(); flush_tlb_all();
vaddr = (unsigned long)area->addr + (addr & ~PAGE_MASK); vaddr = (unsigned long)area->addr + (addr & ~PAGE_MASK);
return vaddr; return (void __iomem *) vaddr;
} }
/* return NULL;
* Not found - assume legacy ioremap.
*/
return addr + TITAN_MEM_BIAS;
} }
void void
titan_iounmap(unsigned long addr) titan_iounmap(volatile void __iomem *xaddr)
{ {
if (((long)addr >> 41) == -2) unsigned long addr = (unsigned long) xaddr;
return; /* kseg map, nothing to do */ if (addr >= VMALLOC_START)
if (addr)
vfree((void *)(PAGE_MASK & addr)); vfree((void *)(PAGE_MASK & addr));
} }
int
titan_is_mmio(const volatile void __iomem *xaddr)
{
unsigned long addr = (unsigned long) xaddr;
if (addr >= VMALLOC_START)
return 1;
else
return (addr & 0x100000000UL) == 0;
}
#ifndef CONFIG_ALPHA_GENERIC #ifndef CONFIG_ALPHA_GENERIC
EXPORT_SYMBOL(titan_ioremap); EXPORT_SYMBOL(titan_ioremap);
EXPORT_SYMBOL(titan_iounmap); EXPORT_SYMBOL(titan_iounmap);
EXPORT_SYMBOL(titan_is_mmio);
#endif #endif
/* /*
......
...@@ -177,7 +177,7 @@ titan_parse_p_perror(int which, int port, u64 perror, int print) ...@@ -177,7 +177,7 @@ titan_parse_p_perror(int which, int port, u64 perror, int print)
#define TITAN__PCHIP_PERROR__CMD__S (52) #define TITAN__PCHIP_PERROR__CMD__S (52)
#define TITAN__PCHIP_PERROR__CMD__M (0x0f) #define TITAN__PCHIP_PERROR__CMD__M (0x0f)
#define TITAN__PCHIP_PERROR__ADDR__S (14) #define TITAN__PCHIP_PERROR__ADDR__S (14)
#define TITAN__PCHIP_PERROR__ADDR__M (0x1ffffffff) #define TITAN__PCHIP_PERROR__ADDR__M (0x1fffffffful)
if (!(perror & TITAN__PCHIP_PERROR__ERRMASK)) if (!(perror & TITAN__PCHIP_PERROR__ERRMASK))
return MCHK_DISPOSITION_UNKNOWN_ERROR; return MCHK_DISPOSITION_UNKNOWN_ERROR;
......
...@@ -44,56 +44,60 @@ ...@@ -44,56 +44,60 @@
#define DO_DEFAULT_RTC rtc_port: 0x70 #define DO_DEFAULT_RTC rtc_port: 0x70
#define DO_EV4_MMU \ #define DO_EV4_MMU \
max_asn: EV4_MAX_ASN, \ .max_asn = EV4_MAX_ASN, \
mv_switch_mm: ev4_switch_mm, \ .mv_switch_mm = ev4_switch_mm, \
mv_activate_mm: ev4_activate_mm, \ .mv_activate_mm = ev4_activate_mm, \
mv_flush_tlb_current: ev4_flush_tlb_current, \ .mv_flush_tlb_current = ev4_flush_tlb_current, \
mv_flush_tlb_current_page: ev4_flush_tlb_current_page .mv_flush_tlb_current_page = ev4_flush_tlb_current_page
#define DO_EV5_MMU \ #define DO_EV5_MMU \
max_asn: EV5_MAX_ASN, \ .max_asn = EV5_MAX_ASN, \
mv_switch_mm: ev5_switch_mm, \ .mv_switch_mm = ev5_switch_mm, \
mv_activate_mm: ev5_activate_mm, \ .mv_activate_mm = ev5_activate_mm, \
mv_flush_tlb_current: ev5_flush_tlb_current, \ .mv_flush_tlb_current = ev5_flush_tlb_current, \
mv_flush_tlb_current_page: ev5_flush_tlb_current_page .mv_flush_tlb_current_page = ev5_flush_tlb_current_page
#define DO_EV6_MMU \ #define DO_EV6_MMU \
max_asn: EV6_MAX_ASN, \ .max_asn = EV6_MAX_ASN, \
mv_switch_mm: ev5_switch_mm, \ .mv_switch_mm = ev5_switch_mm, \
mv_activate_mm: ev5_activate_mm, \ .mv_activate_mm = ev5_activate_mm, \
mv_flush_tlb_current: ev5_flush_tlb_current, \ .mv_flush_tlb_current = ev5_flush_tlb_current, \
mv_flush_tlb_current_page: ev5_flush_tlb_current_page .mv_flush_tlb_current_page = ev5_flush_tlb_current_page
#define DO_EV7_MMU \ #define DO_EV7_MMU \
max_asn: EV6_MAX_ASN, \ .max_asn = EV6_MAX_ASN, \
mv_switch_mm: ev5_switch_mm, \ .mv_switch_mm = ev5_switch_mm, \
mv_activate_mm: ev5_activate_mm, \ .mv_activate_mm = ev5_activate_mm, \
mv_flush_tlb_current: ev5_flush_tlb_current, \ .mv_flush_tlb_current = ev5_flush_tlb_current, \
mv_flush_tlb_current_page: ev5_flush_tlb_current_page .mv_flush_tlb_current_page = ev5_flush_tlb_current_page
#define IO_LITE(UP,low) \ #define IO_LITE(UP,low) \
hae_register: (unsigned long *) CAT(UP,_HAE_ADDRESS), \ .hae_register = (unsigned long *) CAT(UP,_HAE_ADDRESS), \
iack_sc: CAT(UP,_IACK_SC), \ .iack_sc = CAT(UP,_IACK_SC), \
mv_inb: CAT(low,_inb), \ .mv_ioread8 = CAT(low,_ioread8), \
mv_inw: CAT(low,_inw), \ .mv_ioread16 = CAT(low,_ioread16), \
mv_inl: CAT(low,_inl), \ .mv_ioread32 = CAT(low,_ioread32), \
mv_outb: CAT(low,_outb), \ .mv_iowrite8 = CAT(low,_iowrite8), \
mv_outw: CAT(low,_outw), \ .mv_iowrite16 = CAT(low,_iowrite16), \
mv_outl: CAT(low,_outl), \ .mv_iowrite32 = CAT(low,_iowrite32), \
mv_readb: CAT(low,_readb), \ .mv_readb = CAT(low,_readb), \
mv_readw: CAT(low,_readw), \ .mv_readw = CAT(low,_readw), \
mv_readl: CAT(low,_readl), \ .mv_readl = CAT(low,_readl), \
mv_readq: CAT(low,_readq), \ .mv_readq = CAT(low,_readq), \
mv_writeb: CAT(low,_writeb), \ .mv_writeb = CAT(low,_writeb), \
mv_writew: CAT(low,_writew), \ .mv_writew = CAT(low,_writew), \
mv_writel: CAT(low,_writel), \ .mv_writel = CAT(low,_writel), \
mv_writeq: CAT(low,_writeq), \ .mv_writeq = CAT(low,_writeq), \
mv_ioremap: CAT(low,_ioremap), \ .mv_ioportmap = CAT(low,_ioportmap), \
mv_iounmap: CAT(low,_iounmap) \ .mv_ioremap = CAT(low,_ioremap), \
.mv_iounmap = CAT(low,_iounmap), \
.mv_is_ioaddr = CAT(low,_is_ioaddr), \
.mv_is_mmio = CAT(low,_is_mmio) \
#define IO(UP,low) \ #define IO(UP,low) \
IO_LITE(UP,low), \ IO_LITE(UP,low), \
pci_ops: &CAT(low,_pci_ops) .pci_ops = &CAT(low,_pci_ops), \
.mv_pci_tbi = CAT(low,_pci_tbi)
#define DO_APECS_IO IO(APECS,apecs) #define DO_APECS_IO IO(APECS,apecs)
#define DO_CIA_IO IO(CIA,cia) #define DO_CIA_IO IO(CIA,cia)
...@@ -108,23 +112,8 @@ ...@@ -108,23 +112,8 @@
#define DO_WILDFIRE_IO IO(WILDFIRE,wildfire) #define DO_WILDFIRE_IO IO(WILDFIRE,wildfire)
#define DO_PYXIS_IO IO_LITE(CIA,cia_bwx), \ #define DO_PYXIS_IO IO_LITE(CIA,cia_bwx), \
pci_ops: &CAT(cia,_pci_ops) .pci_ops = &cia_pci_ops, \
.mv_pci_tbi = cia_pci_tbi
#define BUS(which) \
mv_is_ioaddr: CAT(which,_is_ioaddr), \
mv_pci_tbi: CAT(which,_pci_tbi)
#define DO_APECS_BUS BUS(apecs)
#define DO_CIA_BUS BUS(cia)
#define DO_IRONGATE_BUS BUS(irongate)
#define DO_LCA_BUS BUS(lca)
#define DO_MARVEL_BUS BUS(marvel)
#define DO_MCPCIA_BUS BUS(mcpcia)
#define DO_POLARIS_BUS BUS(polaris)
#define DO_T2_BUS BUS(t2)
#define DO_TSUNAMI_BUS BUS(tsunami)
#define DO_TITAN_BUS BUS(titan)
#define DO_WILDFIRE_BUS BUS(wildfire)
/* /*
* In a GENERIC kernel, we have lots of these vectors floating about, * In a GENERIC kernel, we have lots of these vectors floating about,
......
...@@ -1110,46 +1110,47 @@ osf_getrusage(int who, struct rusage32 __user *ru) ...@@ -1110,46 +1110,47 @@ osf_getrusage(int who, struct rusage32 __user *ru)
return copy_to_user(ru, &r, sizeof(r)) ? -EFAULT : 0; return copy_to_user(ru, &r, sizeof(r)) ? -EFAULT : 0;
} }
asmlinkage int asmlinkage long
osf_wait4(pid_t pid, int __user *ustatus, int options, struct rusage32 __user *ur) osf_wait4(pid_t pid, int __user *ustatus, int options,
struct rusage32 __user *ur)
{ {
if (!ur) { struct rusage r;
long ret, err;
mm_segment_t old_fs;
if (!ur)
return sys_wait4(pid, ustatus, options, NULL); return sys_wait4(pid, ustatus, options, NULL);
} else {
struct rusage r; old_fs = get_fs();
int ret, status;
mm_segment_t old_fs = get_fs();
set_fs (KERNEL_DS); set_fs (KERNEL_DS);
ret = sys_wait4(pid, &status, options, &r); ret = sys_wait4(pid, ustatus, options, (struct rusage __user *) &r);
set_fs (old_fs); set_fs (old_fs);
if (!access_ok(VERIFY_WRITE, ur, sizeof(*ur))) if (!access_ok(VERIFY_WRITE, ur, sizeof(*ur)))
return -EFAULT; return -EFAULT;
__put_user(r.ru_utime.tv_sec, &ur->ru_utime.tv_sec);
__put_user(r.ru_utime.tv_usec, &ur->ru_utime.tv_usec);
__put_user(r.ru_stime.tv_sec, &ur->ru_stime.tv_sec);
__put_user(r.ru_stime.tv_usec, &ur->ru_stime.tv_usec);
__put_user(r.ru_maxrss, &ur->ru_maxrss);
__put_user(r.ru_ixrss, &ur->ru_ixrss);
__put_user(r.ru_idrss, &ur->ru_idrss);
__put_user(r.ru_isrss, &ur->ru_isrss);
__put_user(r.ru_minflt, &ur->ru_minflt);
__put_user(r.ru_majflt, &ur->ru_majflt);
__put_user(r.ru_nswap, &ur->ru_nswap);
__put_user(r.ru_inblock, &ur->ru_inblock);
__put_user(r.ru_oublock, &ur->ru_oublock);
__put_user(r.ru_msgsnd, &ur->ru_msgsnd);
__put_user(r.ru_msgrcv, &ur->ru_msgrcv);
__put_user(r.ru_nsignals, &ur->ru_nsignals);
__put_user(r.ru_nvcsw, &ur->ru_nvcsw);
if (__put_user(r.ru_nivcsw, &ur->ru_nivcsw))
return -EFAULT;
if (ustatus && put_user(status, ustatus)) err = 0;
return -EFAULT; err |= __put_user(r.ru_utime.tv_sec, &ur->ru_utime.tv_sec);
return ret; err |= __put_user(r.ru_utime.tv_usec, &ur->ru_utime.tv_usec);
} err |= __put_user(r.ru_stime.tv_sec, &ur->ru_stime.tv_sec);
err |= __put_user(r.ru_stime.tv_usec, &ur->ru_stime.tv_usec);
err |= __put_user(r.ru_maxrss, &ur->ru_maxrss);
err |= __put_user(r.ru_ixrss, &ur->ru_ixrss);
err |= __put_user(r.ru_idrss, &ur->ru_idrss);
err |= __put_user(r.ru_isrss, &ur->ru_isrss);
err |= __put_user(r.ru_minflt, &ur->ru_minflt);
err |= __put_user(r.ru_majflt, &ur->ru_majflt);
err |= __put_user(r.ru_nswap, &ur->ru_nswap);
err |= __put_user(r.ru_inblock, &ur->ru_inblock);
err |= __put_user(r.ru_oublock, &ur->ru_oublock);
err |= __put_user(r.ru_msgsnd, &ur->ru_msgsnd);
err |= __put_user(r.ru_msgrcv, &ur->ru_msgrcv);
err |= __put_user(r.ru_nsignals, &ur->ru_nsignals);
err |= __put_user(r.ru_nvcsw, &ur->ru_nvcsw);
err |= __put_user(r.ru_nivcsw, &ur->ru_nivcsw);
return err ? err : ret;
} }
/* /*
......
...@@ -200,3 +200,15 @@ dma_set_mask(struct device *dev, u64 mask) ...@@ -200,3 +200,15 @@ dma_set_mask(struct device *dev, u64 mask)
return 0; return 0;
} }
void __iomem *pci_iomap(struct pci_dev *dev, int bar, unsigned long maxlen)
{
return NULL;
}
void pci_iounmap(struct pci_dev *dev, void __iomem * addr)
{
}
EXPORT_SYMBOL(pci_iomap);
EXPORT_SYMBOL(pci_iounmap);
...@@ -531,3 +531,37 @@ sys_pciconfig_iobase(long which, unsigned long bus, unsigned long dfn) ...@@ -531,3 +531,37 @@ sys_pciconfig_iobase(long which, unsigned long bus, unsigned long dfn)
return -EOPNOTSUPP; return -EOPNOTSUPP;
} }
/* Create an __iomem token from a PCI BAR. Copied from lib/iomap.c with
no changes, since we don't want the other things in that object file. */
void __iomem *pci_iomap(struct pci_dev *dev, int bar, unsigned long maxlen)
{
unsigned long start = pci_resource_start(dev, bar);
unsigned long len = pci_resource_len(dev, bar);
unsigned long flags = pci_resource_flags(dev, bar);
if (!len || !start)
return NULL;
if (maxlen && len > maxlen)
len = maxlen;
if (flags & IORESOURCE_IO)
return ioport_map(start, len);
if (flags & IORESOURCE_MEM) {
/* Not checking IORESOURCE_CACHEABLE because alpha does
not distinguish between ioremap and ioremap_nocache. */
return ioremap(start, len);
}
return NULL;
}
/* Destroy that token. Not copied from lib/iomap.c. */
void pci_iounmap(struct pci_dev *dev, void __iomem * addr)
{
if (__is_mmio(addr))
iounmap(addr);
}
EXPORT_SYMBOL(pci_iomap);
EXPORT_SYMBOL(pci_iounmap);
...@@ -146,7 +146,7 @@ srmcons_write(struct tty_struct *tty, int from_user, ...@@ -146,7 +146,7 @@ srmcons_write(struct tty_struct *tty, int from_user,
if (c > sizeof(tmp)) if (c > sizeof(tmp))
c = sizeof(tmp); c = sizeof(tmp);
c -= copy_from_user(tmp, buf, c); c -= copy_from_user(tmp, (const char __user *) buf, c);
if (!c) { if (!c) {
printk("%s: EFAULT (count %d)\n", printk("%s: EFAULT (count %d)\n",
......
...@@ -274,7 +274,6 @@ struct alpha_machine_vector alcor_mv __initmv = { ...@@ -274,7 +274,6 @@ struct alpha_machine_vector alcor_mv __initmv = {
DO_EV5_MMU, DO_EV5_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_CIA_IO, DO_CIA_IO,
DO_CIA_BUS,
.machine_check = cia_machine_check, .machine_check = cia_machine_check,
.max_isa_dma_address = ALPHA_ALCOR_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_ALCOR_MAX_ISA_DMA_ADDRESS,
.min_io_address = EISA_DEFAULT_IO_BASE, .min_io_address = EISA_DEFAULT_IO_BASE,
...@@ -292,7 +291,7 @@ struct alpha_machine_vector alcor_mv __initmv = { ...@@ -292,7 +291,7 @@ struct alpha_machine_vector alcor_mv __initmv = {
.pci_swizzle = common_swizzle, .pci_swizzle = common_swizzle,
.sys = { .cia = { .sys = { .cia = {
.gru_int_req_bits = ALCOR_GRU_INT_REQ_BITS .gru_int_req_bits = ALCOR_GRU_INT_REQ_BITS
}} }}
}; };
ALIAS_MV(alcor) ALIAS_MV(alcor)
...@@ -302,7 +301,6 @@ struct alpha_machine_vector xlt_mv __initmv = { ...@@ -302,7 +301,6 @@ struct alpha_machine_vector xlt_mv __initmv = {
DO_EV5_MMU, DO_EV5_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_CIA_IO, DO_CIA_IO,
DO_CIA_BUS,
.machine_check = cia_machine_check, .machine_check = cia_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = EISA_DEFAULT_IO_BASE, .min_io_address = EISA_DEFAULT_IO_BASE,
...@@ -320,7 +318,7 @@ struct alpha_machine_vector xlt_mv __initmv = { ...@@ -320,7 +318,7 @@ struct alpha_machine_vector xlt_mv __initmv = {
.pci_swizzle = common_swizzle, .pci_swizzle = common_swizzle,
.sys = { .cia = { .sys = { .cia = {
.gru_int_req_bits = XLT_GRU_INT_REQ_BITS .gru_int_req_bits = XLT_GRU_INT_REQ_BITS
}} }}
}; };
......
...@@ -327,7 +327,6 @@ struct alpha_machine_vector cabriolet_mv __initmv = { ...@@ -327,7 +327,6 @@ struct alpha_machine_vector cabriolet_mv __initmv = {
DO_EV4_MMU, DO_EV4_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_APECS_IO, DO_APECS_IO,
DO_APECS_BUS,
.machine_check = apecs_machine_check, .machine_check = apecs_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
...@@ -354,7 +353,6 @@ struct alpha_machine_vector eb164_mv __initmv = { ...@@ -354,7 +353,6 @@ struct alpha_machine_vector eb164_mv __initmv = {
DO_EV5_MMU, DO_EV5_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_CIA_IO, DO_CIA_IO,
DO_CIA_BUS,
.machine_check = cia_machine_check, .machine_check = cia_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
...@@ -380,7 +378,6 @@ struct alpha_machine_vector eb66p_mv __initmv = { ...@@ -380,7 +378,6 @@ struct alpha_machine_vector eb66p_mv __initmv = {
DO_EV4_MMU, DO_EV4_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_LCA_IO, DO_LCA_IO,
DO_LCA_BUS,
.machine_check = lca_machine_check, .machine_check = lca_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
...@@ -405,7 +402,6 @@ struct alpha_machine_vector lx164_mv __initmv = { ...@@ -405,7 +402,6 @@ struct alpha_machine_vector lx164_mv __initmv = {
DO_EV5_MMU, DO_EV5_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_PYXIS_IO, DO_PYXIS_IO,
DO_CIA_BUS,
.machine_check = cia_machine_check, .machine_check = cia_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
...@@ -432,7 +428,6 @@ struct alpha_machine_vector pc164_mv __initmv = { ...@@ -432,7 +428,6 @@ struct alpha_machine_vector pc164_mv __initmv = {
DO_EV5_MMU, DO_EV5_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_CIA_IO, DO_CIA_IO,
DO_CIA_BUS,
.machine_check = cia_machine_check, .machine_check = cia_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
......
...@@ -569,7 +569,6 @@ struct alpha_machine_vector dp264_mv __initmv = { ...@@ -569,7 +569,6 @@ struct alpha_machine_vector dp264_mv __initmv = {
DO_EV6_MMU, DO_EV6_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_TSUNAMI_IO, DO_TSUNAMI_IO,
DO_TSUNAMI_BUS,
.machine_check = tsunami_machine_check, .machine_check = tsunami_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
...@@ -594,7 +593,6 @@ struct alpha_machine_vector monet_mv __initmv = { ...@@ -594,7 +593,6 @@ struct alpha_machine_vector monet_mv __initmv = {
DO_EV6_MMU, DO_EV6_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_TSUNAMI_IO, DO_TSUNAMI_IO,
DO_TSUNAMI_BUS,
.machine_check = tsunami_machine_check, .machine_check = tsunami_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
...@@ -618,7 +616,6 @@ struct alpha_machine_vector webbrick_mv __initmv = { ...@@ -618,7 +616,6 @@ struct alpha_machine_vector webbrick_mv __initmv = {
DO_EV6_MMU, DO_EV6_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_TSUNAMI_IO, DO_TSUNAMI_IO,
DO_TSUNAMI_BUS,
.machine_check = tsunami_machine_check, .machine_check = tsunami_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
...@@ -642,7 +639,6 @@ struct alpha_machine_vector clipper_mv __initmv = { ...@@ -642,7 +639,6 @@ struct alpha_machine_vector clipper_mv __initmv = {
DO_EV6_MMU, DO_EV6_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_TSUNAMI_IO, DO_TSUNAMI_IO,
DO_TSUNAMI_BUS,
.machine_check = tsunami_machine_check, .machine_check = tsunami_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
...@@ -671,7 +667,6 @@ struct alpha_machine_vector shark_mv __initmv = { ...@@ -671,7 +667,6 @@ struct alpha_machine_vector shark_mv __initmv = {
DO_EV6_MMU, DO_EV6_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_TSUNAMI_IO, DO_TSUNAMI_IO,
DO_TSUNAMI_BUS,
.machine_check = tsunami_machine_check, .machine_check = tsunami_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
......
...@@ -212,7 +212,6 @@ struct alpha_machine_vector eb64p_mv __initmv = { ...@@ -212,7 +212,6 @@ struct alpha_machine_vector eb64p_mv __initmv = {
DO_EV4_MMU, DO_EV4_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_APECS_IO, DO_APECS_IO,
DO_APECS_BUS,
.machine_check = apecs_machine_check, .machine_check = apecs_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
...@@ -238,7 +237,6 @@ struct alpha_machine_vector eb66_mv __initmv = { ...@@ -238,7 +237,6 @@ struct alpha_machine_vector eb66_mv __initmv = {
DO_EV4_MMU, DO_EV4_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_LCA_IO, DO_LCA_IO,
DO_LCA_BUS,
.machine_check = lca_machine_check, .machine_check = lca_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
......
...@@ -222,7 +222,6 @@ struct alpha_machine_vector eiger_mv __initmv = { ...@@ -222,7 +222,6 @@ struct alpha_machine_vector eiger_mv __initmv = {
DO_EV6_MMU, DO_EV6_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_TSUNAMI_IO, DO_TSUNAMI_IO,
DO_TSUNAMI_BUS,
.machine_check = tsunami_machine_check, .machine_check = tsunami_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
......
...@@ -249,8 +249,6 @@ jensen_machine_check (u64 vector, u64 la, struct pt_regs *regs) ...@@ -249,8 +249,6 @@ jensen_machine_check (u64 vector, u64 la, struct pt_regs *regs)
printk(KERN_CRIT "Machine check\n"); printk(KERN_CRIT "Machine check\n");
} }
#define jensen_pci_tbi ((void*)0)
/* /*
* The System Vector * The System Vector
...@@ -260,7 +258,6 @@ struct alpha_machine_vector jensen_mv __initmv = { ...@@ -260,7 +258,6 @@ struct alpha_machine_vector jensen_mv __initmv = {
.vector_name = "Jensen", .vector_name = "Jensen",
DO_EV4_MMU, DO_EV4_MMU,
IO_LITE(JENSEN,jensen), IO_LITE(JENSEN,jensen),
BUS(jensen),
.machine_check = jensen_machine_check, .machine_check = jensen_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.rtc_port = 0x170, .rtc_port = 0x170,
......
...@@ -471,7 +471,6 @@ struct alpha_machine_vector marvel_ev7_mv __initmv = { ...@@ -471,7 +471,6 @@ struct alpha_machine_vector marvel_ev7_mv __initmv = {
DO_EV7_MMU, DO_EV7_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_MARVEL_IO, DO_MARVEL_IO,
DO_MARVEL_BUS,
.machine_check = marvel_machine_check, .machine_check = marvel_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
......
...@@ -269,7 +269,6 @@ struct alpha_machine_vector miata_mv __initmv = { ...@@ -269,7 +269,6 @@ struct alpha_machine_vector miata_mv __initmv = {
DO_EV5_MMU, DO_EV5_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_PYXIS_IO, DO_PYXIS_IO,
DO_CIA_BUS,
.machine_check = cia_machine_check, .machine_check = cia_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
......
...@@ -221,7 +221,6 @@ struct alpha_machine_vector mikasa_mv __initmv = { ...@@ -221,7 +221,6 @@ struct alpha_machine_vector mikasa_mv __initmv = {
DO_EV4_MMU, DO_EV4_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_APECS_IO, DO_APECS_IO,
DO_APECS_BUS,
.machine_check = mikasa_apecs_machine_check, .machine_check = mikasa_apecs_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
...@@ -246,7 +245,6 @@ struct alpha_machine_vector mikasa_primo_mv __initmv = { ...@@ -246,7 +245,6 @@ struct alpha_machine_vector mikasa_primo_mv __initmv = {
DO_EV5_MMU, DO_EV5_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_CIA_IO, DO_CIA_IO,
DO_CIA_BUS,
.machine_check = cia_machine_check, .machine_check = cia_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
......
...@@ -250,7 +250,6 @@ struct alpha_machine_vector nautilus_mv __initmv = { ...@@ -250,7 +250,6 @@ struct alpha_machine_vector nautilus_mv __initmv = {
DO_EV6_MMU, DO_EV6_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_IRONGATE_IO, DO_IRONGATE_IO,
DO_IRONGATE_BUS,
.machine_check = nautilus_machine_check, .machine_check = nautilus_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
......
...@@ -303,7 +303,6 @@ struct alpha_machine_vector noritake_mv __initmv = { ...@@ -303,7 +303,6 @@ struct alpha_machine_vector noritake_mv __initmv = {
DO_EV4_MMU, DO_EV4_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_APECS_IO, DO_APECS_IO,
DO_APECS_BUS,
.machine_check = noritake_apecs_machine_check, .machine_check = noritake_apecs_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = EISA_DEFAULT_IO_BASE, .min_io_address = EISA_DEFAULT_IO_BASE,
...@@ -328,7 +327,6 @@ struct alpha_machine_vector noritake_primo_mv __initmv = { ...@@ -328,7 +327,6 @@ struct alpha_machine_vector noritake_primo_mv __initmv = {
DO_EV5_MMU, DO_EV5_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_CIA_IO, DO_CIA_IO,
DO_CIA_BUS,
.machine_check = cia_machine_check, .machine_check = cia_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = EISA_DEFAULT_IO_BASE, .min_io_address = EISA_DEFAULT_IO_BASE,
......
...@@ -250,7 +250,6 @@ struct alpha_machine_vector rawhide_mv __initmv = { ...@@ -250,7 +250,6 @@ struct alpha_machine_vector rawhide_mv __initmv = {
DO_EV5_MMU, DO_EV5_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_MCPCIA_IO, DO_MCPCIA_IO,
DO_MCPCIA_BUS,
.machine_check = mcpcia_machine_check, .machine_check = mcpcia_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
......
...@@ -220,7 +220,6 @@ struct alpha_machine_vector ruffian_mv __initmv = { ...@@ -220,7 +220,6 @@ struct alpha_machine_vector ruffian_mv __initmv = {
DO_EV5_MMU, DO_EV5_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_PYXIS_IO, DO_PYXIS_IO,
DO_CIA_BUS,
.machine_check = cia_machine_check, .machine_check = cia_machine_check,
.max_isa_dma_address = ALPHA_RUFFIAN_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_RUFFIAN_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
......
...@@ -201,7 +201,6 @@ struct alpha_machine_vector rx164_mv __initmv = { ...@@ -201,7 +201,6 @@ struct alpha_machine_vector rx164_mv __initmv = {
DO_EV5_MMU, DO_EV5_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_POLARIS_IO, DO_POLARIS_IO,
DO_POLARIS_BUS,
.machine_check = polaris_machine_check, .machine_check = polaris_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
......
...@@ -566,7 +566,6 @@ struct alpha_machine_vector sable_mv __initmv = { ...@@ -566,7 +566,6 @@ struct alpha_machine_vector sable_mv __initmv = {
DO_EV4_MMU, DO_EV4_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_T2_IO, DO_T2_IO,
DO_T2_BUS,
.machine_check = t2_machine_check, .machine_check = t2_machine_check,
.max_isa_dma_address = ALPHA_SABLE_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_SABLE_MAX_ISA_DMA_ADDRESS,
.min_io_address = EISA_DEFAULT_IO_BASE, .min_io_address = EISA_DEFAULT_IO_BASE,
...@@ -599,7 +598,6 @@ struct alpha_machine_vector sable_gamma_mv __initmv = { ...@@ -599,7 +598,6 @@ struct alpha_machine_vector sable_gamma_mv __initmv = {
DO_EV5_MMU, DO_EV5_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_T2_IO, DO_T2_IO,
DO_T2_BUS,
.machine_check = t2_machine_check, .machine_check = t2_machine_check,
.max_isa_dma_address = ALPHA_SABLE_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_SABLE_MAX_ISA_DMA_ADDRESS,
.min_io_address = EISA_DEFAULT_IO_BASE, .min_io_address = EISA_DEFAULT_IO_BASE,
...@@ -631,7 +629,6 @@ struct alpha_machine_vector lynx_mv __initmv = { ...@@ -631,7 +629,6 @@ struct alpha_machine_vector lynx_mv __initmv = {
DO_EV4_MMU, DO_EV4_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_T2_IO, DO_T2_IO,
DO_T2_BUS,
.machine_check = t2_machine_check, .machine_check = t2_machine_check,
.max_isa_dma_address = ALPHA_SABLE_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_SABLE_MAX_ISA_DMA_ADDRESS,
.min_io_address = EISA_DEFAULT_IO_BASE, .min_io_address = EISA_DEFAULT_IO_BASE,
......
...@@ -288,7 +288,6 @@ struct alpha_machine_vector alphabook1_mv __initmv = { ...@@ -288,7 +288,6 @@ struct alpha_machine_vector alphabook1_mv __initmv = {
DO_EV4_MMU, DO_EV4_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_LCA_IO, DO_LCA_IO,
DO_LCA_BUS,
.machine_check = lca_machine_check, .machine_check = lca_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
...@@ -319,7 +318,6 @@ struct alpha_machine_vector avanti_mv __initmv = { ...@@ -319,7 +318,6 @@ struct alpha_machine_vector avanti_mv __initmv = {
DO_EV4_MMU, DO_EV4_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_APECS_IO, DO_APECS_IO,
DO_APECS_BUS,
.machine_check = apecs_machine_check, .machine_check = apecs_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
...@@ -349,7 +347,6 @@ struct alpha_machine_vector noname_mv __initmv = { ...@@ -349,7 +347,6 @@ struct alpha_machine_vector noname_mv __initmv = {
DO_EV4_MMU, DO_EV4_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_LCA_IO, DO_LCA_IO,
DO_LCA_BUS,
.machine_check = lca_machine_check, .machine_check = lca_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
...@@ -388,7 +385,6 @@ struct alpha_machine_vector p2k_mv __initmv = { ...@@ -388,7 +385,6 @@ struct alpha_machine_vector p2k_mv __initmv = {
DO_EV4_MMU, DO_EV4_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_LCA_IO, DO_LCA_IO,
DO_LCA_BUS,
.machine_check = lca_machine_check, .machine_check = lca_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
...@@ -418,7 +414,6 @@ struct alpha_machine_vector xl_mv __initmv = { ...@@ -418,7 +414,6 @@ struct alpha_machine_vector xl_mv __initmv = {
DO_EV4_MMU, DO_EV4_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_APECS_IO, DO_APECS_IO,
BUS(apecs),
.machine_check = apecs_machine_check, .machine_check = apecs_machine_check,
.max_isa_dma_address = ALPHA_XL_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_XL_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
......
...@@ -158,7 +158,6 @@ struct alpha_machine_vector sx164_mv __initmv = { ...@@ -158,7 +158,6 @@ struct alpha_machine_vector sx164_mv __initmv = {
DO_EV5_MMU, DO_EV5_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_PYXIS_IO, DO_PYXIS_IO,
DO_CIA_BUS,
.machine_check = cia_machine_check, .machine_check = cia_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
......
...@@ -277,7 +277,6 @@ struct alpha_machine_vector takara_mv __initmv = { ...@@ -277,7 +277,6 @@ struct alpha_machine_vector takara_mv __initmv = {
DO_EV5_MMU, DO_EV5_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_CIA_IO, DO_CIA_IO,
DO_CIA_BUS,
.machine_check = cia_machine_check, .machine_check = cia_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
......
...@@ -66,7 +66,7 @@ titan_update_irq_hw(unsigned long mask) ...@@ -66,7 +66,7 @@ titan_update_irq_hw(unsigned long mask)
register int bcpu = boot_cpuid; register int bcpu = boot_cpuid;
#ifdef CONFIG_SMP #ifdef CONFIG_SMP
register unsigned long cpm = cpu_present_mask; cpumask_t cpm = cpu_present_mask;
volatile unsigned long *dim0, *dim1, *dim2, *dim3; volatile unsigned long *dim0, *dim1, *dim2, *dim3;
unsigned long mask0, mask1, mask2, mask3, dummy; unsigned long mask0, mask1, mask2, mask3, dummy;
...@@ -85,10 +85,10 @@ titan_update_irq_hw(unsigned long mask) ...@@ -85,10 +85,10 @@ titan_update_irq_hw(unsigned long mask)
dim1 = &cchip->dim1.csr; dim1 = &cchip->dim1.csr;
dim2 = &cchip->dim2.csr; dim2 = &cchip->dim2.csr;
dim3 = &cchip->dim3.csr; dim3 = &cchip->dim3.csr;
if ((cpm & 1) == 0) dim0 = &dummy; if (!cpu_isset(0, cpm)) dim0 = &dummy;
if ((cpm & 2) == 0) dim1 = &dummy; if (!cpu_isset(1, cpm)) dim1 = &dummy;
if ((cpm & 4) == 0) dim2 = &dummy; if (!cpu_isset(2, cpm)) dim2 = &dummy;
if ((cpm & 8) == 0) dim3 = &dummy; if (!cpu_isset(3, cpm)) dim3 = &dummy;
*dim0 = mask0; *dim0 = mask0;
*dim1 = mask1; *dim1 = mask1;
...@@ -369,7 +369,6 @@ struct alpha_machine_vector titan_mv __initmv = { ...@@ -369,7 +369,6 @@ struct alpha_machine_vector titan_mv __initmv = {
DO_EV6_MMU, DO_EV6_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_TITAN_IO, DO_TITAN_IO,
DO_TITAN_BUS,
.machine_check = titan_machine_check, .machine_check = titan_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
...@@ -397,7 +396,6 @@ struct alpha_machine_vector privateer_mv __initmv = { ...@@ -397,7 +396,6 @@ struct alpha_machine_vector privateer_mv __initmv = {
DO_EV6_MMU, DO_EV6_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_TITAN_IO, DO_TITAN_IO,
DO_TITAN_BUS,
.machine_check = privateer_machine_check, .machine_check = privateer_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
......
...@@ -337,7 +337,6 @@ struct alpha_machine_vector wildfire_mv __initmv = { ...@@ -337,7 +337,6 @@ struct alpha_machine_vector wildfire_mv __initmv = {
DO_EV6_MMU, DO_EV6_MMU,
DO_DEFAULT_RTC, DO_DEFAULT_RTC,
DO_WILDFIRE_IO, DO_WILDFIRE_IO,
DO_WILDFIRE_BUS,
.machine_check = wildfire_machine_check, .machine_check = wildfire_machine_check,
.max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS, .max_isa_dma_address = ALPHA_MAX_ISA_DMA_ADDRESS,
.min_io_address = DEFAULT_IO_BASE, .min_io_address = DEFAULT_IO_BASE,
......
...@@ -111,7 +111,7 @@ dik_show_code(unsigned int *pc) ...@@ -111,7 +111,7 @@ dik_show_code(unsigned int *pc)
printk("Code:"); printk("Code:");
for (i = -6; i < 2; i++) { for (i = -6; i < 2; i++) {
unsigned int insn; unsigned int insn;
if (__get_user(insn, pc+i)) if (__get_user(insn, (unsigned int __user *)pc + i))
break; break;
printk("%c%08x%c", i ? ' ' : '<', insn, i ? ' ' : '>'); printk("%c%08x%c", i ? ' ' : '<', insn, i ? ' ' : '>');
} }
......
This diff is collapsed.
...@@ -98,7 +98,7 @@ do_page_fault(unsigned long address, unsigned long mmcsr, ...@@ -98,7 +98,7 @@ do_page_fault(unsigned long address, unsigned long mmcsr,
by ignoring such an instruction. */ by ignoring such an instruction. */
if (cause == 0) { if (cause == 0) {
unsigned int insn; unsigned int insn;
__get_user(insn, (unsigned int *)regs->pc); __get_user(insn, (unsigned int __user *)regs->pc);
if ((insn >> 21 & 0x1f) == 0x1f && if ((insn >> 21 & 0x1f) == 0x1f &&
/* ldq ldl ldt lds ldg ldf ldwu ldbu */ /* ldq ldl ldt lds ldg ldf ldwu ldbu */
(1ul << (insn >> 26) & 0x30f00001400ul)) { (1ul << (insn >> 26) & 0x30f00001400ul)) {
......
...@@ -90,4 +90,14 @@ ...@@ -90,4 +90,14 @@
__asm__("stw %1,%0" : "=m"(mem) : "r"(val)) __asm__("stw %1,%0" : "=m"(mem) : "r"(val))
#endif #endif
/* Some idiots over in <linux/compiler.h> thought inline should imply
always_inline. This breaks stuff. We'll include this file whenever
we run into such problems. */
#include <linux/compiler.h>
#undef inline
#undef __inline__
#undef __inline
#endif /* __ALPHA_COMPILER_H */ #endif /* __ALPHA_COMPILER_H */
...@@ -370,178 +370,142 @@ struct el_apecs_procdata ...@@ -370,178 +370,142 @@ struct el_apecs_procdata
* data to/from the right byte-lanes. * data to/from the right byte-lanes.
*/ */
#define vip volatile int * #define vip volatile int __force *
#define vuip volatile unsigned int * #define vuip volatile unsigned int __force *
#define vulp volatile unsigned long * #define vulp volatile unsigned long __force *
__EXTERN_INLINE u8 apecs_inb(unsigned long addr) #define APECS_SET_HAE \
do { \
if (addr >= (1UL << 24)) { \
unsigned long msb = addr & 0xf8000000; \
addr -= msb; \
set_hae(msb); \
} \
} while (0)
__EXTERN_INLINE unsigned int apecs_ioread8(void __iomem *xaddr)
{ {
long result = *(vip) ((addr << 5) + APECS_IO + 0x00); unsigned long addr = (unsigned long) xaddr;
unsigned long result, base_and_type;
if (addr >= APECS_DENSE_MEM) {
addr -= APECS_DENSE_MEM;
APECS_SET_HAE;
base_and_type = APECS_SPARSE_MEM + 0x00;
} else {
addr -= APECS_IO;
base_and_type = APECS_IO + 0x00;
}
result = *(vip) ((addr << 5) + base_and_type);
return __kernel_extbl(result, addr & 3); return __kernel_extbl(result, addr & 3);
} }
__EXTERN_INLINE void apecs_outb(u8 b, unsigned long addr) __EXTERN_INLINE void apecs_iowrite8(u8 b, void __iomem *xaddr)
{ {
unsigned long w; unsigned long addr = (unsigned long) xaddr;
unsigned long w, base_and_type;
if (addr >= APECS_DENSE_MEM) {
addr -= APECS_DENSE_MEM;
APECS_SET_HAE;
base_and_type = APECS_SPARSE_MEM + 0x00;
} else {
addr -= APECS_IO;
base_and_type = APECS_IO + 0x00;
}
w = __kernel_insbl(b, addr & 3); w = __kernel_insbl(b, addr & 3);
*(vuip) ((addr << 5) + APECS_IO + 0x00) = w; *(vuip) ((addr << 5) + base_and_type) = w;
mb();
}
__EXTERN_INLINE u16 apecs_inw(unsigned long addr)
{
long result = *(vip) ((addr << 5) + APECS_IO + 0x08);
return __kernel_extwl(result, addr & 3);
} }
__EXTERN_INLINE void apecs_outw(u16 b, unsigned long addr) __EXTERN_INLINE unsigned int apecs_ioread16(void __iomem *xaddr)
{
unsigned long w;
w = __kernel_inswl(b, addr & 3);
*(vuip) ((addr << 5) + APECS_IO + 0x08) = w;
mb();
}
__EXTERN_INLINE u32 apecs_inl(unsigned long addr)
{
return *(vuip) ((addr << 5) + APECS_IO + 0x18);
}
__EXTERN_INLINE void apecs_outl(u32 b, unsigned long addr)
{
*(vuip) ((addr << 5) + APECS_IO + 0x18) = b;
mb();
}
/*
* Memory functions. 64-bit and 32-bit accesses are done through
* dense memory space, everything else through sparse space.
*/
__EXTERN_INLINE u8 apecs_readb(unsigned long addr)
{ {
unsigned long result, msb; unsigned long addr = (unsigned long) xaddr;
unsigned long result, base_and_type;
addr -= APECS_DENSE_MEM;
if (addr >= (1UL << 24)) { if (addr >= APECS_DENSE_MEM) {
msb = addr & 0xf8000000; addr -= APECS_DENSE_MEM;
addr -= msb; APECS_SET_HAE;
set_hae(msb); base_and_type = APECS_SPARSE_MEM + 0x08;
} else {
addr -= APECS_IO;
base_and_type = APECS_IO + 0x08;
} }
result = *(vip) ((addr << 5) + APECS_SPARSE_MEM + 0x00);
return __kernel_extbl(result, addr & 3);
}
__EXTERN_INLINE u16 apecs_readw(unsigned long addr) result = *(vip) ((addr << 5) + base_and_type);
{
unsigned long result, msb;
addr -= APECS_DENSE_MEM;
if (addr >= (1UL << 24)) {
msb = addr & 0xf8000000;
addr -= msb;
set_hae(msb);
}
result = *(vip) ((addr << 5) + APECS_SPARSE_MEM + 0x08);
return __kernel_extwl(result, addr & 3); return __kernel_extwl(result, addr & 3);
} }
__EXTERN_INLINE u32 apecs_readl(unsigned long addr) __EXTERN_INLINE void apecs_iowrite16(u16 b, void __iomem *xaddr)
{
return (*(vuip)addr) & 0xffffffff;
}
__EXTERN_INLINE u64 apecs_readq(unsigned long addr)
{
return *(vulp)addr;
}
__EXTERN_INLINE void apecs_writeb(u8 b, unsigned long addr)
{ {
unsigned long msb; unsigned long addr = (unsigned long) xaddr;
unsigned long w, base_and_type;
addr -= APECS_DENSE_MEM;
if (addr >= (1UL << 24)) { if (addr >= APECS_DENSE_MEM) {
msb = addr & 0xf8000000; addr -= APECS_DENSE_MEM;
addr -= msb; APECS_SET_HAE;
set_hae(msb); base_and_type = APECS_SPARSE_MEM + 0x08;
} else {
addr -= APECS_IO;
base_and_type = APECS_IO + 0x08;
} }
*(vuip) ((addr << 5) + APECS_SPARSE_MEM + 0x00) = b * 0x01010101;
w = __kernel_inswl(b, addr & 3);
*(vuip) ((addr << 5) + base_and_type) = w;
} }
__EXTERN_INLINE void apecs_writew(u16 b, unsigned long addr) __EXTERN_INLINE unsigned int apecs_ioread32(void __iomem *xaddr)
{ {
unsigned long msb; unsigned long addr = (unsigned long) xaddr;
if (addr < APECS_DENSE_MEM)
addr -= APECS_DENSE_MEM; addr = ((addr - APECS_IO) << 5) + APECS_IO + 0x18;
if (addr >= (1UL << 24)) { return *(vuip)addr;
msb = addr & 0xf8000000;
addr -= msb;
set_hae(msb);
}
*(vuip) ((addr << 5) + APECS_SPARSE_MEM + 0x08) = b * 0x00010001;
} }
__EXTERN_INLINE void apecs_writel(u32 b, unsigned long addr) __EXTERN_INLINE void apecs_iowrite32(u32 b, void __iomem *xaddr)
{ {
unsigned long addr = (unsigned long) xaddr;
if (addr < APECS_DENSE_MEM)
addr = ((addr - APECS_IO) << 5) + APECS_IO + 0x18;
*(vuip)addr = b; *(vuip)addr = b;
} }
__EXTERN_INLINE void apecs_writeq(u64 b, unsigned long addr) __EXTERN_INLINE void __iomem *apecs_ioportmap(unsigned long addr)
{ {
*(vulp)addr = b; return (void __iomem *)(addr + APECS_IO);
} }
__EXTERN_INLINE unsigned long apecs_ioremap(unsigned long addr, __EXTERN_INLINE void __iomem *apecs_ioremap(unsigned long addr,
unsigned long size unsigned long size)
__attribute__((unused)))
{ {
return addr + APECS_DENSE_MEM; return (void __iomem *)(addr + APECS_DENSE_MEM);
} }
__EXTERN_INLINE void apecs_iounmap(unsigned long addr) __EXTERN_INLINE int apecs_is_ioaddr(unsigned long addr)
{ {
return; return addr >= IDENT_ADDR + 0x180000000UL;
} }
__EXTERN_INLINE int apecs_is_ioaddr(unsigned long addr) __EXTERN_INLINE int apecs_is_mmio(const volatile void __iomem *addr)
{ {
return addr >= IDENT_ADDR + 0x180000000UL; return (unsigned long)addr >= APECS_DENSE_MEM;
} }
#undef APECS_SET_HAE
#undef vip #undef vip
#undef vuip #undef vuip
#undef vulp #undef vulp
#ifdef __WANT_IO_DEF #undef __IO_PREFIX
#define __IO_PREFIX apecs
#define __inb(p) apecs_inb((unsigned long)(p)) #define apecs_trivial_io_bw 0
#define __inw(p) apecs_inw((unsigned long)(p)) #define apecs_trivial_io_lq 0
#define __inl(p) apecs_inl((unsigned long)(p)) #define apecs_trivial_rw_bw 2
#define __outb(x,p) apecs_outb((x),(unsigned long)(p)) #define apecs_trivial_rw_lq 1
#define __outw(x,p) apecs_outw((x),(unsigned long)(p)) #define apecs_trivial_iounmap 1
#define __outl(x,p) apecs_outl((x),(unsigned long)(p)) #include <asm/io_trivial.h>
#define __readb(a) apecs_readb((unsigned long)(a))
#define __readw(a) apecs_readw((unsigned long)(a))
#define __readl(a) apecs_readl((unsigned long)(a))
#define __readq(a) apecs_readq((unsigned long)(a))
#define __writeb(x,a) apecs_writeb((x),(unsigned long)(a))
#define __writew(x,a) apecs_writew((x),(unsigned long)(a))
#define __writel(x,a) apecs_writel((x),(unsigned long)(a))
#define __writeq(x,a) apecs_writeq((x),(unsigned long)(a))
#define __ioremap(a,s) apecs_ioremap((unsigned long)(a),(s))
#define __iounmap(a) apecs_iounmap((unsigned long)(a))
#define __is_ioaddr(a) apecs_is_ioaddr((unsigned long)(a))
#define __raw_readl(a) __readl(a)
#define __raw_readq(a) __readq(a)
#define __raw_writel(v,a) __writel((v),(a))
#define __raw_writeq(v,a) __writeq((v),(a))
#endif /* __WANT_IO_DEF */
#ifdef __IO_EXTERN_INLINE #ifdef __IO_EXTERN_INLINE
#undef __EXTERN_INLINE #undef __EXTERN_INLINE
......
This diff is collapsed.
...@@ -190,137 +190,37 @@ struct el_IRONGATE_sysdata_mcheck { ...@@ -190,137 +190,37 @@ struct el_IRONGATE_sysdata_mcheck {
* K7 can only use linear accesses to get at PCI memory and I/O spaces. * K7 can only use linear accesses to get at PCI memory and I/O spaces.
*/ */
#define vucp volatile unsigned char *
#define vusp volatile unsigned short *
#define vuip volatile unsigned int *
#define vulp volatile unsigned long *
__EXTERN_INLINE u8 irongate_inb(unsigned long addr)
{
return __kernel_ldbu(*(vucp)(addr + IRONGATE_IO));
}
__EXTERN_INLINE void irongate_outb(u8 b, unsigned long addr)
{
__kernel_stb(b, *(vucp)(addr + IRONGATE_IO));
mb();
}
__EXTERN_INLINE u16 irongate_inw(unsigned long addr)
{
return __kernel_ldwu(*(vusp)(addr + IRONGATE_IO));
}
__EXTERN_INLINE void irongate_outw(u16 b, unsigned long addr)
{
__kernel_stw(b, *(vusp)(addr + IRONGATE_IO));
mb();
}
__EXTERN_INLINE u32 irongate_inl(unsigned long addr)
{
return *(vuip)(addr + IRONGATE_IO);
}
__EXTERN_INLINE void irongate_outl(u32 b, unsigned long addr)
{
*(vuip)(addr + IRONGATE_IO) = b;
mb();
}
/* /*
* Memory functions. All accesses are done through linear space. * Memory functions. All accesses are done through linear space.
*/ */
__EXTERN_INLINE u8 irongate_readb(unsigned long addr) __EXTERN_INLINE void __iomem *irongate_ioportmap(unsigned long addr)
{ {
return __kernel_ldbu(*(vucp)addr); return (void __iomem *)(addr + IRONGATE_IO);
} }
__EXTERN_INLINE u16 irongate_readw(unsigned long addr) extern void __iomem *irongate_ioremap(unsigned long addr, unsigned long size);
{ extern void irongate_iounmap(volatile void __iomem *addr);
return __kernel_ldwu(*(vusp)addr);
}
__EXTERN_INLINE u32 irongate_readl(unsigned long addr) __EXTERN_INLINE int irongate_is_ioaddr(unsigned long addr)
{
return (*(vuip)addr) & 0xffffffff;
}
__EXTERN_INLINE u64 irongate_readq(unsigned long addr)
{
return *(vulp)addr;
}
__EXTERN_INLINE void irongate_writeb(u8 b, unsigned long addr)
{
__kernel_stb(b, *(vucp)addr);
}
__EXTERN_INLINE void irongate_writew(u16 b, unsigned long addr)
{
__kernel_stw(b, *(vusp)addr);
}
__EXTERN_INLINE void irongate_writel(u32 b, unsigned long addr)
{
*(vuip)addr = b;
}
__EXTERN_INLINE void irongate_writeq(u64 b, unsigned long addr)
{ {
*(vulp)addr = b; return addr >= IRONGATE_MEM;
} }
extern unsigned long irongate_ioremap(unsigned long addr, unsigned long size); __EXTERN_INLINE int irongate_is_mmio(const volatile void __iomem *xaddr)
extern void irongate_iounmap(unsigned long addr);
__EXTERN_INLINE int irongate_is_ioaddr(unsigned long addr)
{ {
return addr >= IRONGATE_MEM; unsigned long addr = (unsigned long)xaddr;
return addr < IRONGATE_IO || addr >= IRONGATE_CONF;
} }
#undef vucp #undef __IO_PREFIX
#undef vusp #define __IO_PREFIX irongate
#undef vuip #define irongate_trivial_rw_bw 1
#undef vulp #define irongate_trivial_rw_lq 1
#define irongate_trivial_io_bw 1
#ifdef __WANT_IO_DEF #define irongate_trivial_io_lq 1
#define irongate_trivial_iounmap 0
#define __inb(p) irongate_inb((unsigned long)(p)) #include <asm/io_trivial.h>
#define __inw(p) irongate_inw((unsigned long)(p))
#define __inl(p) irongate_inl((unsigned long)(p))
#define __outb(x,p) irongate_outb((x),(unsigned long)(p))
#define __outw(x,p) irongate_outw((x),(unsigned long)(p))
#define __outl(x,p) irongate_outl((x),(unsigned long)(p))
#define __readb(a) irongate_readb((unsigned long)(a))
#define __readw(a) irongate_readw((unsigned long)(a))
#define __readl(a) irongate_readl((unsigned long)(a))
#define __readq(a) irongate_readq((unsigned long)(a))
#define __writeb(x,a) irongate_writeb((x),(unsigned long)(a))
#define __writew(x,a) irongate_writew((x),(unsigned long)(a))
#define __writel(x,a) irongate_writel((x),(unsigned long)(a))
#define __writeq(x,a) irongate_writeq((x),(unsigned long)(a))
#define __ioremap(a,s) irongate_ioremap((unsigned long)(a),(s))
#define __iounmap(a) irongate_iounmap((unsigned long)(a))
#define __is_ioaddr(a) irongate_is_ioaddr((unsigned long)(a))
#define inb(p) __inb(p)
#define inw(p) __inw(p)
#define inl(p) __inl(p)
#define outb(x,p) __outb((x),(p))
#define outw(x,p) __outw((x),(p))
#define outl(x,p) __outl((x),(p))
#define __raw_readb(a) __readb(a)
#define __raw_readw(a) __readw(a)
#define __raw_readl(a) __readl(a)
#define __raw_readq(a) __readq(a)
#define __raw_writeb(v,a) __writeb((v),(a))
#define __raw_writew(v,a) __writew((v),(a))
#define __raw_writel(v,a) __writel((v),(a))
#define __raw_writeq(v,a) __writeq((v),(a))
#endif /* __WANT_IO_DEF */
#ifdef __IO_EXTERN_INLINE #ifdef __IO_EXTERN_INLINE
#undef __EXTERN_INLINE #undef __EXTERN_INLINE
......
...@@ -215,145 +215,117 @@ union el_lca { ...@@ -215,145 +215,117 @@ union el_lca {
* data to/from the right byte-lanes. * data to/from the right byte-lanes.
*/ */
#define vip volatile int * #define vip volatile int __force *
#define vuip volatile unsigned int * #define vuip volatile unsigned int __force *
#define vulp volatile unsigned long * #define vulp volatile unsigned long __force *
__EXTERN_INLINE u8 lca_inb(unsigned long addr) #define LCA_SET_HAE \
{ do { \
long result = *(vip) ((addr << 5) + LCA_IO + 0x00); if (addr >= (1UL << 24)) { \
return __kernel_extbl(result, addr & 3); unsigned long msb = addr & 0xf8000000; \
} addr -= msb; \
set_hae(msb); \
} \
} while (0)
__EXTERN_INLINE void lca_outb(u8 b, unsigned long addr)
{
unsigned long w;
w = __kernel_insbl(b, addr & 3);
*(vuip) ((addr << 5) + LCA_IO + 0x00) = w;
mb();
}
__EXTERN_INLINE u16 lca_inw(unsigned long addr)
{
long result = *(vip) ((addr << 5) + LCA_IO + 0x08);
return __kernel_extwl(result, addr & 3);
}
__EXTERN_INLINE void lca_outw(u16 b, unsigned long addr) __EXTERN_INLINE unsigned int lca_ioread8(void __iomem *xaddr)
{ {
unsigned long w; unsigned long addr = (unsigned long) xaddr;
unsigned long result, base_and_type;
if (addr >= LCA_DENSE_MEM) {
addr -= LCA_DENSE_MEM;
LCA_SET_HAE;
base_and_type = LCA_SPARSE_MEM + 0x00;
} else {
addr -= LCA_IO;
base_and_type = LCA_IO + 0x00;
}
w = __kernel_inswl(b, addr & 3); result = *(vip) ((addr << 5) + base_and_type);
*(vuip) ((addr << 5) + LCA_IO + 0x08) = w; return __kernel_extbl(result, addr & 3);
mb();
} }
__EXTERN_INLINE u32 lca_inl(unsigned long addr) __EXTERN_INLINE void lca_iowrite8(u8 b, void __iomem *xaddr)
{ {
return *(vuip) ((addr << 5) + LCA_IO + 0x18); unsigned long addr = (unsigned long) xaddr;
} unsigned long w, base_and_type;
if (addr >= LCA_DENSE_MEM) {
addr -= LCA_DENSE_MEM;
LCA_SET_HAE;
base_and_type = LCA_SPARSE_MEM + 0x00;
} else {
addr -= LCA_IO;
base_and_type = LCA_IO + 0x00;
}
__EXTERN_INLINE void lca_outl(u32 b, unsigned long addr) w = __kernel_insbl(b, addr & 3);
{ *(vuip) ((addr << 5) + base_and_type) = w;
*(vuip) ((addr << 5) + LCA_IO + 0x18) = b;
mb();
} }
__EXTERN_INLINE unsigned int lca_ioread16(void __iomem *xaddr)
/*
* Memory functions. 64-bit and 32-bit accesses are done through
* dense memory space, everything else through sparse space.
*/
__EXTERN_INLINE u8 lca_readb(unsigned long addr)
{ {
unsigned long result, msb; unsigned long addr = (unsigned long) xaddr;
unsigned long result, base_and_type;
addr -= LCA_DENSE_MEM;
if (addr >= (1UL << 24)) { if (addr >= LCA_DENSE_MEM) {
msb = addr & 0xf8000000; addr -= LCA_DENSE_MEM;
addr -= msb; LCA_SET_HAE;
set_hae(msb); base_and_type = LCA_SPARSE_MEM + 0x08;
} else {
addr -= LCA_IO;
base_and_type = LCA_IO + 0x08;
} }
result = *(vip) ((addr << 5) + LCA_SPARSE_MEM + 0x00);
return __kernel_extbl(result, addr & 3);
}
__EXTERN_INLINE u16 lca_readw(unsigned long addr) result = *(vip) ((addr << 5) + base_and_type);
{
unsigned long result, msb;
addr -= LCA_DENSE_MEM;
if (addr >= (1UL << 24)) {
msb = addr & 0xf8000000;
addr -= msb;
set_hae(msb);
}
result = *(vip) ((addr << 5) + LCA_SPARSE_MEM + 0x08);
return __kernel_extwl(result, addr & 3); return __kernel_extwl(result, addr & 3);
} }
__EXTERN_INLINE u32 lca_readl(unsigned long addr) __EXTERN_INLINE void lca_iowrite16(u16 b, void __iomem *xaddr)
{ {
return (*(vuip)addr) & 0xffffffff; unsigned long addr = (unsigned long) xaddr;
} unsigned long w, base_and_type;
__EXTERN_INLINE u64 lca_readq(unsigned long addr) if (addr >= LCA_DENSE_MEM) {
{ addr -= LCA_DENSE_MEM;
return *(vulp)addr; LCA_SET_HAE;
} base_and_type = LCA_SPARSE_MEM + 0x08;
} else {
__EXTERN_INLINE void lca_writeb(u8 b, unsigned long addr) addr -= LCA_IO;
{ base_and_type = LCA_IO + 0x08;
unsigned long msb;
unsigned long w;
addr -= LCA_DENSE_MEM;
if (addr >= (1UL << 24)) {
msb = addr & 0xf8000000;
addr -= msb;
set_hae(msb);
} }
w = __kernel_insbl(b, addr & 3);
*(vuip) ((addr << 5) + LCA_SPARSE_MEM + 0x00) = w;
}
__EXTERN_INLINE void lca_writew(u16 b, unsigned long addr)
{
unsigned long msb;
unsigned long w;
addr -= LCA_DENSE_MEM;
if (addr >= (1UL << 24)) {
msb = addr & 0xf8000000;
addr -= msb;
set_hae(msb);
}
w = __kernel_inswl(b, addr & 3); w = __kernel_inswl(b, addr & 3);
*(vuip) ((addr << 5) + LCA_SPARSE_MEM + 0x08) = w; *(vuip) ((addr << 5) + base_and_type) = w;
} }
__EXTERN_INLINE void lca_writel(u32 b, unsigned long addr) __EXTERN_INLINE unsigned int lca_ioread32(void __iomem *xaddr)
{ {
*(vuip)addr = b; unsigned long addr = (unsigned long) xaddr;
if (addr < LCA_DENSE_MEM)
addr = ((addr - LCA_IO) << 5) + LCA_IO + 0x18;
return *(vuip)addr;
} }
__EXTERN_INLINE void lca_writeq(u64 b, unsigned long addr) __EXTERN_INLINE void lca_iowrite32(u32 b, void __iomem *xaddr)
{ {
*(vulp)addr = b; unsigned long addr = (unsigned long) xaddr;
if (addr < LCA_DENSE_MEM)
addr = ((addr - LCA_IO) << 5) + LCA_IO + 0x18;
*(vuip)addr = b;
} }
__EXTERN_INLINE unsigned long lca_ioremap(unsigned long addr, __EXTERN_INLINE void __iomem *lca_ioportmap(unsigned long addr)
unsigned long size
__attribute__((unused)))
{ {
return addr + LCA_DENSE_MEM; return (void __iomem *)(addr + LCA_IO);
} }
__EXTERN_INLINE void lca_iounmap(unsigned long addr) __EXTERN_INLINE void __iomem *lca_ioremap(unsigned long addr,
unsigned long size)
{ {
return; return (void __iomem *)(addr + LCA_DENSE_MEM);
} }
__EXTERN_INLINE int lca_is_ioaddr(unsigned long addr) __EXTERN_INLINE int lca_is_ioaddr(unsigned long addr)
...@@ -361,36 +333,23 @@ __EXTERN_INLINE int lca_is_ioaddr(unsigned long addr) ...@@ -361,36 +333,23 @@ __EXTERN_INLINE int lca_is_ioaddr(unsigned long addr)
return addr >= IDENT_ADDR + 0x120000000UL; return addr >= IDENT_ADDR + 0x120000000UL;
} }
__EXTERN_INLINE int lca_is_mmio(const volatile void __iomem *addr)
{
return (unsigned long)addr >= LCA_DENSE_MEM;
}
#undef vip #undef vip
#undef vuip #undef vuip
#undef vulp #undef vulp
#ifdef __WANT_IO_DEF #undef __IO_PREFIX
#define __IO_PREFIX lca
#define __inb(p) lca_inb((unsigned long)(p)) #define lca_trivial_rw_bw 2
#define __inw(p) lca_inw((unsigned long)(p)) #define lca_trivial_rw_lq 1
#define __inl(p) lca_inl((unsigned long)(p)) #define lca_trivial_io_bw 0
#define __outb(x,p) lca_outb((x),(unsigned long)(p)) #define lca_trivial_io_lq 0
#define __outw(x,p) lca_outw((x),(unsigned long)(p)) #define lca_trivial_iounmap 1
#define __outl(x,p) lca_outl((x),(unsigned long)(p)) #include <asm/io_trivial.h>
#define __readb(a) lca_readb((unsigned long)(a))
#define __readw(a) lca_readw((unsigned long)(a))
#define __readl(a) lca_readl((unsigned long)(a))
#define __readq(a) lca_readq((unsigned long)(a))
#define __writeb(x,a) lca_writeb((x),(unsigned long)(a))
#define __writew(x,a) lca_writew((x),(unsigned long)(a))
#define __writel(x,a) lca_writel((x),(unsigned long)(a))
#define __writeq(x,a) lca_writeq((x),(unsigned long)(a))
#define __ioremap(a,s) lca_ioremap((unsigned long)(a),(s))
#define __iounmap(a) lca_iounmap((unsigned long)(a))
#define __is_ioaddr(a) lca_is_ioaddr((unsigned long)(a))
#define __raw_readl(a) __readl(a)
#define __raw_readq(a) __readq(a)
#define __raw_writel(v,a) __writel((v),(a))
#define __raw_writeq(v,a) __writeq((v),(a))
#endif /* __WANT_IO_DEF */
#ifdef __IO_EXTERN_INLINE #ifdef __IO_EXTERN_INLINE
#undef __EXTERN_INLINE #undef __EXTERN_INLINE
......
...@@ -325,249 +325,48 @@ struct io7 { ...@@ -325,249 +325,48 @@ struct io7 {
* I/O functions. All access through linear space. * I/O functions. All access through linear space.
*/ */
#define vucp volatile unsigned char *
#define vusp volatile unsigned short *
#define vuip volatile unsigned int *
#define vulp volatile unsigned long *
#ifdef CONFIG_VGA_HOSE
extern struct pci_controller *pci_vga_hose;
# define __marvel_is_port_vga(a) \
(((a) >= 0x3b0) && ((a) < 0x3e0) && ((a) != 0x3b3) && ((a) != 0x3d3))
# define __marvel_is_mem_vga(a) (((a) >= 0xa0000) && ((a) <= 0xc0000))
# define FIXUP_IOADDR_VGA(a) do { \
if (pci_vga_hose && __marvel_is_port_vga(a)) \
a += pci_vga_hose->io_space->start; \
} while(0)
#else
# define FIXUP_IOADDR_VGA(a)
#endif
#define __marvel_is_port_kbd(a) (((a) == 0x60) || ((a) == 0x64))
#define __marvel_is_port_rtc(a) (((a) == 0x70) || ((a) == 0x71))
#define FIXUP_IOADDR_LEGACY(a)
#define FIXUP_IOADDR(a) do { \
FIXUP_IOADDR_VGA(a); \
FIXUP_IOADDR_LEGACY(a); \
} while(0)
#if 0
# define IOBUG(x) printk x
# define IOBUG_FILTER_IOADDR(a, x) \
if (!__marvel_is_port_kbd(a) && !__marvel_is_port_rtc(a)) IOBUG(x)
#else
# define IOBUG(x)
# define IOBUG_FILTER_IOADDR(a, x)
#endif
extern u8 __marvel_rtc_io(int write, u8 b, unsigned long addr);
#define __marvel_rtc_inb(a) __marvel_rtc_io(0, 0, (a))
#define __marvel_rtc_outb(b, a) __marvel_rtc_io(1, (b), (a))
__EXTERN_INLINE int marvel_is_ioaddr(unsigned long addr)
{
return (addr & (1UL << 40)) != 0; /*FIXME - hardwire*/
}
__EXTERN_INLINE u8 marvel_inb(unsigned long addr)
{
FIXUP_IOADDR(addr);
if (!marvel_is_ioaddr(addr)) {
if (__marvel_is_port_kbd(addr))
return (u8)0;
if (__marvel_is_port_rtc(addr))
return __marvel_rtc_inb(addr);
IOBUG_FILTER_IOADDR(addr,
("Bad IO addr %lx - reading -1\n", addr));
return (u8)-1;
}
return __kernel_ldbu(*(vucp)addr);
}
__EXTERN_INLINE void marvel_outb(u8 b, unsigned long addr)
{
FIXUP_IOADDR(addr);
if (!marvel_is_ioaddr(addr)) {
if (__marvel_is_port_rtc(addr))
return (void)__marvel_rtc_outb(b, addr);
IOBUG_FILTER_IOADDR(addr,
("Bad IO addr %lx - reading -1\n", addr));
return;
}
__kernel_stb(b, *(vucp)addr);
mb();
}
__EXTERN_INLINE u16 marvel_inw(unsigned long addr)
{
FIXUP_IOADDR(addr);
if (!marvel_is_ioaddr(addr)) {
IOBUG_FILTER_IOADDR(addr,
("Bad IO addr %lx - reading -1\n", addr));
return (u16)-1;
}
return __kernel_ldwu(*(vusp)addr);
}
__EXTERN_INLINE void marvel_outw(u16 w, unsigned long addr)
{
FIXUP_IOADDR(addr);
if (!marvel_is_ioaddr(addr)) {
IOBUG_FILTER_IOADDR(addr,
("Bad IO addr %lx - reading -1\n", addr));
return;
}
__kernel_stw(w, *(vusp)addr);
mb();
}
__EXTERN_INLINE u32 marvel_inl(unsigned long addr)
{
FIXUP_IOADDR(addr);
if (!marvel_is_ioaddr(addr)) {
IOBUG_FILTER_IOADDR(addr,
("Bad IO addr %lx - reading -1\n", addr));
return (u32)-1;
}
return *(vuip)addr;
}
__EXTERN_INLINE void marvel_outl(u32 l, unsigned long addr)
{
FIXUP_IOADDR(addr);
if (!marvel_is_ioaddr(addr)) {
IOBUG_FILTER_IOADDR(addr,
("Bad IO addr %lx - reading -1\n", addr));
return;
}
*(vuip)addr = l;
mb();
}
/* /*
* Memory functions. All accesses through linear space. * Memory functions. All accesses through linear space.
*/ */
extern unsigned long marvel_ioremap(unsigned long addr, unsigned long size); #define vucp volatile unsigned char __force *
extern void marvel_iounmap(unsigned long addr); #define vusp volatile unsigned short __force *
__EXTERN_INLINE u8 marvel_readb(unsigned long addr) extern unsigned int marvel_ioread8(void __iomem *);
{ extern void marvel_iowrite8(u8 b, void __iomem *);
if (!marvel_is_ioaddr(addr)) {
IOBUG(("Bad MEM addr %lx - reading -1\n", addr));
return (u8)-1;
}
return __kernel_ldbu(*(vucp)addr);
}
__EXTERN_INLINE u16 marvel_readw(unsigned long addr) __EXTERN_INLINE unsigned int marvel_ioread16(void __iomem *addr)
{ {
if (!marvel_is_ioaddr(addr)) {
IOBUG(("Bad MEM addr %lx - reading -1\n", addr));
return (u16)-1;
}
return __kernel_ldwu(*(vusp)addr); return __kernel_ldwu(*(vusp)addr);
} }
__EXTERN_INLINE u32 marvel_readl(unsigned long addr) __EXTERN_INLINE void marvel_iowrite16(u16 b, void __iomem *addr)
{
if (!marvel_is_ioaddr(addr)) {
IOBUG(("Bad MEM addr %lx - reading -1\n", addr));
return (u32)-1;
}
return *(vuip)addr;
}
__EXTERN_INLINE u64 marvel_readq(unsigned long addr)
{ {
if (!marvel_is_ioaddr(addr)) { __kernel_stw(b, *(vusp)addr);
IOBUG(("Bad MEM addr %lx - reading -1\n", addr));
return (u64)-1;
}
return *(vulp)addr;
} }
__EXTERN_INLINE void marvel_writeb(u8 b, unsigned long addr) extern void __iomem *marvel_ioremap(unsigned long addr, unsigned long size);
{ extern void marvel_iounmap(volatile void __iomem *addr);
if (!marvel_is_ioaddr(addr)) { extern void __iomem *marvel_ioportmap (unsigned long addr);
IOBUG(("Bad MEM addr %lx - dropping store\n", addr));
return;
}
__kernel_stb(b, *(vucp)addr);
}
__EXTERN_INLINE void marvel_writew(u16 w, unsigned long addr)
{
if (!marvel_is_ioaddr(addr)) {
IOBUG(("Bad MEM addr %lx - dropping store\n", addr));
return;
}
__kernel_stw(w, *(vusp)addr);
}
__EXTERN_INLINE void marvel_writel(u32 l, unsigned long addr) __EXTERN_INLINE int marvel_is_ioaddr(unsigned long addr)
{
if (!marvel_is_ioaddr(addr)) {
IOBUG(("Bad MEM addr %lx - dropping store\n", addr));
return;
}
*(vuip)addr = l;
}
__EXTERN_INLINE void marvel_writeq(u64 q, unsigned long addr)
{ {
if (!marvel_is_ioaddr(addr)) { return (addr >> 40) & 1;
IOBUG(("Bad MEM addr %lx - dropping store\n", addr));
return;
}
*(vulp)addr = q;
} }
#undef FIXUP_IOADDR extern int marvel_is_mmio(const volatile void __iomem *);
#undef FIXUP_IOADDR_LEGACY
#undef FIXUP_IOADDR_VGA
#undef vucp #undef vucp
#undef vusp #undef vusp
#undef vuip
#undef vulp
#ifdef __WANT_IO_DEF
#define __inb(p) marvel_inb((unsigned long)(p))
#define __inw(p) marvel_inw((unsigned long)(p))
#define __inl(p) marvel_inl((unsigned long)(p))
#define __outb(x,p) marvel_outb((x),(unsigned long)(p))
#define __outw(x,p) marvel_outw((x),(unsigned long)(p))
#define __outl(x,p) marvel_outl((x),(unsigned long)(p))
#define __readb(a) marvel_readb((unsigned long)(a))
#define __readw(a) marvel_readw((unsigned long)(a))
#define __readl(a) marvel_readl((unsigned long)(a))
#define __readq(a) marvel_readq((unsigned long)(a))
#define __writeb(x,a) marvel_writeb((x),(unsigned long)(a))
#define __writew(x,a) marvel_writew((x),(unsigned long)(a))
#define __writel(x,a) marvel_writel((x),(unsigned long)(a))
#define __writeq(x,a) marvel_writeq((x),(unsigned long)(a))
#define __ioremap(a,s) marvel_ioremap((unsigned long)(a),(s))
#define __iounmap(a) marvel_iounmap((unsigned long)(a))
#define __is_ioaddr(a) marvel_is_ioaddr((unsigned long)(a))
/* Disable direct inlining of these calls with the debug checks present. */
#if 0
#define __raw_readb(a) __readb(a)
#define __raw_readw(a) __readw(a)
#define __raw_readl(a) __readl(a)
#define __raw_readq(a) __readq(a)
#define __raw_writeb(v,a) __writeb(v,a)
#define __raw_writew(v,a) __writew(v,a)
#define __raw_writel(v,a) __writel(v,a)
#define __raw_writeq(v,a) __writeq(v,a)
#endif
#endif /* __WANT_IO_DEF */ #undef __IO_PREFIX
#define __IO_PREFIX marvel
#define marvel_trivial_rw_bw 1
#define marvel_trivial_rw_lq 1
#define marvel_trivial_io_bw 0
#define marvel_trivial_io_lq 1
#define marvel_trivial_iounmap 0
#include <asm/io_trivial.h>
#ifdef __IO_EXTERN_INLINE #ifdef __IO_EXTERN_INLINE
# undef __EXTERN_INLINE # undef __EXTERN_INLINE
......
...@@ -211,91 +211,6 @@ struct el_MCPCIA_uncorrected_frame_mcheck { ...@@ -211,91 +211,6 @@ struct el_MCPCIA_uncorrected_frame_mcheck {
* Unfortunately, we can't use BWIO with EV5, so for now, we always use SPARSE. * Unfortunately, we can't use BWIO with EV5, so for now, we always use SPARSE.
*/ */
#define vucp volatile unsigned char *
#define vusp volatile unsigned short *
#define vip volatile int *
#define vuip volatile unsigned int *
#define vulp volatile unsigned long *
__EXTERN_INLINE u8 mcpcia_inb(unsigned long in_addr)
{
unsigned long addr, hose, result;
addr = in_addr & 0xffffUL;
hose = in_addr & ~0xffffUL;
/* ??? I wish I could get rid of this. But there's no ioremap
equivalent for I/O space. PCI I/O can be forced into the
correct hose's I/O region, but that doesn't take care of
legacy ISA crap. */
hose += MCPCIA_IO_BIAS;
result = *(vip) ((addr << 5) + hose + 0x00);
return __kernel_extbl(result, addr & 3);
}
__EXTERN_INLINE void mcpcia_outb(u8 b, unsigned long in_addr)
{
unsigned long addr, hose, w;
addr = in_addr & 0xffffUL;
hose = in_addr & ~0xffffUL;
hose += MCPCIA_IO_BIAS;
w = __kernel_insbl(b, addr & 3);
*(vuip) ((addr << 5) + hose + 0x00) = w;
mb();
}
__EXTERN_INLINE u16 mcpcia_inw(unsigned long in_addr)
{
unsigned long addr, hose, result;
addr = in_addr & 0xffffUL;
hose = in_addr & ~0xffffUL;
hose += MCPCIA_IO_BIAS;
result = *(vip) ((addr << 5) + hose + 0x08);
return __kernel_extwl(result, addr & 3);
}
__EXTERN_INLINE void mcpcia_outw(u16 b, unsigned long in_addr)
{
unsigned long addr, hose, w;
addr = in_addr & 0xffffUL;
hose = in_addr & ~0xffffUL;
hose += MCPCIA_IO_BIAS;
w = __kernel_inswl(b, addr & 3);
*(vuip) ((addr << 5) + hose + 0x08) = w;
mb();
}
__EXTERN_INLINE u32 mcpcia_inl(unsigned long in_addr)
{
unsigned long addr, hose;
addr = in_addr & 0xffffUL;
hose = in_addr & ~0xffffUL;
hose += MCPCIA_IO_BIAS;
return *(vuip) ((addr << 5) + hose + 0x18);
}
__EXTERN_INLINE void mcpcia_outl(u32 b, unsigned long in_addr)
{
unsigned long addr, hose;
addr = in_addr & 0xffffUL;
hose = in_addr & ~0xffffUL;
hose += MCPCIA_IO_BIAS;
*(vuip) ((addr << 5) + hose + 0x18) = b;
mb();
}
/* /*
* Memory functions. 64-bit and 32-bit accesses are done through * Memory functions. 64-bit and 32-bit accesses are done through
* dense memory space, everything else through sparse space. * dense memory space, everything else through sparse space.
...@@ -328,149 +243,131 @@ __EXTERN_INLINE void mcpcia_outl(u32 b, unsigned long in_addr) ...@@ -328,149 +243,131 @@ __EXTERN_INLINE void mcpcia_outl(u32 b, unsigned long in_addr)
* *
*/ */
__EXTERN_INLINE unsigned long mcpcia_ioremap(unsigned long addr, #define vip volatile int __force *
unsigned long size #define vuip volatile unsigned int __force *
__attribute__((unused)))
{ #ifdef MCPCIA_ONE_HAE_WINDOW
return addr + MCPCIA_MEM_BIAS; #define MCPCIA_FROB_MMIO \
} if (__mcpcia_is_mmio(hose)) { \
set_hae(hose & 0xffffffff); \
hose = hose - MCPCIA_DENSE(4) + MCPCIA_SPARSE(4); \
}
#else
#define MCPCIA_FROB_MMIO \
if (__mcpcia_is_mmio(hose)) { \
hose = hose - MCPCIA_DENSE(4) + MCPCIA_SPARSE(4); \
}
#endif
__EXTERN_INLINE void mcpcia_iounmap(unsigned long addr) static inline int __mcpcia_is_mmio(unsigned long addr)
{ {
return; return (addr & 0x80000000UL) == 0;
} }
__EXTERN_INLINE int mcpcia_is_ioaddr(unsigned long addr) __EXTERN_INLINE unsigned int mcpcia_ioread8(void __iomem *xaddr)
{ {
return addr >= MCPCIA_SPARSE(0); unsigned long addr = (unsigned long)xaddr & MCPCIA_MEM_MASK;
unsigned long hose = (unsigned long)xaddr & ~MCPCIA_MEM_MASK;
unsigned long result;
MCPCIA_FROB_MMIO;
result = *(vip) ((addr << 5) + hose + 0x00);
return __kernel_extbl(result, addr & 3);
} }
__EXTERN_INLINE u8 mcpcia_readb(unsigned long in_addr) __EXTERN_INLINE void mcpcia_iowrite8(u8 b, void __iomem *xaddr)
{ {
unsigned long addr = in_addr & 0xffffffffUL; unsigned long addr = (unsigned long)xaddr & MCPCIA_MEM_MASK;
unsigned long hose = in_addr & ~0xffffffffUL; unsigned long hose = (unsigned long)xaddr & ~MCPCIA_MEM_MASK;
unsigned long result, work; unsigned long w;
#ifndef MCPCIA_ONE_HAE_WINDOW MCPCIA_FROB_MMIO;
unsigned long msb;
msb = addr & ~MCPCIA_MEM_MASK;
set_hae(msb);
#endif
addr = addr & MCPCIA_MEM_MASK;
hose = hose - MCPCIA_DENSE(4) + MCPCIA_SPARSE(4); w = __kernel_insbl(b, addr & 3);
work = ((addr << 5) + hose + 0x00); *(vuip) ((addr << 5) + hose + 0x00) = w;
result = *(vip) work;
return __kernel_extbl(result, addr & 3);
} }
__EXTERN_INLINE u16 mcpcia_readw(unsigned long in_addr) __EXTERN_INLINE unsigned int mcpcia_ioread16(void __iomem *xaddr)
{ {
unsigned long addr = in_addr & 0xffffffffUL; unsigned long addr = (unsigned long)xaddr & MCPCIA_MEM_MASK;
unsigned long hose = in_addr & ~0xffffffffUL; unsigned long hose = (unsigned long)xaddr & ~MCPCIA_MEM_MASK;
unsigned long result, work; unsigned long result;
#ifndef MCPCIA_ONE_HAE_WINDOW MCPCIA_FROB_MMIO;
unsigned long msb;
msb = addr & ~MCPCIA_MEM_MASK;
set_hae(msb);
#endif
addr = addr & MCPCIA_MEM_MASK;
hose = hose - MCPCIA_DENSE(4) + MCPCIA_SPARSE(4); result = *(vip) ((addr << 5) + hose + 0x08);
work = ((addr << 5) + hose + 0x08);
result = *(vip) work;
return __kernel_extwl(result, addr & 3); return __kernel_extwl(result, addr & 3);
} }
__EXTERN_INLINE void mcpcia_writeb(u8 b, unsigned long in_addr) __EXTERN_INLINE void mcpcia_iowrite16(u16 b, void __iomem *xaddr)
{ {
unsigned long addr = in_addr & 0xffffffffUL; unsigned long addr = (unsigned long)xaddr & MCPCIA_MEM_MASK;
unsigned long hose = in_addr & ~0xffffffffUL; unsigned long hose = (unsigned long)xaddr & ~MCPCIA_MEM_MASK;
unsigned long w; unsigned long w;
#ifndef MCPCIA_ONE_HAE_WINDOW MCPCIA_FROB_MMIO;
unsigned long msb;
msb = addr & ~MCPCIA_MEM_MASK;
set_hae(msb);
#endif
addr = addr & MCPCIA_MEM_MASK;
w = __kernel_insbl(b, in_addr & 3); w = __kernel_inswl(b, addr & 3);
hose = hose - MCPCIA_DENSE(4) + MCPCIA_SPARSE(4); *(vuip) ((addr << 5) + hose + 0x08) = w;
*(vuip) ((addr << 5) + hose + 0x00) = w;
} }
__EXTERN_INLINE void mcpcia_writew(u16 b, unsigned long in_addr) __EXTERN_INLINE unsigned int mcpcia_ioread32(void __iomem *xaddr)
{ {
unsigned long addr = in_addr & 0xffffffffUL; unsigned long addr = (unsigned long)xaddr;
unsigned long hose = in_addr & ~0xffffffffUL;
unsigned long w;
#ifndef MCPCIA_ONE_HAE_WINDOW if (!__mcpcia_is_mmio(addr))
unsigned long msb; addr = ((addr & 0xffff) << 5) + (addr & ~0xfffful) + 0x18;
msb = addr & ~MCPCIA_MEM_MASK;
set_hae(msb);
#endif
addr = addr & MCPCIA_MEM_MASK;
w = __kernel_inswl(b, in_addr & 3); return *(vuip)addr;
hose = hose - MCPCIA_DENSE(4) + MCPCIA_SPARSE(4); }
*(vuip) ((addr << 5) + hose + 0x08) = w;
__EXTERN_INLINE void mcpcia_iowrite32(u32 b, void __iomem *xaddr)
{
unsigned long addr = (unsigned long)xaddr;
if (!__mcpcia_is_mmio(addr))
addr = ((addr & 0xffff) << 5) + (addr & ~0xfffful) + 0x18;
*(vuip)addr = b;
} }
__EXTERN_INLINE u32 mcpcia_readl(unsigned long addr)
__EXTERN_INLINE void __iomem *mcpcia_ioportmap(unsigned long addr)
{ {
return (*(vuip)addr) & 0xffffffff; return (void __iomem *)(addr + MCPCIA_IO_BIAS);
} }
__EXTERN_INLINE u64 mcpcia_readq(unsigned long addr) __EXTERN_INLINE void __iomem *mcpcia_ioremap(unsigned long addr,
unsigned long size)
{ {
return *(vulp)addr; return (void __iomem *)(addr + MCPCIA_MEM_BIAS);
} }
__EXTERN_INLINE void mcpcia_writel(u32 b, unsigned long addr) __EXTERN_INLINE int mcpcia_is_ioaddr(unsigned long addr)
{ {
*(vuip)addr = b; return addr >= MCPCIA_SPARSE(0);
} }
__EXTERN_INLINE void mcpcia_writeq(u64 b, unsigned long addr) __EXTERN_INLINE int mcpcia_is_mmio(const volatile void __iomem *xaddr)
{ {
*(vulp)addr = b; unsigned long addr = (unsigned long) xaddr;
return __mcpcia_is_mmio(addr);
} }
#undef vucp #undef MCPCIA_FROB_MMIO
#undef vusp
#undef vip #undef vip
#undef vuip #undef vuip
#undef vulp
#undef __IO_PREFIX
#ifdef __WANT_IO_DEF #define __IO_PREFIX mcpcia
#define mcpcia_trivial_rw_bw 2
#define __inb(p) mcpcia_inb((unsigned long)(p)) #define mcpcia_trivial_rw_lq 1
#define __inw(p) mcpcia_inw((unsigned long)(p)) #define mcpcia_trivial_io_bw 0
#define __inl(p) mcpcia_inl((unsigned long)(p)) #define mcpcia_trivial_io_lq 0
#define __outb(x,p) mcpcia_outb((x),(unsigned long)(p)) #define mcpcia_trivial_iounmap 1
#define __outw(x,p) mcpcia_outw((x),(unsigned long)(p)) #include <asm/io_trivial.h>
#define __outl(x,p) mcpcia_outl((x),(unsigned long)(p))
#define __readb(a) mcpcia_readb((unsigned long)(a))
#define __readw(a) mcpcia_readw((unsigned long)(a))
#define __readl(a) mcpcia_readl((unsigned long)(a))
#define __readq(a) mcpcia_readq((unsigned long)(a))
#define __writeb(x,a) mcpcia_writeb((x),(unsigned long)(a))
#define __writew(x,a) mcpcia_writew((x),(unsigned long)(a))
#define __writel(x,a) mcpcia_writel((x),(unsigned long)(a))
#define __writeq(x,a) mcpcia_writeq((x),(unsigned long)(a))
#define __ioremap(a,s) mcpcia_ioremap((unsigned long)(a),(s))
#define __iounmap(a) mcpcia_iounmap((unsigned long)(a))
#define __is_ioaddr(a) mcpcia_is_ioaddr((unsigned long)(a))
#define __raw_readl(a) __readl(a)
#define __raw_readq(a) __readq(a)
#define __raw_writel(v,a) __writel((v),(a))
#define __raw_writeq(v,a) __writeq((v),(a))
#endif /* __WANT_IO_DEF */
#ifdef __IO_EXTERN_INLINE #ifdef __IO_EXTERN_INLINE
#undef __EXTERN_INLINE #undef __EXTERN_INLINE
......
...@@ -63,49 +63,6 @@ struct el_POLARIS_sysdata_mcheck { ...@@ -63,49 +63,6 @@ struct el_POLARIS_sysdata_mcheck {
* However, we will support only the BWX form. * However, we will support only the BWX form.
*/ */
#define vucp volatile unsigned char *
#define vusp volatile unsigned short *
#define vuip volatile unsigned int *
#define vulp volatile unsigned long *
__EXTERN_INLINE u8 polaris_inb(unsigned long addr)
{
/* ??? I wish I could get rid of this. But there's no ioremap
equivalent for I/O space. PCI I/O can be forced into the
POLARIS I/O region, but that doesn't take care of legacy
ISA crap. */
return __kernel_ldbu(*(vucp)(addr + POLARIS_DENSE_IO_BASE));
}
__EXTERN_INLINE void polaris_outb(u8 b, unsigned long addr)
{
__kernel_stb(b, *(vucp)(addr + POLARIS_DENSE_IO_BASE));
mb();
}
__EXTERN_INLINE u16 polaris_inw(unsigned long addr)
{
return __kernel_ldwu(*(vusp)(addr + POLARIS_DENSE_IO_BASE));
}
__EXTERN_INLINE void polaris_outw(u16 b, unsigned long addr)
{
__kernel_stw(b, *(vusp)(addr + POLARIS_DENSE_IO_BASE));
mb();
}
__EXTERN_INLINE u32 polaris_inl(unsigned long addr)
{
return *(vuip)(addr + POLARIS_DENSE_IO_BASE);
}
__EXTERN_INLINE void polaris_outl(u32 b, unsigned long addr)
{
*(vuip)(addr + POLARIS_DENSE_IO_BASE) = b;
mb();
}
/* /*
* Memory functions. Polaris allows all accesses (byte/word * Memory functions. Polaris allows all accesses (byte/word
* as well as long/quad) to be done through dense space. * as well as long/quad) to be done through dense space.
...@@ -113,104 +70,35 @@ __EXTERN_INLINE void polaris_outl(u32 b, unsigned long addr) ...@@ -113,104 +70,35 @@ __EXTERN_INLINE void polaris_outl(u32 b, unsigned long addr)
* We will only support DENSE access via BWX insns. * We will only support DENSE access via BWX insns.
*/ */
__EXTERN_INLINE u8 polaris_readb(unsigned long addr) __EXTERN_INLINE void __iomem *polaris_ioportmap(unsigned long addr)
{
return __kernel_ldbu(*(vucp)addr);
}
__EXTERN_INLINE u16 polaris_readw(unsigned long addr)
{ {
return __kernel_ldwu(*(vusp)addr); return (void __iomem *)(addr + POLARIS_DENSE_IO_BASE);
} }
__EXTERN_INLINE u32 polaris_readl(unsigned long addr) __EXTERN_INLINE void __iomem *polaris_ioremap(unsigned long addr,
unsigned long size)
{ {
return (*(vuip)addr) & 0xffffffff; return (void __iomem *)(addr + POLARIS_DENSE_MEM_BASE);
} }
__EXTERN_INLINE u64 polaris_readq(unsigned long addr) __EXTERN_INLINE int polaris_is_ioaddr(unsigned long addr)
{
return *(vulp)addr;
}
__EXTERN_INLINE void polaris_writeb(u8 b, unsigned long addr)
{
__kernel_stb(b, *(vucp)addr);
}
__EXTERN_INLINE void polaris_writew(u16 b, unsigned long addr)
{
__kernel_stw(b, *(vusp)addr);
}
__EXTERN_INLINE void polaris_writel(u32 b, unsigned long addr)
{
*(vuip)addr = b;
}
__EXTERN_INLINE void polaris_writeq(u64 b, unsigned long addr)
{
*(vulp)addr = b;
}
__EXTERN_INLINE unsigned long polaris_ioremap(unsigned long addr,
unsigned long size
__attribute__((unused)))
{
return addr + POLARIS_DENSE_MEM_BASE;
}
__EXTERN_INLINE void polaris_iounmap(unsigned long addr)
{ {
return; return addr >= POLARIS_SPARSE_MEM_BASE;
} }
__EXTERN_INLINE int polaris_is_ioaddr(unsigned long addr) __EXTERN_INLINE int polaris_is_mmio(const volatile void __iomem *addr)
{ {
return addr >= POLARIS_SPARSE_MEM_BASE; return (unsigned long)addr < POLARIS_SPARSE_IO_BASE;
} }
#undef vucp #undef __IO_PREFIX
#undef vusp #define __IO_PREFIX polaris
#undef vuip #define polaris_trivial_rw_bw 1
#undef vulp #define polaris_trivial_rw_lq 1
#define polaris_trivial_io_bw 1
#ifdef __WANT_IO_DEF #define polaris_trivial_io_lq 1
#define polaris_trivial_iounmap 1
#define __inb(p) polaris_inb((unsigned long)(p)) #include <asm/io_trivial.h>
#define __inw(p) polaris_inw((unsigned long)(p))
#define __inl(p) polaris_inl((unsigned long)(p))
#define __outb(x,p) polaris_outb((x),(unsigned long)(p))
#define __outw(x,p) polaris_outw((x),(unsigned long)(p))
#define __outl(x,p) polaris_outl((x),(unsigned long)(p))
#define __readb(a) polaris_readb((unsigned long)(a))
#define __readw(a) polaris_readw((unsigned long)(a))
#define __readl(a) polaris_readl((unsigned long)(a))
#define __readq(a) polaris_readq((unsigned long)(a))
#define __writeb(x,a) polaris_writeb((x),(unsigned long)(a))
#define __writew(x,a) polaris_writew((x),(unsigned long)(a))
#define __writel(x,a) polaris_writel((x),(unsigned long)(a))
#define __writeq(x,a) polaris_writeq((x),(unsigned long)(a))
#define __ioremap(a,s) polaris_ioremap((unsigned long)(a),(s))
#define __iounmap(a) polaris_iounmap((unsigned long)(a))
#define __is_ioaddr(a) polaris_is_ioaddr((unsigned long)(a))
#define inb(p) __inb(p)
#define inw(p) __inw(p)
#define inl(p) __inl(p)
#define outb(x,p) __outb((x),(p))
#define outw(x,p) __outw((x),(p))
#define outl(x,p) __outl((x),(p))
#define __raw_readb(a) __readb(a)
#define __raw_readw(a) __readw(a)
#define __raw_readl(a) __readl(a)
#define __raw_readq(a) __readq(a)
#define __raw_writeb(v,a) __writeb((v),(a))
#define __raw_writew(v,a) __writew((v),(a))
#define __raw_writel(v,a) __writel((v),(a))
#define __raw_writeq(v,a) __writeq((v),(a))
#endif /* __WANT_IO_DEF */
#ifdef __IO_EXTERN_INLINE #ifdef __IO_EXTERN_INLINE
#undef __EXTERN_INLINE #undef __EXTERN_INLINE
......
...@@ -199,8 +199,8 @@ struct el_t2_procdata_mcheck { ...@@ -199,8 +199,8 @@ struct el_t2_procdata_mcheck {
struct el_t2_logout_header { struct el_t2_logout_header {
unsigned int elfl_size; /* size in bytes of logout area. */ unsigned int elfl_size; /* size in bytes of logout area. */
int elfl_sbz1:31; /* Should be zero. */ unsigned int elfl_sbz1:31; /* Should be zero. */
char elfl_retry:1; /* Retry flag. */ unsigned int elfl_retry:1; /* Retry flag. */
unsigned int elfl_procoffset; /* Processor-specific offset. */ unsigned int elfl_procoffset; /* Processor-specific offset. */
unsigned int elfl_sysoffset; /* Offset of system-specific. */ unsigned int elfl_sysoffset; /* Offset of system-specific. */
unsigned int elfl_error_type; /* PAL error type code. */ unsigned int elfl_error_type; /* PAL error type code. */
...@@ -357,13 +357,13 @@ struct el_t2_frame_corrected { ...@@ -357,13 +357,13 @@ struct el_t2_frame_corrected {
#define vip volatile int * #define vip volatile int *
#define vuip volatile unsigned int * #define vuip volatile unsigned int *
__EXTERN_INLINE u8 t2_inb(unsigned long addr) static inline u8 t2_inb(unsigned long addr)
{ {
long result = *(vip) ((addr << 5) + T2_IO + 0x00); long result = *(vip) ((addr << 5) + T2_IO + 0x00);
return __kernel_extbl(result, addr & 3); return __kernel_extbl(result, addr & 3);
} }
__EXTERN_INLINE void t2_outb(u8 b, unsigned long addr) static inline void t2_outb(u8 b, unsigned long addr)
{ {
unsigned long w; unsigned long w;
...@@ -372,13 +372,13 @@ __EXTERN_INLINE void t2_outb(u8 b, unsigned long addr) ...@@ -372,13 +372,13 @@ __EXTERN_INLINE void t2_outb(u8 b, unsigned long addr)
mb(); mb();
} }
__EXTERN_INLINE u16 t2_inw(unsigned long addr) static inline u16 t2_inw(unsigned long addr)
{ {
long result = *(vip) ((addr << 5) + T2_IO + 0x08); long result = *(vip) ((addr << 5) + T2_IO + 0x08);
return __kernel_extwl(result, addr & 3); return __kernel_extwl(result, addr & 3);
} }
__EXTERN_INLINE void t2_outw(u16 b, unsigned long addr) static inline void t2_outw(u16 b, unsigned long addr)
{ {
unsigned long w; unsigned long w;
...@@ -387,12 +387,12 @@ __EXTERN_INLINE void t2_outw(u16 b, unsigned long addr) ...@@ -387,12 +387,12 @@ __EXTERN_INLINE void t2_outw(u16 b, unsigned long addr)
mb(); mb();
} }
__EXTERN_INLINE u32 t2_inl(unsigned long addr) static inline u32 t2_inl(unsigned long addr)
{ {
return *(vuip) ((addr << 5) + T2_IO + 0x18); return *(vuip) ((addr << 5) + T2_IO + 0x18);
} }
__EXTERN_INLINE void t2_outl(u32 b, unsigned long addr) static inline void t2_outl(u32 b, unsigned long addr)
{ {
*(vuip) ((addr << 5) + T2_IO + 0x18) = b; *(vuip) ((addr << 5) + T2_IO + 0x18) = b;
mb(); mb();
...@@ -438,8 +438,9 @@ __EXTERN_INLINE void t2_outl(u32 b, unsigned long addr) ...@@ -438,8 +438,9 @@ __EXTERN_INLINE void t2_outl(u32 b, unsigned long addr)
static spinlock_t t2_hae_lock = SPIN_LOCK_UNLOCKED; static spinlock_t t2_hae_lock = SPIN_LOCK_UNLOCKED;
__EXTERN_INLINE u8 t2_readb(unsigned long addr) __EXTERN_INLINE u8 t2_readb(const volatile void __iomem *xaddr)
{ {
unsigned long addr = (unsigned long) xaddr;
unsigned long result, msb; unsigned long result, msb;
unsigned long flags; unsigned long flags;
spin_lock_irqsave(&t2_hae_lock, flags); spin_lock_irqsave(&t2_hae_lock, flags);
...@@ -451,8 +452,9 @@ __EXTERN_INLINE u8 t2_readb(unsigned long addr) ...@@ -451,8 +452,9 @@ __EXTERN_INLINE u8 t2_readb(unsigned long addr)
return __kernel_extbl(result, addr & 3); return __kernel_extbl(result, addr & 3);
} }
__EXTERN_INLINE u16 t2_readw(unsigned long addr) __EXTERN_INLINE u16 t2_readw(const volatile void __iomem *xaddr)
{ {
unsigned long addr = (unsigned long) xaddr;
unsigned long result, msb; unsigned long result, msb;
unsigned long flags; unsigned long flags;
spin_lock_irqsave(&t2_hae_lock, flags); spin_lock_irqsave(&t2_hae_lock, flags);
...@@ -468,8 +470,9 @@ __EXTERN_INLINE u16 t2_readw(unsigned long addr) ...@@ -468,8 +470,9 @@ __EXTERN_INLINE u16 t2_readw(unsigned long addr)
* On SABLE with T2, we must use SPARSE memory even for 32-bit access, * On SABLE with T2, we must use SPARSE memory even for 32-bit access,
* because we cannot access all of DENSE without changing its HAE. * because we cannot access all of DENSE without changing its HAE.
*/ */
__EXTERN_INLINE u32 t2_readl(unsigned long addr) __EXTERN_INLINE u32 t2_readl(const volatile void __iomem *xaddr)
{ {
unsigned long addr = (unsigned long) xaddr;
unsigned long result, msb; unsigned long result, msb;
unsigned long flags; unsigned long flags;
spin_lock_irqsave(&t2_hae_lock, flags); spin_lock_irqsave(&t2_hae_lock, flags);
...@@ -481,8 +484,9 @@ __EXTERN_INLINE u32 t2_readl(unsigned long addr) ...@@ -481,8 +484,9 @@ __EXTERN_INLINE u32 t2_readl(unsigned long addr)
return result & 0xffffffffUL; return result & 0xffffffffUL;
} }
__EXTERN_INLINE u64 t2_readq(unsigned long addr) __EXTERN_INLINE u64 t2_readq(const volatile void __iomem *xaddr)
{ {
unsigned long addr = (unsigned long) xaddr;
unsigned long r0, r1, work, msb; unsigned long r0, r1, work, msb;
unsigned long flags; unsigned long flags;
spin_lock_irqsave(&t2_hae_lock, flags); spin_lock_irqsave(&t2_hae_lock, flags);
...@@ -496,8 +500,9 @@ __EXTERN_INLINE u64 t2_readq(unsigned long addr) ...@@ -496,8 +500,9 @@ __EXTERN_INLINE u64 t2_readq(unsigned long addr)
return r1 << 32 | r0; return r1 << 32 | r0;
} }
__EXTERN_INLINE void t2_writeb(u8 b, unsigned long addr) __EXTERN_INLINE void t2_writeb(u8 b, volatile void __iomem *xaddr)
{ {
unsigned long addr = (unsigned long) xaddr;
unsigned long msb, w; unsigned long msb, w;
unsigned long flags; unsigned long flags;
spin_lock_irqsave(&t2_hae_lock, flags); spin_lock_irqsave(&t2_hae_lock, flags);
...@@ -509,8 +514,9 @@ __EXTERN_INLINE void t2_writeb(u8 b, unsigned long addr) ...@@ -509,8 +514,9 @@ __EXTERN_INLINE void t2_writeb(u8 b, unsigned long addr)
spin_unlock_irqrestore(&t2_hae_lock, flags); spin_unlock_irqrestore(&t2_hae_lock, flags);
} }
__EXTERN_INLINE void t2_writew(u16 b, unsigned long addr) __EXTERN_INLINE void t2_writew(u16 b, volatile void __iomem *xaddr)
{ {
unsigned long addr = (unsigned long) xaddr;
unsigned long msb, w; unsigned long msb, w;
unsigned long flags; unsigned long flags;
spin_lock_irqsave(&t2_hae_lock, flags); spin_lock_irqsave(&t2_hae_lock, flags);
...@@ -526,8 +532,9 @@ __EXTERN_INLINE void t2_writew(u16 b, unsigned long addr) ...@@ -526,8 +532,9 @@ __EXTERN_INLINE void t2_writew(u16 b, unsigned long addr)
* On SABLE with T2, we must use SPARSE memory even for 32-bit access, * On SABLE with T2, we must use SPARSE memory even for 32-bit access,
* because we cannot access all of DENSE without changing its HAE. * because we cannot access all of DENSE without changing its HAE.
*/ */
__EXTERN_INLINE void t2_writel(u32 b, unsigned long addr) __EXTERN_INLINE void t2_writel(u32 b, volatile void __iomem *xaddr)
{ {
unsigned long addr = (unsigned long) xaddr;
unsigned long msb; unsigned long msb;
unsigned long flags; unsigned long flags;
spin_lock_irqsave(&t2_hae_lock, flags); spin_lock_irqsave(&t2_hae_lock, flags);
...@@ -538,8 +545,9 @@ __EXTERN_INLINE void t2_writel(u32 b, unsigned long addr) ...@@ -538,8 +545,9 @@ __EXTERN_INLINE void t2_writel(u32 b, unsigned long addr)
spin_unlock_irqrestore(&t2_hae_lock, flags); spin_unlock_irqrestore(&t2_hae_lock, flags);
} }
__EXTERN_INLINE void t2_writeq(u64 b, unsigned long addr) __EXTERN_INLINE void t2_writeq(u64 b, volatile void __iomem *xaddr)
{ {
unsigned long addr = (unsigned long) xaddr;
unsigned long msb, work; unsigned long msb, work;
unsigned long flags; unsigned long flags;
spin_lock_irqsave(&t2_hae_lock, flags); spin_lock_irqsave(&t2_hae_lock, flags);
...@@ -552,16 +560,15 @@ __EXTERN_INLINE void t2_writeq(u64 b, unsigned long addr) ...@@ -552,16 +560,15 @@ __EXTERN_INLINE void t2_writeq(u64 b, unsigned long addr)
spin_unlock_irqrestore(&t2_hae_lock, flags); spin_unlock_irqrestore(&t2_hae_lock, flags);
} }
__EXTERN_INLINE unsigned long t2_ioremap(unsigned long addr, __EXTERN_INLINE void __iomem *t2_ioportmap(unsigned long addr)
unsigned long size
__attribute__((unused)))
{ {
return addr; return (void __iomem *)(addr + T2_IO);
} }
__EXTERN_INLINE void t2_iounmap(unsigned long addr) __EXTERN_INLINE void __iomem *t2_ioremap(unsigned long addr,
unsigned long size)
{ {
return; return (void __iomem *)(addr + T2_DENSE_MEM);
} }
__EXTERN_INLINE int t2_is_ioaddr(unsigned long addr) __EXTERN_INLINE int t2_is_ioaddr(unsigned long addr)
...@@ -569,30 +576,47 @@ __EXTERN_INLINE int t2_is_ioaddr(unsigned long addr) ...@@ -569,30 +576,47 @@ __EXTERN_INLINE int t2_is_ioaddr(unsigned long addr)
return (long)addr >= 0; return (long)addr >= 0;
} }
__EXTERN_INLINE int t2_is_mmio(const volatile void __iomem *addr)
{
return (unsigned long)addr >= T2_DENSE_MEM;
}
/* New-style ioread interface. The mmio routines are so ugly for T2 that
it doesn't make sense to merge the pio and mmio routines. */
#define IOPORT(OS, NS) \
__EXTERN_INLINE unsigned int t2_ioread##NS(void __iomem *xaddr) \
{ \
if (t2_is_mmio(xaddr)) \
return t2_read##OS(xaddr - T2_DENSE_MEM); \
else \
return t2_in##OS((unsigned long)xaddr - T2_IO); \
} \
__EXTERN_INLINE void t2_iowrite##NS(u##NS b, void __iomem *xaddr) \
{ \
if (t2_is_mmio(xaddr)) \
t2_write##OS(b, xaddr - T2_DENSE_MEM); \
else \
t2_out##OS(b, (unsigned long)xaddr - T2_IO); \
}
IOPORT(b, 8)
IOPORT(w, 16)
IOPORT(l, 32)
#undef IOPORT
#undef vip #undef vip
#undef vuip #undef vuip
#ifdef __WANT_IO_DEF #undef __IO_PREFIX
#define __IO_PREFIX t2
#define __inb(p) t2_inb((unsigned long)(p)) #define t2_trivial_rw_bw 0
#define __inw(p) t2_inw((unsigned long)(p)) #define t2_trivial_rw_lq 0
#define __inl(p) t2_inl((unsigned long)(p)) #define t2_trivial_io_bw 0
#define __outb(x,p) t2_outb((x),(unsigned long)(p)) #define t2_trivial_io_lq 0
#define __outw(x,p) t2_outw((x),(unsigned long)(p)) #define t2_trivial_iounmap 1
#define __outl(x,p) t2_outl((x),(unsigned long)(p)) #include <asm/io_trivial.h>
#define __readb(a) t2_readb((unsigned long)(a))
#define __readw(a) t2_readw((unsigned long)(a))
#define __readl(a) t2_readl((unsigned long)(a))
#define __readq(a) t2_readq((unsigned long)(a))
#define __writeb(x,a) t2_writeb((x),(unsigned long)(a))
#define __writew(x,a) t2_writew((x),(unsigned long)(a))
#define __writel(x,a) t2_writel((x),(unsigned long)(a))
#define __writeq(x,a) t2_writeq((x),(unsigned long)(a))
#define __ioremap(a,s) t2_ioremap((unsigned long)(a),(s))
#define __iounmap(a) t2_iounmap((unsigned long)(a))
#define __is_ioaddr(a) t2_is_ioaddr((unsigned long)(a))
#endif /* __WANT_IO_DEF */
#ifdef __IO_EXTERN_INLINE #ifdef __IO_EXTERN_INLINE
#undef __EXTERN_INLINE #undef __EXTERN_INLINE
......
...@@ -377,149 +377,33 @@ struct el_PRIVATEER_envdata_mcheck { ...@@ -377,149 +377,33 @@ struct el_PRIVATEER_envdata_mcheck {
* can only use linear accesses to get at PCI/AGP memory and I/O spaces. * can only use linear accesses to get at PCI/AGP memory and I/O spaces.
*/ */
#define vucp volatile unsigned char *
#define vusp volatile unsigned short *
#define vuip volatile unsigned int *
#define vulp volatile unsigned long *
__EXTERN_INLINE u8 titan_inb(unsigned long addr)
{
/* ??? I wish I could get rid of this. But there's no ioremap
equivalent for I/O space. PCI I/O can be forced into the
correct hose's I/O region, but that doesn't take care of
legacy ISA crap. */
addr += TITAN_IO_BIAS;
return __kernel_ldbu(*(vucp)addr);
}
__EXTERN_INLINE void titan_outb(u8 b, unsigned long addr)
{
addr += TITAN_IO_BIAS;
__kernel_stb(b, *(vucp)addr);
mb();
}
__EXTERN_INLINE u16 titan_inw(unsigned long addr)
{
addr += TITAN_IO_BIAS;
return __kernel_ldwu(*(vusp)addr);
}
__EXTERN_INLINE void titan_outw(u16 b, unsigned long addr)
{
addr += TITAN_IO_BIAS;
__kernel_stw(b, *(vusp)addr);
mb();
}
__EXTERN_INLINE u32 titan_inl(unsigned long addr)
{
addr += TITAN_IO_BIAS;
return *(vuip)addr;
}
__EXTERN_INLINE void titan_outl(u32 b, unsigned long addr)
{
addr += TITAN_IO_BIAS;
*(vuip)addr = b;
mb();
}
/* /*
* Memory functions. all accesses are done through linear space. * Memory functions. all accesses are done through linear space.
*/ */
extern unsigned long titan_ioremap(unsigned long addr, unsigned long size); __EXTERN_INLINE void __iomem *titan_ioportmap(unsigned long addr)
extern void titan_iounmap(unsigned long addr);
__EXTERN_INLINE int titan_is_ioaddr(unsigned long addr)
{
return addr >= TITAN_BASE;
}
__EXTERN_INLINE u8 titan_readb(unsigned long addr)
{
return __kernel_ldbu(*(vucp)addr);
}
__EXTERN_INLINE u16 titan_readw(unsigned long addr)
{
return __kernel_ldwu(*(vusp)addr);
}
__EXTERN_INLINE u32 titan_readl(unsigned long addr)
{
return (*(vuip)addr) & 0xffffffff;
}
__EXTERN_INLINE u64 titan_readq(unsigned long addr)
{ {
return *(vulp)addr; return (void __iomem *)(addr + TITAN_IO_BIAS);
} }
__EXTERN_INLINE void titan_writeb(u8 b, unsigned long addr) extern void __iomem *titan_ioremap(unsigned long addr, unsigned long size);
{ extern void titan_iounmap(volatile void __iomem *addr);
__kernel_stb(b, *(vucp)addr);
}
__EXTERN_INLINE void titan_writew(u16 b, unsigned long addr) __EXTERN_INLINE int titan_is_ioaddr(unsigned long addr)
{
__kernel_stw(b, *(vusp)addr);
}
__EXTERN_INLINE void titan_writel(u32 b, unsigned long addr)
{ {
*(vuip)addr = b; return addr >= TITAN_BASE;
} }
__EXTERN_INLINE void titan_writeq(u64 b, unsigned long addr) extern int titan_is_mmio(const volatile void __iomem *addr);
{
*(vulp)addr = b;
}
#undef vucp #undef __IO_PREFIX
#undef vusp #define __IO_PREFIX titan
#undef vuip #define titan_trivial_rw_bw 1
#undef vulp #define titan_trivial_rw_lq 1
#define titan_trivial_io_bw 1
#ifdef __WANT_IO_DEF #define titan_trivial_io_lq 1
#define titan_trivial_iounmap 0
#define __inb(p) titan_inb((unsigned long)(p)) #include <asm/io_trivial.h>
#define __inw(p) titan_inw((unsigned long)(p))
#define __inl(p) titan_inl((unsigned long)(p))
#define __outb(x,p) titan_outb((x),(unsigned long)(p))
#define __outw(x,p) titan_outw((x),(unsigned long)(p))
#define __outl(x,p) titan_outl((x),(unsigned long)(p))
#define __readb(a) titan_readb((unsigned long)(a))
#define __readw(a) titan_readw((unsigned long)(a))
#define __readl(a) titan_readl((unsigned long)(a))
#define __readq(a) titan_readq((unsigned long)(a))
#define __writeb(x,a) titan_writeb((x),(unsigned long)(a))
#define __writew(x,a) titan_writew((x),(unsigned long)(a))
#define __writel(x,a) titan_writel((x),(unsigned long)(a))
#define __writeq(x,a) titan_writeq((x),(unsigned long)(a))
#define __ioremap(a,s) titan_ioremap((unsigned long)(a),(s))
#define __iounmap(a) titan_iounmap((unsigned long)(a))
#define __is_ioaddr(a) titan_is_ioaddr((unsigned long)(a))
#define inb(port) __inb((port))
#define inw(port) __inw((port))
#define inl(port) __inl((port))
#define outb(v, port) __outb((v),(port))
#define outw(v, port) __outw((v),(port))
#define outl(v, port) __outl((v),(port))
#define __raw_readb(a) __readb((unsigned long)(a))
#define __raw_readw(a) __readw((unsigned long)(a))
#define __raw_readl(a) __readl((unsigned long)(a))
#define __raw_readq(a) __readq((unsigned long)(a))
#define __raw_writeb(v,a) __writeb((v),(unsigned long)(a))
#define __raw_writew(v,a) __writew((v),(unsigned long)(a))
#define __raw_writel(v,a) __writel((v),(unsigned long)(a))
#define __raw_writeq(v,a) __writeq((v),(unsigned long)(a))
#endif /* __WANT_IO_DEF */
#ifdef __IO_EXTERN_INLINE #ifdef __IO_EXTERN_INLINE
#undef __EXTERN_INLINE #undef __EXTERN_INLINE
......
...@@ -299,69 +299,19 @@ struct el_TSUNAMI_sysdata_mcheck { ...@@ -299,69 +299,19 @@ struct el_TSUNAMI_sysdata_mcheck {
* can only use linear accesses to get at PCI memory and I/O spaces. * can only use linear accesses to get at PCI memory and I/O spaces.
*/ */
#define vucp volatile unsigned char *
#define vusp volatile unsigned short *
#define vuip volatile unsigned int *
#define vulp volatile unsigned long *
__EXTERN_INLINE u8 tsunami_inb(unsigned long addr)
{
/* ??? I wish I could get rid of this. But there's no ioremap
equivalent for I/O space. PCI I/O can be forced into the
correct hose's I/O region, but that doesn't take care of
legacy ISA crap. */
addr += TSUNAMI_IO_BIAS;
return __kernel_ldbu(*(vucp)addr);
}
__EXTERN_INLINE void tsunami_outb(u8 b, unsigned long addr)
{
addr += TSUNAMI_IO_BIAS;
__kernel_stb(b, *(vucp)addr);
mb();
}
__EXTERN_INLINE u16 tsunami_inw(unsigned long addr)
{
addr += TSUNAMI_IO_BIAS;
return __kernel_ldwu(*(vusp)addr);
}
__EXTERN_INLINE void tsunami_outw(u16 b, unsigned long addr)
{
addr += TSUNAMI_IO_BIAS;
__kernel_stw(b, *(vusp)addr);
mb();
}
__EXTERN_INLINE u32 tsunami_inl(unsigned long addr)
{
addr += TSUNAMI_IO_BIAS;
return *(vuip)addr;
}
__EXTERN_INLINE void tsunami_outl(u32 b, unsigned long addr)
{
addr += TSUNAMI_IO_BIAS;
*(vuip)addr = b;
mb();
}
/* /*
* Memory functions. all accesses are done through linear space. * Memory functions. all accesses are done through linear space.
*/ */
__EXTERN_INLINE unsigned long tsunami_ioremap(unsigned long addr, __EXTERN_INLINE void __iomem *tsunami_ioportmap(unsigned long addr)
unsigned long size
__attribute__((unused)))
{ {
return addr + TSUNAMI_MEM_BIAS; return (void __iomem *)(addr + TSUNAMI_IO_BIAS);
} }
__EXTERN_INLINE void tsunami_iounmap(unsigned long addr) __EXTERN_INLINE void __iomem *tsunami_ioremap(unsigned long addr,
unsigned long size)
{ {
return; return (void __iomem *)(addr + TSUNAMI_MEM_BIAS);
} }
__EXTERN_INLINE int tsunami_is_ioaddr(unsigned long addr) __EXTERN_INLINE int tsunami_is_ioaddr(unsigned long addr)
...@@ -369,87 +319,20 @@ __EXTERN_INLINE int tsunami_is_ioaddr(unsigned long addr) ...@@ -369,87 +319,20 @@ __EXTERN_INLINE int tsunami_is_ioaddr(unsigned long addr)
return addr >= TSUNAMI_BASE; return addr >= TSUNAMI_BASE;
} }
__EXTERN_INLINE u8 tsunami_readb(unsigned long addr) __EXTERN_INLINE int tsunami_is_mmio(const volatile void __iomem *xaddr)
{
return __kernel_ldbu(*(vucp)addr);
}
__EXTERN_INLINE u16 tsunami_readw(unsigned long addr)
{
return __kernel_ldwu(*(vusp)addr);
}
__EXTERN_INLINE u32 tsunami_readl(unsigned long addr)
{
return *(vuip)addr;
}
__EXTERN_INLINE u64 tsunami_readq(unsigned long addr)
{
return *(vulp)addr;
}
__EXTERN_INLINE void tsunami_writeb(u8 b, unsigned long addr)
{
__kernel_stb(b, *(vucp)addr);
}
__EXTERN_INLINE void tsunami_writew(u16 b, unsigned long addr)
{
__kernel_stw(b, *(vusp)addr);
}
__EXTERN_INLINE void tsunami_writel(u32 b, unsigned long addr)
{
*(vuip)addr = b;
}
__EXTERN_INLINE void tsunami_writeq(u64 b, unsigned long addr)
{ {
*(vulp)addr = b; unsigned long addr = (unsigned long) xaddr;
return (addr & 0x100000000UL) == 0;
} }
#undef vucp #undef __IO_PREFIX
#undef vusp #define __IO_PREFIX tsunami
#undef vuip #define tsunami_trivial_rw_bw 1
#undef vulp #define tsunami_trivial_rw_lq 1
#define tsunami_trivial_io_bw 1
#ifdef __WANT_IO_DEF #define tsunami_trivial_io_lq 1
#define tsunami_trivial_iounmap 1
#define __inb(p) tsunami_inb((unsigned long)(p)) #include <asm/io_trivial.h>
#define __inw(p) tsunami_inw((unsigned long)(p))
#define __inl(p) tsunami_inl((unsigned long)(p))
#define __outb(x,p) tsunami_outb((x),(unsigned long)(p))
#define __outw(x,p) tsunami_outw((x),(unsigned long)(p))
#define __outl(x,p) tsunami_outl((x),(unsigned long)(p))
#define __readb(a) tsunami_readb((unsigned long)(a))
#define __readw(a) tsunami_readw((unsigned long)(a))
#define __readl(a) tsunami_readl((unsigned long)(a))
#define __readq(a) tsunami_readq((unsigned long)(a))
#define __writeb(x,a) tsunami_writeb((x),(unsigned long)(a))
#define __writew(x,a) tsunami_writew((x),(unsigned long)(a))
#define __writel(x,a) tsunami_writel((x),(unsigned long)(a))
#define __writeq(x,a) tsunami_writeq((x),(unsigned long)(a))
#define __ioremap(a,s) tsunami_ioremap((unsigned long)(a),(s))
#define __iounmap(a) tsunami_iounmap((unsigned long)(a))
#define __is_ioaddr(a) tsunami_is_ioaddr((unsigned long)(a))
#define inb(p) __inb(p)
#define inw(p) __inw(p)
#define inl(p) __inl(p)
#define outb(x,p) __outb((x),(p))
#define outw(x,p) __outw((x),(p))
#define outl(x,p) __outl((x),(p))
#define __raw_readb(a) __readb(a)
#define __raw_readw(a) __readw(a)
#define __raw_readl(a) __readl(a)
#define __raw_readq(a) __readq(a)
#define __raw_writeb(v,a) __writeb((v),(a))
#define __raw_writew(v,a) __writew((v),(a))
#define __raw_writel(v,a) __writel((v),(a))
#define __raw_writeq(v,a) __writeq((v),(a))
#endif /* __WANT_IO_DEF */
#ifdef __IO_EXTERN_INLINE #ifdef __IO_EXTERN_INLINE
#undef __EXTERN_INLINE #undef __EXTERN_INLINE
......
...@@ -273,69 +273,19 @@ typedef struct { ...@@ -273,69 +273,19 @@ typedef struct {
#define __IO_EXTERN_INLINE #define __IO_EXTERN_INLINE
#endif #endif
#define vucp volatile unsigned char *
#define vusp volatile unsigned short *
#define vuip volatile unsigned int *
#define vulp volatile unsigned long *
__EXTERN_INLINE u8 wildfire_inb(unsigned long addr)
{
/* ??? I wish I could get rid of this. But there's no ioremap
equivalent for I/O space. PCI I/O can be forced into the
correct hose's I/O region, but that doesn't take care of
legacy ISA crap. */
addr += WILDFIRE_IO_BIAS;
return __kernel_ldbu(*(vucp)addr);
}
__EXTERN_INLINE void wildfire_outb(u8 b, unsigned long addr)
{
addr += WILDFIRE_IO_BIAS;
__kernel_stb(b, *(vucp)addr);
mb();
}
__EXTERN_INLINE u16 wildfire_inw(unsigned long addr)
{
addr += WILDFIRE_IO_BIAS;
return __kernel_ldwu(*(vusp)addr);
}
__EXTERN_INLINE void wildfire_outw(u16 b, unsigned long addr)
{
addr += WILDFIRE_IO_BIAS;
__kernel_stw(b, *(vusp)addr);
mb();
}
__EXTERN_INLINE u32 wildfire_inl(unsigned long addr)
{
addr += WILDFIRE_IO_BIAS;
return *(vuip)addr;
}
__EXTERN_INLINE void wildfire_outl(u32 b, unsigned long addr)
{
addr += WILDFIRE_IO_BIAS;
*(vuip)addr = b;
mb();
}
/* /*
* Memory functions. all accesses are done through linear space. * Memory functions. all accesses are done through linear space.
*/ */
__EXTERN_INLINE unsigned long wildfire_ioremap(unsigned long addr, __EXTERN_INLINE void __iomem *wildfire_ioportmap(unsigned long addr)
unsigned long size
__attribute__((unused)))
{ {
return addr + WILDFIRE_MEM_BIAS; return (void __iomem *)(addr + WILDFIRE_IO_BIAS);
} }
__EXTERN_INLINE void wildfire_iounmap(unsigned long addr) __EXTERN_INLINE void __iomem *wildfire_ioremap(unsigned long addr,
unsigned long size)
{ {
return; return (void __iomem *)(addr + WILDFIRE_MEM_BIAS);
} }
__EXTERN_INLINE int wildfire_is_ioaddr(unsigned long addr) __EXTERN_INLINE int wildfire_is_ioaddr(unsigned long addr)
...@@ -343,87 +293,20 @@ __EXTERN_INLINE int wildfire_is_ioaddr(unsigned long addr) ...@@ -343,87 +293,20 @@ __EXTERN_INLINE int wildfire_is_ioaddr(unsigned long addr)
return addr >= WILDFIRE_BASE; return addr >= WILDFIRE_BASE;
} }
__EXTERN_INLINE u8 wildfire_readb(unsigned long addr) __EXTERN_INLINE int wildfire_is_mmio(const volatile void __iomem *xaddr)
{
return __kernel_ldbu(*(vucp)addr);
}
__EXTERN_INLINE u16 wildfire_readw(unsigned long addr)
{
return __kernel_ldwu(*(vusp)addr);
}
__EXTERN_INLINE u32 wildfire_readl(unsigned long addr)
{
return (*(vuip)addr) & 0xffffffff;
}
__EXTERN_INLINE u64 wildfire_readq(unsigned long addr)
{
return *(vulp)addr;
}
__EXTERN_INLINE void wildfire_writeb(u8 b, unsigned long addr)
{
__kernel_stb(b, *(vucp)addr);
}
__EXTERN_INLINE void wildfire_writew(u16 b, unsigned long addr)
{
__kernel_stw(b, *(vusp)addr);
}
__EXTERN_INLINE void wildfire_writel(u32 b, unsigned long addr)
{
*(vuip)addr = b;
}
__EXTERN_INLINE void wildfire_writeq(u64 b, unsigned long addr)
{ {
*(vulp)addr = b; unsigned long addr = (unsigned long)addr;
return (addr & 0x100000000UL) == 0;
} }
#undef vucp #undef __IO_PREFIX
#undef vusp #define __IO_PREFIX wildfire
#undef vuip #define wildfire_trivial_rw_bw 1
#undef vulp #define wildfire_trivial_rw_lq 1
#define wildfire_trivial_io_bw 1
#ifdef __WANT_IO_DEF #define wildfire_trivial_io_lq 1
#define wildfire_trivial_iounmap 1
#define __inb(p) wildfire_inb((unsigned long)(p)) #include <asm/io_trivial.h>
#define __inw(p) wildfire_inw((unsigned long)(p))
#define __inl(p) wildfire_inl((unsigned long)(p))
#define __outb(x,p) wildfire_outb((x),(unsigned long)(p))
#define __outw(x,p) wildfire_outw((x),(unsigned long)(p))
#define __outl(x,p) wildfire_outl((x),(unsigned long)(p))
#define __readb(a) wildfire_readb((unsigned long)(a))
#define __readw(a) wildfire_readw((unsigned long)(a))
#define __readl(a) wildfire_readl((unsigned long)(a))
#define __readq(a) wildfire_readq((unsigned long)(a))
#define __writeb(x,a) wildfire_writeb((x),(unsigned long)(a))
#define __writew(x,a) wildfire_writew((x),(unsigned long)(a))
#define __writel(x,a) wildfire_writel((x),(unsigned long)(a))
#define __writeq(x,a) wildfire_writeq((x),(unsigned long)(a))
#define __ioremap(a,s) wildfire_ioremap((unsigned long)(a),(s))
#define __iounmap(a) wildfire_iounmap((unsigned long)(a))
#define __is_ioaddr(a) wildfire_is_ioaddr((unsigned long)(a))
#define inb(p) __inb(p)
#define inw(p) __inw(p)
#define inl(p) __inl(p)
#define outb(x,p) __outb((x),(p))
#define outw(x,p) __outw((x),(p))
#define outl(x,p) __outl((x),(p))
#define __raw_readb(a) __readb(a)
#define __raw_readw(a) __readw(a)
#define __raw_readl(a) __readl(a)
#define __raw_readq(a) __readq(a)
#define __raw_writeb(v,a) __writeb((v),(a))
#define __raw_writew(v,a) __writew((v),(a))
#define __raw_writel(v,a) __writel((v),(a))
#define __raw_writeq(v,a) __writeq((v),(a))
#endif /* __WANT_IO_DEF */
#ifdef __IO_EXTERN_INLINE #ifdef __IO_EXTERN_INLINE
#undef __EXTERN_INLINE #undef __EXTERN_INLINE
......
This diff is collapsed.
/* Trivial implementations of basic i/o routines. Assumes that all
of the hard work has been done by ioremap and ioportmap, and that
access to i/o space is linear. */
/* This file may be included multiple times. */
#if IO_CONCAT(__IO_PREFIX,trivial_io_bw)
__EXTERN_INLINE unsigned int
IO_CONCAT(__IO_PREFIX,ioread8)(void __iomem *a)
{
return __kernel_ldbu(*(volatile u8 __force *)a);
}
__EXTERN_INLINE unsigned int
IO_CONCAT(__IO_PREFIX,ioread16)(void __iomem *a)
{
return __kernel_ldwu(*(volatile u16 __force *)a);
}
__EXTERN_INLINE void
IO_CONCAT(__IO_PREFIX,iowrite8)(u8 b, void __iomem *a)
{
__kernel_stb(b, *(volatile u8 __force *)a);
}
__EXTERN_INLINE void
IO_CONCAT(__IO_PREFIX,iowrite16)(u16 b, void __iomem *a)
{
__kernel_stb(b, *(volatile u16 __force *)a);
}
#endif
#if IO_CONCAT(__IO_PREFIX,trivial_io_lq)
__EXTERN_INLINE unsigned int
IO_CONCAT(__IO_PREFIX,ioread32)(void __iomem *a)
{
return *(volatile u32 __force *)a;
}
__EXTERN_INLINE void
IO_CONCAT(__IO_PREFIX,iowrite32)(u32 b, void __iomem *a)
{
*(volatile u32 __force *)a = b;
}
#endif
#if IO_CONCAT(__IO_PREFIX,trivial_rw_bw) == 1
__EXTERN_INLINE u8
IO_CONCAT(__IO_PREFIX,readb)(const volatile void __iomem *a)
{
return __kernel_ldbu(*(const volatile u8 __force *)a);
}
__EXTERN_INLINE u16
IO_CONCAT(__IO_PREFIX,readw)(const volatile void __iomem *a)
{
return __kernel_ldwu(*(const volatile u16 __force *)a);
}
__EXTERN_INLINE void
IO_CONCAT(__IO_PREFIX,writeb)(u8 b, volatile void __iomem *a)
{
__kernel_stb(b, *(volatile u8 __force *)a);
}
__EXTERN_INLINE void
IO_CONCAT(__IO_PREFIX,writew)(u16 b, volatile void __iomem *a)
{
__kernel_stb(b, *(volatile u16 __force *)a);
}
#elif IO_CONCAT(__IO_PREFIX,trivial_rw_bw) == 2
__EXTERN_INLINE u8
IO_CONCAT(__IO_PREFIX,readb)(const volatile void __iomem *a)
{
return IO_CONCAT(__IO_PREFIX,ioread8)((void __iomem *)a);
}
__EXTERN_INLINE u16
IO_CONCAT(__IO_PREFIX,readw)(const volatile void __iomem *a)
{
return IO_CONCAT(__IO_PREFIX,ioread16)((void __iomem *)a);
}
__EXTERN_INLINE void
IO_CONCAT(__IO_PREFIX,writeb)(u8 b, volatile void __iomem *a)
{
IO_CONCAT(__IO_PREFIX,iowrite8)(b, (void __iomem *)a);
}
__EXTERN_INLINE void
IO_CONCAT(__IO_PREFIX,writew)(u16 b, volatile void __iomem *a)
{
IO_CONCAT(__IO_PREFIX,iowrite16)(b, (void __iomem *)a);
}
#endif
#if IO_CONCAT(__IO_PREFIX,trivial_rw_lq) == 1
__EXTERN_INLINE u32
IO_CONCAT(__IO_PREFIX,readl)(const volatile void __iomem *a)
{
return *(const volatile u32 __force *)a;
}
__EXTERN_INLINE u64
IO_CONCAT(__IO_PREFIX,readq)(const volatile void __iomem *a)
{
return *(const volatile u64 __force *)a;
}
__EXTERN_INLINE void
IO_CONCAT(__IO_PREFIX,writel)(u32 b, volatile void __iomem *a)
{
*(volatile u32 __force *)a = b;
}
__EXTERN_INLINE void
IO_CONCAT(__IO_PREFIX,writeq)(u64 b, volatile void __iomem *a)
{
*(volatile u64 __force *)a = b;
}
#endif
#if IO_CONCAT(__IO_PREFIX,trivial_iounmap)
__EXTERN_INLINE void IO_CONCAT(__IO_PREFIX,iounmap)(volatile void __iomem *a)
{
}
#endif
...@@ -200,8 +200,9 @@ __EXTERN_INLINE void jensen_outl(u32 b, unsigned long addr) ...@@ -200,8 +200,9 @@ __EXTERN_INLINE void jensen_outl(u32 b, unsigned long addr)
* Memory functions. * Memory functions.
*/ */
__EXTERN_INLINE u8 jensen_readb(unsigned long addr) __EXTERN_INLINE u8 jensen_readb(const volatile void __iomem *xaddr)
{ {
unsigned long addr = (unsigned long) xaddr;
long result; long result;
jensen_set_hae(addr); jensen_set_hae(addr);
...@@ -211,8 +212,9 @@ __EXTERN_INLINE u8 jensen_readb(unsigned long addr) ...@@ -211,8 +212,9 @@ __EXTERN_INLINE u8 jensen_readb(unsigned long addr)
return 0xffUL & result; return 0xffUL & result;
} }
__EXTERN_INLINE u16 jensen_readw(unsigned long addr) __EXTERN_INLINE u16 jensen_readw(const volatile void __iomem *xaddr)
{ {
unsigned long addr = (unsigned long) xaddr;
long result; long result;
jensen_set_hae(addr); jensen_set_hae(addr);
...@@ -222,15 +224,17 @@ __EXTERN_INLINE u16 jensen_readw(unsigned long addr) ...@@ -222,15 +224,17 @@ __EXTERN_INLINE u16 jensen_readw(unsigned long addr)
return 0xffffUL & result; return 0xffffUL & result;
} }
__EXTERN_INLINE u32 jensen_readl(unsigned long addr) __EXTERN_INLINE u32 jensen_readl(const volatile void __iomem *xaddr)
{ {
unsigned long addr = (unsigned long) xaddr;
jensen_set_hae(addr); jensen_set_hae(addr);
addr &= JENSEN_HAE_MASK; addr &= JENSEN_HAE_MASK;
return *(vuip) ((addr << 7) + EISA_MEM + 0x60); return *(vuip) ((addr << 7) + EISA_MEM + 0x60);
} }
__EXTERN_INLINE u64 jensen_readq(unsigned long addr) __EXTERN_INLINE u64 jensen_readq(const volatile void __iomem *xaddr)
{ {
unsigned long addr = (unsigned long) xaddr;
unsigned long r0, r1; unsigned long r0, r1;
jensen_set_hae(addr); jensen_set_hae(addr);
...@@ -241,29 +245,33 @@ __EXTERN_INLINE u64 jensen_readq(unsigned long addr) ...@@ -241,29 +245,33 @@ __EXTERN_INLINE u64 jensen_readq(unsigned long addr)
return r1 << 32 | r0; return r1 << 32 | r0;
} }
__EXTERN_INLINE void jensen_writeb(u8 b, unsigned long addr) __EXTERN_INLINE void jensen_writeb(u8 b, volatile void __iomem *xaddr)
{ {
unsigned long addr = (unsigned long) xaddr;
jensen_set_hae(addr); jensen_set_hae(addr);
addr &= JENSEN_HAE_MASK; addr &= JENSEN_HAE_MASK;
*(vuip) ((addr << 7) + EISA_MEM + 0x00) = b * 0x01010101; *(vuip) ((addr << 7) + EISA_MEM + 0x00) = b * 0x01010101;
} }
__EXTERN_INLINE void jensen_writew(u16 b, unsigned long addr) __EXTERN_INLINE void jensen_writew(u16 b, volatile void __iomem *xaddr)
{ {
unsigned long addr = (unsigned long) xaddr;
jensen_set_hae(addr); jensen_set_hae(addr);
addr &= JENSEN_HAE_MASK; addr &= JENSEN_HAE_MASK;
*(vuip) ((addr << 7) + EISA_MEM + 0x20) = b * 0x00010001; *(vuip) ((addr << 7) + EISA_MEM + 0x20) = b * 0x00010001;
} }
__EXTERN_INLINE void jensen_writel(u32 b, unsigned long addr) __EXTERN_INLINE void jensen_writel(u32 b, volatile void __iomem *xaddr)
{ {
unsigned long addr = (unsigned long) xaddr;
jensen_set_hae(addr); jensen_set_hae(addr);
addr &= JENSEN_HAE_MASK; addr &= JENSEN_HAE_MASK;
*(vuip) ((addr << 7) + EISA_MEM + 0x60) = b; *(vuip) ((addr << 7) + EISA_MEM + 0x60) = b;
} }
__EXTERN_INLINE void jensen_writeq(u64 b, unsigned long addr) __EXTERN_INLINE void jensen_writeq(u64 b, volatile void __iomem *xaddr)
{ {
unsigned long addr = (unsigned long) xaddr;
jensen_set_hae(addr); jensen_set_hae(addr);
addr &= JENSEN_HAE_MASK; addr &= JENSEN_HAE_MASK;
addr = (addr << 7) + EISA_MEM + 0x60; addr = (addr << 7) + EISA_MEM + 0x60;
...@@ -271,15 +279,15 @@ __EXTERN_INLINE void jensen_writeq(u64 b, unsigned long addr) ...@@ -271,15 +279,15 @@ __EXTERN_INLINE void jensen_writeq(u64 b, unsigned long addr)
*(vuip) (addr + (4 << 7)) = b >> 32; *(vuip) (addr + (4 << 7)) = b >> 32;
} }
__EXTERN_INLINE unsigned long jensen_ioremap(unsigned long addr, __EXTERN_INLINE void __iomem *jensen_ioportmap(unsigned long addr)
unsigned long size)
{ {
return addr; return (void __iomem *)addr;
} }
__EXTERN_INLINE void jensen_iounmap(unsigned long addr) __EXTERN_INLINE void __iomem *jensen_ioremap(unsigned long addr,
unsigned long size)
{ {
return; return (void __iomem *)(addr + 0x100000000ul);
} }
__EXTERN_INLINE int jensen_is_ioaddr(unsigned long addr) __EXTERN_INLINE int jensen_is_ioaddr(unsigned long addr)
...@@ -287,39 +295,46 @@ __EXTERN_INLINE int jensen_is_ioaddr(unsigned long addr) ...@@ -287,39 +295,46 @@ __EXTERN_INLINE int jensen_is_ioaddr(unsigned long addr)
return (long)addr >= 0; return (long)addr >= 0;
} }
#undef vuip __EXTERN_INLINE int jensen_is_mmio(const volatile void __iomem *addr)
{
return (unsigned long)addr >= 0x100000000ul;
}
#ifdef __WANT_IO_DEF /* New-style ioread interface. All the routines are so ugly for Jensen
that it doesn't make sense to merge them. */
#define __inb jensen_inb
#define __inw jensen_inw #define IOPORT(OS, NS) \
#define __inl jensen_inl __EXTERN_INLINE unsigned int jensen_ioread##NS(void __iomem *xaddr) \
#define __outb jensen_outb { \
#define __outw jensen_outw if (jensen_is_mmio(xaddr)) \
#define __outl jensen_outl return jensen_read##OS(xaddr - 0x100000000ul); \
#define __readb jensen_readb else \
#define __readw jensen_readw return jensen_in##OS((unsigned long)xaddr); \
#define __writeb jensen_writeb } \
#define __writew jensen_writew __EXTERN_INLINE void jensen_iowrite##NS(u##NS b, void __iomem *xaddr) \
#define __readl jensen_readl { \
#define __readq jensen_readq if (jensen_is_mmio(xaddr)) \
#define __writel jensen_writel jensen_write##OS(b, xaddr - 0x100000000ul); \
#define __writeq jensen_writeq else \
#define __ioremap jensen_ioremap jensen_out##OS(b, (unsigned long)xaddr); \
#define __iounmap(a) jensen_iounmap((unsigned long)a) }
#define __is_ioaddr jensen_is_ioaddr
/* IOPORT(b, 8)
* The above have so much overhead that it probably doesn't make IOPORT(w, 16)
* sense to have them inlined (better icache behaviour). IOPORT(l, 32)
*/
#define inb(port) \
(__builtin_constant_p((port))?__inb(port):_inb(port))
#define outb(x, port) \ #undef IOPORT
(__builtin_constant_p((port))?__outb((x),(port)):_outb((x),(port)))
#undef vuip
#endif /* __WANT_IO_DEF */ #undef __IO_PREFIX
#define __IO_PREFIX jensen
#define jensen_trivial_rw_bw 0
#define jensen_trivial_rw_lq 0
#define jensen_trivial_io_bw 0
#define jensen_trivial_io_lq 0
#define jensen_trivial_iounmap 1
#include <asm/io_trivial.h>
#ifdef __IO_EXTERN_INLINE #ifdef __IO_EXTERN_INLINE
#undef __EXTERN_INLINE #undef __EXTERN_INLINE
......
...@@ -45,27 +45,29 @@ struct alpha_machine_vector ...@@ -45,27 +45,29 @@ struct alpha_machine_vector
void (*mv_pci_tbi)(struct pci_controller *hose, void (*mv_pci_tbi)(struct pci_controller *hose,
dma_addr_t start, dma_addr_t end); dma_addr_t start, dma_addr_t end);
u8 (*mv_inb)(unsigned long); unsigned int (*mv_ioread8)(void __iomem *);
u16 (*mv_inw)(unsigned long); unsigned int (*mv_ioread16)(void __iomem *);
u32 (*mv_inl)(unsigned long); unsigned int (*mv_ioread32)(void __iomem *);
void (*mv_outb)(u8, unsigned long); void (*mv_iowrite8)(u8, void __iomem *);
void (*mv_outw)(u16, unsigned long); void (*mv_iowrite16)(u16, void __iomem *);
void (*mv_outl)(u32, unsigned long); void (*mv_iowrite32)(u32, void __iomem *);
u8 (*mv_readb)(unsigned long); u8 (*mv_readb)(const volatile void __iomem *);
u16 (*mv_readw)(unsigned long); u16 (*mv_readw)(const volatile void __iomem *);
u32 (*mv_readl)(unsigned long); u32 (*mv_readl)(const volatile void __iomem *);
u64 (*mv_readq)(unsigned long); u64 (*mv_readq)(const volatile void __iomem *);
void (*mv_writeb)(u8, unsigned long); void (*mv_writeb)(u8, volatile void __iomem *);
void (*mv_writew)(u16, unsigned long); void (*mv_writew)(u16, volatile void __iomem *);
void (*mv_writel)(u32, unsigned long); void (*mv_writel)(u32, volatile void __iomem *);
void (*mv_writeq)(u64, unsigned long); void (*mv_writeq)(u64, volatile void __iomem *);
unsigned long (*mv_ioremap)(unsigned long, unsigned long); void __iomem *(*mv_ioportmap)(unsigned long);
void (*mv_iounmap)(unsigned long); void __iomem *(*mv_ioremap)(unsigned long, unsigned long);
void (*mv_iounmap)(volatile void __iomem *);
int (*mv_is_ioaddr)(unsigned long); int (*mv_is_ioaddr)(unsigned long);
int (*mv_is_mmio)(const volatile void __iomem *);
void (*mv_switch_mm)(struct mm_struct *, struct mm_struct *, void (*mv_switch_mm)(struct mm_struct *, struct mm_struct *,
struct task_struct *); struct task_struct *);
......
...@@ -10,6 +10,7 @@ ...@@ -10,6 +10,7 @@
#include <linux/config.h> #include <linux/config.h>
#include <asm/system.h> #include <asm/system.h>
#include <asm/machvec.h> #include <asm/machvec.h>
#include <asm/compiler.h>
/* /*
* Force a context reload. This is needed when we change the page * Force a context reload. This is needed when we change the page
......
...@@ -95,7 +95,7 @@ static inline int _raw_spin_trylock(spinlock_t *lock) ...@@ -95,7 +95,7 @@ static inline int _raw_spin_trylock(spinlock_t *lock)
/***********************************************************/ /***********************************************************/
typedef struct { typedef struct {
volatile int write_lock:1, read_counter:31; volatile unsigned int write_lock:1, read_counter:31;
} /*__attribute__((aligned(32)))*/ rwlock_t; } /*__attribute__((aligned(32)))*/ rwlock_t;
#define RW_LOCK_UNLOCKED (rwlock_t) { 0, 0 } #define RW_LOCK_UNLOCKED (rwlock_t) { 0, 0 }
...@@ -124,7 +124,7 @@ static inline void _raw_write_lock(rwlock_t * lock) ...@@ -124,7 +124,7 @@ static inline void _raw_write_lock(rwlock_t * lock)
" br 1b\n" " br 1b\n"
".previous" ".previous"
: "=m" (*lock), "=&r" (regx) : "=m" (*lock), "=&r" (regx)
: "0" (*lock) : "memory"); : "m" (*lock) : "memory");
} }
static inline void _raw_read_lock(rwlock_t * lock) static inline void _raw_read_lock(rwlock_t * lock)
......
...@@ -55,9 +55,9 @@ ...@@ -55,9 +55,9 @@
*/ */
struct el_common { struct el_common {
unsigned int size; /* size in bytes of logout area */ unsigned int size; /* size in bytes of logout area */
int sbz1 : 30; /* should be zero */ unsigned int sbz1 : 30; /* should be zero */
int err2 : 1; /* second error */ unsigned int err2 : 1; /* second error */
int retry : 1; /* retry flag */ unsigned int retry : 1; /* retry flag */
unsigned int proc_offset; /* processor-specific offset */ unsigned int proc_offset; /* processor-specific offset */
unsigned int sys_offset; /* system-specific offset */ unsigned int sys_offset; /* system-specific offset */
unsigned int code; /* machine check code */ unsigned int code; /* machine check code */
......
...@@ -3,6 +3,7 @@ ...@@ -3,6 +3,7 @@
#include <linux/config.h> #include <linux/config.h>
#include <linux/mm.h> #include <linux/mm.h>
#include <asm/compiler.h>
#ifndef __EXTERN_INLINE #ifndef __EXTERN_INLINE
#define __EXTERN_INLINE extern inline #define __EXTERN_INLINE extern inline
......
...@@ -15,24 +15,24 @@ ...@@ -15,24 +15,24 @@
extern inline void scr_writew(u16 val, volatile u16 *addr) extern inline void scr_writew(u16 val, volatile u16 *addr)
{ {
if (__is_ioaddr((unsigned long) addr)) if (__is_ioaddr(addr))
__raw_writew(val, (unsigned long) addr); __raw_writew(val, (volatile u16 __iomem *) addr);
else else
*addr = val; *addr = val;
} }
extern inline u16 scr_readw(volatile const u16 *addr) extern inline u16 scr_readw(volatile const u16 *addr)
{ {
if (__is_ioaddr((unsigned long) addr)) if (__is_ioaddr(addr))
return __raw_readw((unsigned long) addr); return __raw_readw((volatile const u16 __iomem *) addr);
else else
return *addr; return *addr;
} }
extern inline void scr_memsetw(u16 *s, u16 c, unsigned int count) extern inline void scr_memsetw(u16 *s, u16 c, unsigned int count)
{ {
if (__is_ioaddr((unsigned long) s)) if (__is_ioaddr(s))
memsetw_io(s, c, count); memsetw_io((u16 __iomem *) s, c, count);
else else
memsetw(s, c, count); memsetw(s, c, count);
} }
...@@ -43,9 +43,9 @@ extern void scr_memcpyw(u16 *d, const u16 *s, unsigned int count); ...@@ -43,9 +43,9 @@ extern void scr_memcpyw(u16 *d, const u16 *s, unsigned int count);
/* ??? These are currently only used for downloading character sets. As /* ??? These are currently only used for downloading character sets. As
such, they don't need memory barriers. Is this all they are intended such, they don't need memory barriers. Is this all they are intended
to be used for? */ to be used for? */
#define vga_readb readb #define vga_readb(a) readb((u8 __iomem *)(a))
#define vga_writeb writeb #define vga_writeb(v,a) writeb(v, (u8 __iomem *)(a))
#define VGA_MAP_MEM(x) ((unsigned long) ioremap((x), 0)) #define VGA_MAP_MEM(x) ((unsigned long) ioremap(x, 0))
#endif #endif
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment