Commit b3967dc5 authored by Linus Torvalds's avatar Linus Torvalds

Merge branch 'release' of git://git.kernel.org/pub/scm/linux/kernel/git/aegl/linux-2.6

* 'release' of git://git.kernel.org/pub/scm/linux/kernel/git/aegl/linux-2.6:
  [IA64] Prefetch mmap_sem in ia64_do_page_fault()
  [IA64] Failure to resume after INIT in user space
  [IA64] Pass more data to the MCA/INIT notify_die hooks
  [IA64] always map VGA framebuffer UC, even if it supports WB
  [IA64] fix bug in ia64 __mutex_fastpath_trylock
  [IA64] for_each_possible_cpu: ia64
  [IA64] update HP CSR space discovery via ACPI
  [IA64] Wire up new syscalls {set,get}_robust_list
  [IA64] 'msg' may be used uninitialized in xpc_initiate_allocate()
  [IA64] Wire up new syscall sync_file_range()
parents cde227af 0ffe9849
/* /*
* arch/ia64/kernel/acpi-ext.c * (c) Copyright 2003, 2006 Hewlett-Packard Development Company, L.P.
* Alex Williamson <alex.williamson@hp.com>
* Bjorn Helgaas <bjorn.helgaas@hp.com>
* *
* Copyright (C) 2003 Hewlett-Packard * This program is free software; you can redistribute it and/or modify
* Copyright (C) Alex Williamson * it under the terms of the GNU General Public License version 2 as
* Copyright (C) Bjorn Helgaas * published by the Free Software Foundation.
*
* Vendor specific extensions to ACPI.
*/ */
#include <linux/config.h> #include <linux/config.h>
#include <linux/module.h> #include <linux/module.h>
#include <linux/types.h> #include <linux/types.h>
#include <linux/acpi.h> #include <linux/acpi.h>
#include <linux/efi.h>
#include <asm/acpi-ext.h> #include <asm/acpi-ext.h>
struct acpi_vendor_descriptor { /*
u8 guid_id; * Device CSRs that do not appear in PCI config space should be described
efi_guid_t guid; * via ACPI. This would normally be done with Address Space Descriptors
}; * marked as "consumer-only," but old versions of Windows and Linux ignore
* the producer/consumer flag, so HP invented a vendor-defined resource to
* describe the location and size of CSR space.
*/
struct acpi_vendor_info { struct acpi_vendor_uuid hp_ccsr_uuid = {
struct acpi_vendor_descriptor *descriptor; .subtype = 2,
u8 *data; .data = { 0xf9, 0xad, 0xe9, 0x69, 0x4f, 0x92, 0x5f, 0xab, 0xf6, 0x4a,
u32 length; 0x24, 0xd2, 0x01, 0x37, 0x0e, 0xad },
}; };
acpi_status static acpi_status hp_ccsr_locate(acpi_handle obj, u64 *base, u64 *length)
acpi_vendor_resource_match(struct acpi_resource *resource, void *context)
{ {
struct acpi_vendor_info *info = (struct acpi_vendor_info *)context; acpi_status status;
struct acpi_resource_vendor *vendor; struct acpi_buffer buffer = { ACPI_ALLOCATE_BUFFER, NULL };
struct acpi_vendor_descriptor *descriptor; struct acpi_resource *resource;
u32 byte_length; struct acpi_resource_vendor_typed *vendor;
if (resource->type != ACPI_RESOURCE_TYPE_VENDOR) status = acpi_get_vendor_resource(obj, METHOD_NAME__CRS, &hp_ccsr_uuid,
return AE_OK; &buffer);
vendor = (struct acpi_resource_vendor *)&resource->data; resource = buffer.pointer;
descriptor = (struct acpi_vendor_descriptor *)vendor->byte_data; vendor = &resource->data.vendor_typed;
if (vendor->byte_length <= sizeof(*info->descriptor) ||
descriptor->guid_id != info->descriptor->guid_id ||
efi_guidcmp(descriptor->guid, info->descriptor->guid))
return AE_OK;
byte_length = vendor->byte_length - sizeof(struct acpi_vendor_descriptor); if (ACPI_FAILURE(status) || vendor->byte_length < 16) {
info->data = acpi_os_allocate(byte_length); status = AE_NOT_FOUND;
if (!info->data) goto exit;
return AE_NO_MEMORY; }
memcpy(info->data, memcpy(base, vendor->byte_data, sizeof(*base));
vendor->byte_data + sizeof(struct acpi_vendor_descriptor), memcpy(length, vendor->byte_data + 8, sizeof(*length));
byte_length);
info->length = byte_length; exit:
return AE_CTRL_TERMINATE; acpi_os_free(buffer.pointer);
return status;
} }
acpi_status struct csr_space {
acpi_find_vendor_resource(acpi_handle obj, struct acpi_vendor_descriptor * id, u64 base;
u8 ** data, u32 * byte_length) u64 length;
};
static acpi_status find_csr_space(struct acpi_resource *resource, void *data)
{ {
struct acpi_vendor_info info; struct csr_space *space = data;
struct acpi_resource_address64 addr;
acpi_status status;
info.descriptor = id; status = acpi_resource_to_address64(resource, &addr);
info.data = NULL; if (ACPI_SUCCESS(status) &&
addr.resource_type == ACPI_MEMORY_RANGE &&
addr.address_length &&
addr.producer_consumer == ACPI_CONSUMER) {
space->base = addr.minimum;
space->length = addr.address_length;
return AE_CTRL_TERMINATE;
}
return AE_OK; /* keep looking */
}
static acpi_status hp_crs_locate(acpi_handle obj, u64 *base, u64 *length)
{
struct csr_space space = { 0, 0 };
acpi_walk_resources(obj, METHOD_NAME__CRS, acpi_vendor_resource_match, acpi_walk_resources(obj, METHOD_NAME__CRS, find_csr_space, &space);
&info); if (!space.length)
if (!info.data)
return AE_NOT_FOUND; return AE_NOT_FOUND;
*data = info.data; *base = space.base;
*byte_length = info.length; *length = space.length;
return AE_OK; return AE_OK;
} }
struct acpi_vendor_descriptor hp_ccsr_descriptor = { acpi_status hp_acpi_csr_space(acpi_handle obj, u64 *csr_base, u64 *csr_length)
.guid_id = 2,
.guid =
EFI_GUID(0x69e9adf9, 0x924f, 0xab5f, 0xf6, 0x4a, 0x24, 0xd2, 0x01,
0x37, 0x0e, 0xad)
};
acpi_status hp_acpi_csr_space(acpi_handle obj, u64 * csr_base, u64 * csr_length)
{ {
acpi_status status; acpi_status status;
u8 *data;
u32 length;
status = status = hp_ccsr_locate(obj, csr_base, csr_length);
acpi_find_vendor_resource(obj, &hp_ccsr_descriptor, &data, &length); if (ACPI_SUCCESS(status))
return status;
if (ACPI_FAILURE(status) || length != 16) return hp_crs_locate(obj, csr_base, csr_length);
return AE_NOT_FOUND;
memcpy(csr_base, data, sizeof(*csr_base));
memcpy(csr_length, data + 8, sizeof(*csr_length));
acpi_os_free(data);
return AE_OK;
} }
EXPORT_SYMBOL(hp_acpi_csr_space); EXPORT_SYMBOL(hp_acpi_csr_space);
...@@ -1606,5 +1606,8 @@ sys_call_table: ...@@ -1606,5 +1606,8 @@ sys_call_table:
data8 sys_ni_syscall // 1295 reserved for ppoll data8 sys_ni_syscall // 1295 reserved for ppoll
data8 sys_unshare data8 sys_unshare
data8 sys_splice data8 sys_splice
data8 sys_set_robust_list
data8 sys_get_robust_list
data8 sys_sync_file_range // 1300
.org sys_call_table + 8*NR_syscalls // guard against failures to increase NR_syscalls .org sys_call_table + 8*NR_syscalls // guard against failures to increase NR_syscalls
...@@ -581,10 +581,12 @@ ia64_mca_rendez_int_handler(int rendez_irq, void *arg, struct pt_regs *regs) ...@@ -581,10 +581,12 @@ ia64_mca_rendez_int_handler(int rendez_irq, void *arg, struct pt_regs *regs)
{ {
unsigned long flags; unsigned long flags;
int cpu = smp_processor_id(); int cpu = smp_processor_id();
struct ia64_mca_notify_die nd =
{ .sos = NULL, .monarch_cpu = &monarch_cpu };
/* Mask all interrupts */ /* Mask all interrupts */
local_irq_save(flags); local_irq_save(flags);
if (notify_die(DIE_MCA_RENDZVOUS_ENTER, "MCA", regs, 0, 0, 0) if (notify_die(DIE_MCA_RENDZVOUS_ENTER, "MCA", regs, (long)&nd, 0, 0)
== NOTIFY_STOP) == NOTIFY_STOP)
ia64_mca_spin(__FUNCTION__); ia64_mca_spin(__FUNCTION__);
...@@ -594,7 +596,7 @@ ia64_mca_rendez_int_handler(int rendez_irq, void *arg, struct pt_regs *regs) ...@@ -594,7 +596,7 @@ ia64_mca_rendez_int_handler(int rendez_irq, void *arg, struct pt_regs *regs)
*/ */
ia64_sal_mc_rendez(); ia64_sal_mc_rendez();
if (notify_die(DIE_MCA_RENDZVOUS_PROCESS, "MCA", regs, 0, 0, 0) if (notify_die(DIE_MCA_RENDZVOUS_PROCESS, "MCA", regs, (long)&nd, 0, 0)
== NOTIFY_STOP) == NOTIFY_STOP)
ia64_mca_spin(__FUNCTION__); ia64_mca_spin(__FUNCTION__);
...@@ -602,7 +604,7 @@ ia64_mca_rendez_int_handler(int rendez_irq, void *arg, struct pt_regs *regs) ...@@ -602,7 +604,7 @@ ia64_mca_rendez_int_handler(int rendez_irq, void *arg, struct pt_regs *regs)
while (monarch_cpu != -1) while (monarch_cpu != -1)
cpu_relax(); /* spin until monarch leaves */ cpu_relax(); /* spin until monarch leaves */
if (notify_die(DIE_MCA_RENDZVOUS_LEAVE, "MCA", regs, 0, 0, 0) if (notify_die(DIE_MCA_RENDZVOUS_LEAVE, "MCA", regs, (long)&nd, 0, 0)
== NOTIFY_STOP) == NOTIFY_STOP)
ia64_mca_spin(__FUNCTION__); ia64_mca_spin(__FUNCTION__);
...@@ -1023,6 +1025,8 @@ ia64_mca_handler(struct pt_regs *regs, struct switch_stack *sw, ...@@ -1023,6 +1025,8 @@ ia64_mca_handler(struct pt_regs *regs, struct switch_stack *sw,
&sos->proc_state_param; &sos->proc_state_param;
int recover, cpu = smp_processor_id(); int recover, cpu = smp_processor_id();
task_t *previous_current; task_t *previous_current;
struct ia64_mca_notify_die nd =
{ .sos = sos, .monarch_cpu = &monarch_cpu };
oops_in_progress = 1; /* FIXME: make printk NMI/MCA/INIT safe */ oops_in_progress = 1; /* FIXME: make printk NMI/MCA/INIT safe */
console_loglevel = 15; /* make sure printks make it to console */ console_loglevel = 15; /* make sure printks make it to console */
...@@ -1031,7 +1035,7 @@ ia64_mca_handler(struct pt_regs *regs, struct switch_stack *sw, ...@@ -1031,7 +1035,7 @@ ia64_mca_handler(struct pt_regs *regs, struct switch_stack *sw,
previous_current = ia64_mca_modify_original_stack(regs, sw, sos, "MCA"); previous_current = ia64_mca_modify_original_stack(regs, sw, sos, "MCA");
monarch_cpu = cpu; monarch_cpu = cpu;
if (notify_die(DIE_MCA_MONARCH_ENTER, "MCA", regs, 0, 0, 0) if (notify_die(DIE_MCA_MONARCH_ENTER, "MCA", regs, (long)&nd, 0, 0)
== NOTIFY_STOP) == NOTIFY_STOP)
ia64_mca_spin(__FUNCTION__); ia64_mca_spin(__FUNCTION__);
ia64_wait_for_slaves(cpu); ia64_wait_for_slaves(cpu);
...@@ -1043,7 +1047,7 @@ ia64_mca_handler(struct pt_regs *regs, struct switch_stack *sw, ...@@ -1043,7 +1047,7 @@ ia64_mca_handler(struct pt_regs *regs, struct switch_stack *sw,
* spinning in SAL does not work. * spinning in SAL does not work.
*/ */
ia64_mca_wakeup_all(); ia64_mca_wakeup_all();
if (notify_die(DIE_MCA_MONARCH_PROCESS, "MCA", regs, 0, 0, 0) if (notify_die(DIE_MCA_MONARCH_PROCESS, "MCA", regs, (long)&nd, 0, 0)
== NOTIFY_STOP) == NOTIFY_STOP)
ia64_mca_spin(__FUNCTION__); ia64_mca_spin(__FUNCTION__);
...@@ -1064,7 +1068,7 @@ ia64_mca_handler(struct pt_regs *regs, struct switch_stack *sw, ...@@ -1064,7 +1068,7 @@ ia64_mca_handler(struct pt_regs *regs, struct switch_stack *sw,
ia64_sal_clear_state_info(SAL_INFO_TYPE_MCA); ia64_sal_clear_state_info(SAL_INFO_TYPE_MCA);
sos->os_status = IA64_MCA_CORRECTED; sos->os_status = IA64_MCA_CORRECTED;
} }
if (notify_die(DIE_MCA_MONARCH_LEAVE, "MCA", regs, 0, 0, recover) if (notify_die(DIE_MCA_MONARCH_LEAVE, "MCA", regs, (long)&nd, 0, recover)
== NOTIFY_STOP) == NOTIFY_STOP)
ia64_mca_spin(__FUNCTION__); ia64_mca_spin(__FUNCTION__);
...@@ -1351,10 +1355,14 @@ ia64_init_handler(struct pt_regs *regs, struct switch_stack *sw, ...@@ -1351,10 +1355,14 @@ ia64_init_handler(struct pt_regs *regs, struct switch_stack *sw,
static atomic_t monarchs; static atomic_t monarchs;
task_t *previous_current; task_t *previous_current;
int cpu = smp_processor_id(); int cpu = smp_processor_id();
struct ia64_mca_notify_die nd =
{ .sos = sos, .monarch_cpu = &monarch_cpu };
oops_in_progress = 1; /* FIXME: make printk NMI/MCA/INIT safe */ oops_in_progress = 1; /* FIXME: make printk NMI/MCA/INIT safe */
console_loglevel = 15; /* make sure printks make it to console */ console_loglevel = 15; /* make sure printks make it to console */
(void) notify_die(DIE_INIT_ENTER, "INIT", regs, (long)&nd, 0, 0);
printk(KERN_INFO "Entered OS INIT handler. PSP=%lx cpu=%d monarch=%ld\n", printk(KERN_INFO "Entered OS INIT handler. PSP=%lx cpu=%d monarch=%ld\n",
sos->proc_state_param, cpu, sos->monarch); sos->proc_state_param, cpu, sos->monarch);
salinfo_log_wakeup(SAL_INFO_TYPE_INIT, NULL, 0, 0); salinfo_log_wakeup(SAL_INFO_TYPE_INIT, NULL, 0, 0);
...@@ -1390,15 +1398,15 @@ ia64_init_handler(struct pt_regs *regs, struct switch_stack *sw, ...@@ -1390,15 +1398,15 @@ ia64_init_handler(struct pt_regs *regs, struct switch_stack *sw,
ia64_mc_info.imi_rendez_checkin[cpu] = IA64_MCA_RENDEZ_CHECKIN_INIT; ia64_mc_info.imi_rendez_checkin[cpu] = IA64_MCA_RENDEZ_CHECKIN_INIT;
while (monarch_cpu == -1) while (monarch_cpu == -1)
cpu_relax(); /* spin until monarch enters */ cpu_relax(); /* spin until monarch enters */
if (notify_die(DIE_INIT_SLAVE_ENTER, "INIT", regs, 0, 0, 0) if (notify_die(DIE_INIT_SLAVE_ENTER, "INIT", regs, (long)&nd, 0, 0)
== NOTIFY_STOP) == NOTIFY_STOP)
ia64_mca_spin(__FUNCTION__); ia64_mca_spin(__FUNCTION__);
if (notify_die(DIE_INIT_SLAVE_PROCESS, "INIT", regs, 0, 0, 0) if (notify_die(DIE_INIT_SLAVE_PROCESS, "INIT", regs, (long)&nd, 0, 0)
== NOTIFY_STOP) == NOTIFY_STOP)
ia64_mca_spin(__FUNCTION__); ia64_mca_spin(__FUNCTION__);
while (monarch_cpu != -1) while (monarch_cpu != -1)
cpu_relax(); /* spin until monarch leaves */ cpu_relax(); /* spin until monarch leaves */
if (notify_die(DIE_INIT_SLAVE_LEAVE, "INIT", regs, 0, 0, 0) if (notify_die(DIE_INIT_SLAVE_LEAVE, "INIT", regs, (long)&nd, 0, 0)
== NOTIFY_STOP) == NOTIFY_STOP)
ia64_mca_spin(__FUNCTION__); ia64_mca_spin(__FUNCTION__);
printk("Slave on cpu %d returning to normal service.\n", cpu); printk("Slave on cpu %d returning to normal service.\n", cpu);
...@@ -1409,7 +1417,7 @@ ia64_init_handler(struct pt_regs *regs, struct switch_stack *sw, ...@@ -1409,7 +1417,7 @@ ia64_init_handler(struct pt_regs *regs, struct switch_stack *sw,
} }
monarch_cpu = cpu; monarch_cpu = cpu;
if (notify_die(DIE_INIT_MONARCH_ENTER, "INIT", regs, 0, 0, 0) if (notify_die(DIE_INIT_MONARCH_ENTER, "INIT", regs, (long)&nd, 0, 0)
== NOTIFY_STOP) == NOTIFY_STOP)
ia64_mca_spin(__FUNCTION__); ia64_mca_spin(__FUNCTION__);
...@@ -1426,10 +1434,10 @@ ia64_init_handler(struct pt_regs *regs, struct switch_stack *sw, ...@@ -1426,10 +1434,10 @@ ia64_init_handler(struct pt_regs *regs, struct switch_stack *sw,
* to default_monarch_init_process() above and just print all the * to default_monarch_init_process() above and just print all the
* tasks. * tasks.
*/ */
if (notify_die(DIE_INIT_MONARCH_PROCESS, "INIT", regs, 0, 0, 0) if (notify_die(DIE_INIT_MONARCH_PROCESS, "INIT", regs, (long)&nd, 0, 0)
== NOTIFY_STOP) == NOTIFY_STOP)
ia64_mca_spin(__FUNCTION__); ia64_mca_spin(__FUNCTION__);
if (notify_die(DIE_INIT_MONARCH_LEAVE, "INIT", regs, 0, 0, 0) if (notify_die(DIE_INIT_MONARCH_LEAVE, "INIT", regs, (long)&nd, 0, 0)
== NOTIFY_STOP) == NOTIFY_STOP)
ia64_mca_spin(__FUNCTION__); ia64_mca_spin(__FUNCTION__);
printk("\nINIT dump complete. Monarch on cpu %d returning to normal service.\n", cpu); printk("\nINIT dump complete. Monarch on cpu %d returning to normal service.\n", cpu);
...@@ -1631,6 +1639,7 @@ ia64_mca_init(void) ...@@ -1631,6 +1639,7 @@ ia64_mca_init(void)
printk(KERN_INFO "Increasing MCA rendezvous timeout from " printk(KERN_INFO "Increasing MCA rendezvous timeout from "
"%ld to %ld milliseconds\n", timeout, isrv.v0); "%ld to %ld milliseconds\n", timeout, isrv.v0);
timeout = isrv.v0; timeout = isrv.v0;
(void) notify_die(DIE_MCA_NEW_TIMEOUT, "MCA", NULL, timeout, 0, 0);
continue; continue;
} }
printk(KERN_ERR "Failed to register rendezvous interrupt " printk(KERN_ERR "Failed to register rendezvous interrupt "
......
...@@ -827,7 +827,7 @@ ia64_state_restore: ...@@ -827,7 +827,7 @@ ia64_state_restore:
ld8 r9=[temp2],16 // sal_gp ld8 r9=[temp2],16 // sal_gp
;; ;;
ld8 r22=[temp1],16 // pal_min_state, virtual ld8 r22=[temp1],16 // pal_min_state, virtual
ld8 r21=[temp2],16 // prev_IA64_KR_CURRENT ld8 r13=[temp2],16 // prev_IA64_KR_CURRENT
;; ;;
ld8 r16=[temp1],16 // prev_IA64_KR_CURRENT_STACK ld8 r16=[temp1],16 // prev_IA64_KR_CURRENT_STACK
ld8 r20=[temp2],16 // prev_task ld8 r20=[temp2],16 // prev_task
...@@ -848,7 +848,7 @@ ia64_state_restore: ...@@ -848,7 +848,7 @@ ia64_state_restore:
mov cr.iim=temp3 mov cr.iim=temp3
mov cr.iha=temp4 mov cr.iha=temp4
dep r22=0,r22,62,1 // pal_min_state, physical, uncached dep r22=0,r22,62,1 // pal_min_state, physical, uncached
mov IA64_KR(CURRENT)=r21 mov IA64_KR(CURRENT)=r13
ld8 r8=[temp1] // os_status ld8 r8=[temp1] // os_status
ld8 r10=[temp2] // context ld8 r10=[temp2] // context
...@@ -856,7 +856,7 @@ ia64_state_restore: ...@@ -856,7 +856,7 @@ ia64_state_restore:
* avoid any dependencies on the algorithm in ia64_switch_to(), just * avoid any dependencies on the algorithm in ia64_switch_to(), just
* purge any existing CURRENT_STACK mapping and insert the new one. * purge any existing CURRENT_STACK mapping and insert the new one.
* *
* r16 contains prev_IA64_KR_CURRENT_STACK, r21 contains * r16 contains prev_IA64_KR_CURRENT_STACK, r13 contains
* prev_IA64_KR_CURRENT, these values may have been changed by the C * prev_IA64_KR_CURRENT, these values may have been changed by the C
* code. Do not use r8, r9, r10, r22, they contain values ready for * code. Do not use r8, r9, r10, r22, they contain values ready for
* the return to SAL. * the return to SAL.
...@@ -873,7 +873,7 @@ ia64_state_restore: ...@@ -873,7 +873,7 @@ ia64_state_restore:
;; ;;
srlz.d srlz.d
extr.u r19=r21,61,3 // r21 = prev_IA64_KR_CURRENT extr.u r19=r13,61,3 // r13 = prev_IA64_KR_CURRENT
shl r20=r16,IA64_GRANULE_SHIFT // r16 = prev_IA64_KR_CURRENT_STACK shl r20=r16,IA64_GRANULE_SHIFT // r16 = prev_IA64_KR_CURRENT_STACK
movl r21=PAGE_KERNEL // page properties movl r21=PAGE_KERNEL // page properties
;; ;;
...@@ -883,7 +883,7 @@ ia64_state_restore: ...@@ -883,7 +883,7 @@ ia64_state_restore:
(p6) br.spnt 1f // the dreaded cpu 0 idle task in region 5:( (p6) br.spnt 1f // the dreaded cpu 0 idle task in region 5:(
;; ;;
mov cr.itir=r18 mov cr.itir=r18
mov cr.ifa=r21 mov cr.ifa=r13
mov r20=IA64_TR_CURRENT_STACK mov r20=IA64_TR_CURRENT_STACK
;; ;;
itr.d dtr[r20]=r21 itr.d dtr[r20]=r21
......
...@@ -947,7 +947,7 @@ void ...@@ -947,7 +947,7 @@ void
percpu_modcopy (void *pcpudst, const void *src, unsigned long size) percpu_modcopy (void *pcpudst, const void *src, unsigned long size)
{ {
unsigned int i; unsigned int i;
for_each_cpu(i) { for_each_possible_cpu(i) {
memcpy(pcpudst + __per_cpu_offset[i], src, size); memcpy(pcpudst + __per_cpu_offset[i], src, size);
} }
} }
......
...@@ -60,6 +60,9 @@ ia64_do_page_fault (unsigned long address, unsigned long isr, struct pt_regs *re ...@@ -60,6 +60,9 @@ ia64_do_page_fault (unsigned long address, unsigned long isr, struct pt_regs *re
struct siginfo si; struct siginfo si;
unsigned long mask; unsigned long mask;
/* mmap_sem is performance critical.... */
prefetchw(&mm->mmap_sem);
/* /*
* If we're in an interrupt or have no user context, we must not take the fault.. * If we're in an interrupt or have no user context, we must not take the fault..
*/ */
......
...@@ -1831,7 +1831,7 @@ xpc_initiate_allocate(partid_t partid, int ch_number, u32 flags, void **payload) ...@@ -1831,7 +1831,7 @@ xpc_initiate_allocate(partid_t partid, int ch_number, u32 flags, void **payload)
{ {
struct xpc_partition *part = &xpc_partitions[partid]; struct xpc_partition *part = &xpc_partitions[partid];
enum xpc_retval ret = xpcUnknownReason; enum xpc_retval ret = xpcUnknownReason;
struct xpc_msg *msg; struct xpc_msg *msg = NULL;
DBUG_ON(partid <= 0 || partid >= XP_MAX_PARTITIONS); DBUG_ON(partid <= 0 || partid >= XP_MAX_PARTITIONS);
......
/* /*
* ia64/platform/hp/common/hp_acpi.h * (c) Copyright 2003, 2006 Hewlett-Packard Development Company, L.P.
* Alex Williamson <alex.williamson@hp.com>
* Bjorn Helgaas <bjorn.helgaas@hp.com>
* *
* Copyright (C) 2003 Hewlett-Packard * This program is free software; you can redistribute it and/or modify
* Copyright (C) Alex Williamson * it under the terms of the GNU General Public License version 2 as
* Copyright (C) Bjorn Helgaas * published by the Free Software Foundation.
* *
* Vendor specific extensions to ACPI. * Vendor specific extensions to ACPI.
*/ */
#ifndef _ASM_IA64_ACPI_EXT_H #ifndef _ASM_IA64_ACPI_EXT_H
#define _ASM_IA64_ACPI_EXT_H #define _ASM_IA64_ACPI_EXT_H
......
...@@ -58,6 +58,8 @@ enum die_val { ...@@ -58,6 +58,8 @@ enum die_val {
DIE_MCA_RENDZVOUS_ENTER, DIE_MCA_RENDZVOUS_ENTER,
DIE_MCA_RENDZVOUS_PROCESS, DIE_MCA_RENDZVOUS_PROCESS,
DIE_MCA_RENDZVOUS_LEAVE, DIE_MCA_RENDZVOUS_LEAVE,
DIE_MCA_NEW_TIMEOUT,
DIE_INIT_ENTER,
DIE_INIT_MONARCH_ENTER, DIE_INIT_MONARCH_ENTER,
DIE_INIT_MONARCH_PROCESS, DIE_INIT_MONARCH_PROCESS,
DIE_INIT_MONARCH_LEAVE, DIE_INIT_MONARCH_LEAVE,
......
...@@ -148,6 +148,11 @@ extern int ia64_reg_MCA_extension(int (*fn)(void *, struct ia64_sal_os_state *) ...@@ -148,6 +148,11 @@ extern int ia64_reg_MCA_extension(int (*fn)(void *, struct ia64_sal_os_state *)
extern void ia64_unreg_MCA_extension(void); extern void ia64_unreg_MCA_extension(void);
extern u64 ia64_get_rnat(u64 *); extern u64 ia64_get_rnat(u64 *);
struct ia64_mca_notify_die {
struct ia64_sal_os_state *sos;
int *monarch_cpu;
};
#else /* __ASSEMBLY__ */ #else /* __ASSEMBLY__ */
#define IA64_MCA_CORRECTED 0x0 /* Error has been corrected by OS_MCA */ #define IA64_MCA_CORRECTED 0x0 /* Error has been corrected by OS_MCA */
......
...@@ -84,7 +84,7 @@ __mutex_fastpath_unlock(atomic_t *count, void (*fail_fn)(atomic_t *)) ...@@ -84,7 +84,7 @@ __mutex_fastpath_unlock(atomic_t *count, void (*fail_fn)(atomic_t *))
static inline int static inline int
__mutex_fastpath_trylock(atomic_t *count, int (*fail_fn)(atomic_t *)) __mutex_fastpath_trylock(atomic_t *count, int (*fail_fn)(atomic_t *))
{ {
if (likely(cmpxchg_acq(count, 1, 0)) == 1) if (cmpxchg_acq(count, 1, 0) == 1)
return 1; return 1;
return 0; return 0;
} }
......
...@@ -286,12 +286,15 @@ ...@@ -286,12 +286,15 @@
/* 1294, 1295 reserved for pselect/ppoll */ /* 1294, 1295 reserved for pselect/ppoll */
#define __NR_unshare 1296 #define __NR_unshare 1296
#define __NR_splice 1297 #define __NR_splice 1297
#define __NR_set_robust_list 1298
#define __NR_get_robust_list 1299
#define __NR_sync_file_range 1300
#ifdef __KERNEL__ #ifdef __KERNEL__
#include <linux/config.h> #include <linux/config.h>
#define NR_syscalls 274 /* length of syscall table */ #define NR_syscalls 277 /* length of syscall table */
#define __ARCH_WANT_SYS_RT_SIGACTION #define __ARCH_WANT_SYS_RT_SIGACTION
......
...@@ -17,7 +17,7 @@ ...@@ -17,7 +17,7 @@
extern unsigned long vga_console_iobase; extern unsigned long vga_console_iobase;
extern unsigned long vga_console_membase; extern unsigned long vga_console_membase;
#define VGA_MAP_MEM(x) ((unsigned long) ioremap(vga_console_membase + (x), 0)) #define VGA_MAP_MEM(x) ((unsigned long) ioremap_nocache(vga_console_membase + (x), 0))
#define vga_readb(x) (*(x)) #define vga_readb(x) (*(x))
#define vga_writeb(x,y) (*(y) = (x)) #define vga_writeb(x,y) (*(y) = (x))
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment