Commit 43ff2f4d authored by Linus Torvalds's avatar Linus Torvalds

Merge branch 'x86-platform-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip

Pull x86 platform updates from Ingo Molnar:
 "The main changes in this cycle were:

   - a refactoring of the early virt init code by merging 'struct
     x86_hyper' into 'struct x86_platform' and 'struct x86_init', which
     allows simplifications and also the addition of a new
     ->guest_late_init() callback. (Juergen Gross)

   - timer_setup() conversion of the UV code (Kees Cook)"

* 'x86-platform-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip:
  x86/virt/xen: Use guest_late_init to detect Xen PVH guest
  x86/virt, x86/platform: Add ->guest_late_init() callback to hypervisor_x86 structure
  x86/virt, x86/acpi: Add test for ACPI_FADT_NO_VGA
  x86/virt: Add enum for hypervisors to replace x86_hyper
  x86/virt, x86/platform: Merge 'struct x86_hyper' into 'struct x86_platform' and 'struct x86_init'
  x86/platform/UV: Convert timers to use timer_setup()
parents 13e57da4 418492ba
...@@ -113,7 +113,7 @@ void hyperv_init(void) ...@@ -113,7 +113,7 @@ void hyperv_init(void)
u64 guest_id; u64 guest_id;
union hv_x64_msr_hypercall_contents hypercall_msr; union hv_x64_msr_hypercall_contents hypercall_msr;
if (x86_hyper != &x86_hyper_ms_hyperv) if (x86_hyper_type != X86_HYPER_MS_HYPERV)
return; return;
/* Allocate percpu VP index */ /* Allocate percpu VP index */
......
...@@ -23,11 +23,22 @@ ...@@ -23,11 +23,22 @@
#ifdef CONFIG_HYPERVISOR_GUEST #ifdef CONFIG_HYPERVISOR_GUEST
#include <asm/kvm_para.h> #include <asm/kvm_para.h>
#include <asm/x86_init.h>
#include <asm/xen/hypervisor.h> #include <asm/xen/hypervisor.h>
/* /*
* x86 hypervisor information * x86 hypervisor information
*/ */
enum x86_hypervisor_type {
X86_HYPER_NATIVE = 0,
X86_HYPER_VMWARE,
X86_HYPER_MS_HYPERV,
X86_HYPER_XEN_PV,
X86_HYPER_XEN_HVM,
X86_HYPER_KVM,
};
struct hypervisor_x86 { struct hypervisor_x86 {
/* Hypervisor name */ /* Hypervisor name */
const char *name; const char *name;
...@@ -35,40 +46,19 @@ struct hypervisor_x86 { ...@@ -35,40 +46,19 @@ struct hypervisor_x86 {
/* Detection routine */ /* Detection routine */
uint32_t (*detect)(void); uint32_t (*detect)(void);
/* Platform setup (run once per boot) */ /* Hypervisor type */
void (*init_platform)(void); enum x86_hypervisor_type type;
/* X2APIC detection (run once per boot) */
bool (*x2apic_available)(void);
/* pin current vcpu to specified physical cpu (run rarely) */ /* init time callbacks */
void (*pin_vcpu)(int); struct x86_hyper_init init;
/* called during init_mem_mapping() to setup early mappings. */ /* runtime callbacks */
void (*init_mem_mapping)(void); struct x86_hyper_runtime runtime;
}; };
extern const struct hypervisor_x86 *x86_hyper; extern enum x86_hypervisor_type x86_hyper_type;
/* Recognized hypervisors */
extern const struct hypervisor_x86 x86_hyper_vmware;
extern const struct hypervisor_x86 x86_hyper_ms_hyperv;
extern const struct hypervisor_x86 x86_hyper_xen_pv;
extern const struct hypervisor_x86 x86_hyper_xen_hvm;
extern const struct hypervisor_x86 x86_hyper_kvm;
extern void init_hypervisor_platform(void); extern void init_hypervisor_platform(void);
extern bool hypervisor_x2apic_available(void);
extern void hypervisor_pin_vcpu(int cpu);
static inline void hypervisor_init_mem_mapping(void)
{
if (x86_hyper && x86_hyper->init_mem_mapping)
x86_hyper->init_mem_mapping();
}
#else #else
static inline void init_hypervisor_platform(void) { } static inline void init_hypervisor_platform(void) { }
static inline bool hypervisor_x2apic_available(void) { return false; }
static inline void hypervisor_init_mem_mapping(void) { }
#endif /* CONFIG_HYPERVISOR_GUEST */ #endif /* CONFIG_HYPERVISOR_GUEST */
#endif /* _ASM_X86_HYPERVISOR_H */ #endif /* _ASM_X86_HYPERVISOR_H */
...@@ -88,7 +88,6 @@ static inline long kvm_hypercall4(unsigned int nr, unsigned long p1, ...@@ -88,7 +88,6 @@ static inline long kvm_hypercall4(unsigned int nr, unsigned long p1,
#ifdef CONFIG_KVM_GUEST #ifdef CONFIG_KVM_GUEST
bool kvm_para_available(void); bool kvm_para_available(void);
unsigned int kvm_arch_para_features(void); unsigned int kvm_arch_para_features(void);
void __init kvm_guest_init(void);
void kvm_async_pf_task_wait(u32 token, int interrupt_kernel); void kvm_async_pf_task_wait(u32 token, int interrupt_kernel);
void kvm_async_pf_task_wake(u32 token); void kvm_async_pf_task_wake(u32 token);
u32 kvm_read_and_reset_pf_reason(void); u32 kvm_read_and_reset_pf_reason(void);
...@@ -103,7 +102,6 @@ static inline void kvm_spinlock_init(void) ...@@ -103,7 +102,6 @@ static inline void kvm_spinlock_init(void)
#endif /* CONFIG_PARAVIRT_SPINLOCKS */ #endif /* CONFIG_PARAVIRT_SPINLOCKS */
#else /* CONFIG_KVM_GUEST */ #else /* CONFIG_KVM_GUEST */
#define kvm_guest_init() do {} while (0)
#define kvm_async_pf_task_wait(T, I) do {} while(0) #define kvm_async_pf_task_wait(T, I) do {} while(0)
#define kvm_async_pf_task_wake(T) do {} while(0) #define kvm_async_pf_task_wake(T) do {} while(0)
......
...@@ -114,6 +114,20 @@ struct x86_init_pci { ...@@ -114,6 +114,20 @@ struct x86_init_pci {
void (*fixup_irqs)(void); void (*fixup_irqs)(void);
}; };
/**
* struct x86_hyper_init - x86 hypervisor init functions
* @init_platform: platform setup
* @guest_late_init: guest late init
* @x2apic_available: X2APIC detection
* @init_mem_mapping: setup early mappings during init_mem_mapping()
*/
struct x86_hyper_init {
void (*init_platform)(void);
void (*guest_late_init)(void);
bool (*x2apic_available)(void);
void (*init_mem_mapping)(void);
};
/** /**
* struct x86_init_ops - functions for platform specific setup * struct x86_init_ops - functions for platform specific setup
* *
...@@ -127,6 +141,7 @@ struct x86_init_ops { ...@@ -127,6 +141,7 @@ struct x86_init_ops {
struct x86_init_timers timers; struct x86_init_timers timers;
struct x86_init_iommu iommu; struct x86_init_iommu iommu;
struct x86_init_pci pci; struct x86_init_pci pci;
struct x86_hyper_init hyper;
}; };
/** /**
...@@ -195,10 +210,20 @@ enum x86_legacy_i8042_state { ...@@ -195,10 +210,20 @@ enum x86_legacy_i8042_state {
struct x86_legacy_features { struct x86_legacy_features {
enum x86_legacy_i8042_state i8042; enum x86_legacy_i8042_state i8042;
int rtc; int rtc;
int no_vga;
int reserve_bios_regions; int reserve_bios_regions;
struct x86_legacy_devices devices; struct x86_legacy_devices devices;
}; };
/**
* struct x86_hyper_runtime - x86 hypervisor specific runtime callbacks
*
* @pin_vcpu: pin current vcpu to specified physical cpu (run rarely)
*/
struct x86_hyper_runtime {
void (*pin_vcpu)(int cpu);
};
/** /**
* struct x86_platform_ops - platform specific runtime functions * struct x86_platform_ops - platform specific runtime functions
* @calibrate_cpu: calibrate CPU * @calibrate_cpu: calibrate CPU
...@@ -218,6 +243,7 @@ struct x86_legacy_features { ...@@ -218,6 +243,7 @@ struct x86_legacy_features {
* possible in x86_early_init_platform_quirks() by * possible in x86_early_init_platform_quirks() by
* only using the current x86_hardware_subarch * only using the current x86_hardware_subarch
* semantics. * semantics.
* @hyper: x86 hypervisor specific runtime callbacks
*/ */
struct x86_platform_ops { struct x86_platform_ops {
unsigned long (*calibrate_cpu)(void); unsigned long (*calibrate_cpu)(void);
...@@ -233,6 +259,7 @@ struct x86_platform_ops { ...@@ -233,6 +259,7 @@ struct x86_platform_ops {
void (*apic_post_init)(void); void (*apic_post_init)(void);
struct x86_legacy_features legacy; struct x86_legacy_features legacy;
void (*set_legacy_features)(void); void (*set_legacy_features)(void);
struct x86_hyper_runtime hyper;
}; };
struct pci_dev; struct pci_dev;
......
...@@ -961,6 +961,11 @@ static int __init acpi_parse_fadt(struct acpi_table_header *table) ...@@ -961,6 +961,11 @@ static int __init acpi_parse_fadt(struct acpi_table_header *table)
x86_platform.legacy.rtc = 0; x86_platform.legacy.rtc = 0;
} }
if (acpi_gbl_FADT.boot_flags & ACPI_FADT_NO_VGA) {
pr_debug("ACPI: probing for VGA not safe\n");
x86_platform.legacy.no_vga = 1;
}
#ifdef CONFIG_X86_PM_TIMER #ifdef CONFIG_X86_PM_TIMER
/* detect the location of the ACPI PM Timer */ /* detect the location of the ACPI PM Timer */
if (acpi_gbl_FADT.header.revision >= FADT2_REVISION_ID) { if (acpi_gbl_FADT.header.revision >= FADT2_REVISION_ID) {
......
...@@ -1645,7 +1645,7 @@ static __init void try_to_enable_x2apic(int remap_mode) ...@@ -1645,7 +1645,7 @@ static __init void try_to_enable_x2apic(int remap_mode)
* under KVM * under KVM
*/ */
if (max_physical_apicid > 255 || if (max_physical_apicid > 255 ||
!hypervisor_x2apic_available()) { !x86_init.hyper.x2apic_available()) {
pr_info("x2apic: IRQ remapping doesn't support X2APIC mode\n"); pr_info("x2apic: IRQ remapping doesn't support X2APIC mode\n");
x2apic_disable(); x2apic_disable();
return; return;
......
...@@ -920,9 +920,8 @@ static __init void uv_rtc_init(void) ...@@ -920,9 +920,8 @@ static __init void uv_rtc_init(void)
/* /*
* percpu heartbeat timer * percpu heartbeat timer
*/ */
static void uv_heartbeat(unsigned long ignored) static void uv_heartbeat(struct timer_list *timer)
{ {
struct timer_list *timer = &uv_scir_info->timer;
unsigned char bits = uv_scir_info->state; unsigned char bits = uv_scir_info->state;
/* Flip heartbeat bit: */ /* Flip heartbeat bit: */
...@@ -947,7 +946,7 @@ static int uv_heartbeat_enable(unsigned int cpu) ...@@ -947,7 +946,7 @@ static int uv_heartbeat_enable(unsigned int cpu)
struct timer_list *timer = &uv_cpu_scir_info(cpu)->timer; struct timer_list *timer = &uv_cpu_scir_info(cpu)->timer;
uv_set_cpu_scir_bits(cpu, SCIR_CPU_HEARTBEAT|SCIR_CPU_ACTIVITY); uv_set_cpu_scir_bits(cpu, SCIR_CPU_HEARTBEAT|SCIR_CPU_ACTIVITY);
setup_pinned_timer(timer, uv_heartbeat, cpu); timer_setup(timer, uv_heartbeat, TIMER_PINNED);
timer->expires = jiffies + SCIR_CPU_HB_INTERVAL; timer->expires = jiffies + SCIR_CPU_HB_INTERVAL;
add_timer_on(timer, cpu); add_timer_on(timer, cpu);
uv_cpu_scir_info(cpu)->enabled = 1; uv_cpu_scir_info(cpu)->enabled = 1;
......
...@@ -26,6 +26,12 @@ ...@@ -26,6 +26,12 @@
#include <asm/processor.h> #include <asm/processor.h>
#include <asm/hypervisor.h> #include <asm/hypervisor.h>
extern const struct hypervisor_x86 x86_hyper_vmware;
extern const struct hypervisor_x86 x86_hyper_ms_hyperv;
extern const struct hypervisor_x86 x86_hyper_xen_pv;
extern const struct hypervisor_x86 x86_hyper_xen_hvm;
extern const struct hypervisor_x86 x86_hyper_kvm;
static const __initconst struct hypervisor_x86 * const hypervisors[] = static const __initconst struct hypervisor_x86 * const hypervisors[] =
{ {
#ifdef CONFIG_XEN_PV #ifdef CONFIG_XEN_PV
...@@ -41,54 +47,52 @@ static const __initconst struct hypervisor_x86 * const hypervisors[] = ...@@ -41,54 +47,52 @@ static const __initconst struct hypervisor_x86 * const hypervisors[] =
#endif #endif
}; };
const struct hypervisor_x86 *x86_hyper; enum x86_hypervisor_type x86_hyper_type;
EXPORT_SYMBOL(x86_hyper); EXPORT_SYMBOL(x86_hyper_type);
static inline void __init static inline const struct hypervisor_x86 * __init
detect_hypervisor_vendor(void) detect_hypervisor_vendor(void)
{ {
const struct hypervisor_x86 *h, * const *p; const struct hypervisor_x86 *h = NULL, * const *p;
uint32_t pri, max_pri = 0; uint32_t pri, max_pri = 0;
for (p = hypervisors; p < hypervisors + ARRAY_SIZE(hypervisors); p++) { for (p = hypervisors; p < hypervisors + ARRAY_SIZE(hypervisors); p++) {
h = *p; pri = (*p)->detect();
pri = h->detect(); if (pri > max_pri) {
if (pri != 0 && pri > max_pri) {
max_pri = pri; max_pri = pri;
x86_hyper = h; h = *p;
} }
} }
if (max_pri) if (h)
pr_info("Hypervisor detected: %s\n", x86_hyper->name); pr_info("Hypervisor detected: %s\n", h->name);
return h;
} }
void __init init_hypervisor_platform(void) static void __init copy_array(const void *src, void *target, unsigned int size)
{ {
unsigned int i, n = size / sizeof(void *);
const void * const *from = (const void * const *)src;
const void **to = (const void **)target;
detect_hypervisor_vendor(); for (i = 0; i < n; i++)
if (from[i])
if (!x86_hyper) to[i] = from[i];
return;
if (x86_hyper->init_platform)
x86_hyper->init_platform();
} }
bool __init hypervisor_x2apic_available(void) void __init init_hypervisor_platform(void)
{ {
return x86_hyper && const struct hypervisor_x86 *h;
x86_hyper->x2apic_available &&
x86_hyper->x2apic_available();
}
void hypervisor_pin_vcpu(int cpu) h = detect_hypervisor_vendor();
{
if (!x86_hyper) if (!h)
return; return;
if (x86_hyper->pin_vcpu) copy_array(&h->init, &x86_init.hyper, sizeof(h->init));
x86_hyper->pin_vcpu(cpu); copy_array(&h->runtime, &x86_platform.hyper, sizeof(h->runtime));
else
WARN_ONCE(1, "vcpu pinning requested but not supported!\n"); x86_hyper_type = h->type;
x86_init.hyper.init_platform();
} }
...@@ -254,9 +254,9 @@ static void __init ms_hyperv_init_platform(void) ...@@ -254,9 +254,9 @@ static void __init ms_hyperv_init_platform(void)
#endif #endif
} }
const __refconst struct hypervisor_x86 x86_hyper_ms_hyperv = { const __initconst struct hypervisor_x86 x86_hyper_ms_hyperv = {
.name = "Microsoft Hyper-V", .name = "Microsoft Hyper-V",
.detect = ms_hyperv_platform, .detect = ms_hyperv_platform,
.init_platform = ms_hyperv_init_platform, .type = X86_HYPER_MS_HYPERV,
.init.init_platform = ms_hyperv_init_platform,
}; };
EXPORT_SYMBOL(x86_hyper_ms_hyperv);
...@@ -205,10 +205,10 @@ static bool __init vmware_legacy_x2apic_available(void) ...@@ -205,10 +205,10 @@ static bool __init vmware_legacy_x2apic_available(void)
(eax & (1 << VMWARE_PORT_CMD_LEGACY_X2APIC)) != 0; (eax & (1 << VMWARE_PORT_CMD_LEGACY_X2APIC)) != 0;
} }
const __refconst struct hypervisor_x86 x86_hyper_vmware = { const __initconst struct hypervisor_x86 x86_hyper_vmware = {
.name = "VMware", .name = "VMware",
.detect = vmware_platform, .detect = vmware_platform,
.init_platform = vmware_platform_setup, .type = X86_HYPER_VMWARE,
.x2apic_available = vmware_legacy_x2apic_available, .init.init_platform = vmware_platform_setup,
.init.x2apic_available = vmware_legacy_x2apic_available,
}; };
EXPORT_SYMBOL(x86_hyper_vmware);
...@@ -498,7 +498,7 @@ static void __init kvm_apf_trap_init(void) ...@@ -498,7 +498,7 @@ static void __init kvm_apf_trap_init(void)
update_intr_gate(X86_TRAP_PF, async_page_fault); update_intr_gate(X86_TRAP_PF, async_page_fault);
} }
void __init kvm_guest_init(void) static void __init kvm_guest_init(void)
{ {
int i; int i;
...@@ -578,12 +578,13 @@ static uint32_t __init kvm_detect(void) ...@@ -578,12 +578,13 @@ static uint32_t __init kvm_detect(void)
return kvm_cpuid_base(); return kvm_cpuid_base();
} }
const struct hypervisor_x86 x86_hyper_kvm __refconst = { const __initconst struct hypervisor_x86 x86_hyper_kvm = {
.name = "KVM", .name = "KVM",
.detect = kvm_detect, .detect = kvm_detect,
.x2apic_available = kvm_para_available, .type = X86_HYPER_KVM,
.init.guest_late_init = kvm_guest_init,
.init.x2apic_available = kvm_para_available,
}; };
EXPORT_SYMBOL_GPL(x86_hyper_kvm);
static __init int activate_jump_labels(void) static __init int activate_jump_labels(void)
{ {
......
...@@ -1296,7 +1296,7 @@ void __init setup_arch(char **cmdline_p) ...@@ -1296,7 +1296,7 @@ void __init setup_arch(char **cmdline_p)
io_apic_init_mappings(); io_apic_init_mappings();
kvm_guest_init(); x86_init.hyper.guest_late_init();
e820__reserve_resources(); e820__reserve_resources();
e820__register_nosave_regions(max_low_pfn); e820__register_nosave_regions(max_low_pfn);
......
...@@ -28,6 +28,8 @@ void x86_init_noop(void) { } ...@@ -28,6 +28,8 @@ void x86_init_noop(void) { }
void __init x86_init_uint_noop(unsigned int unused) { } void __init x86_init_uint_noop(unsigned int unused) { }
int __init iommu_init_noop(void) { return 0; } int __init iommu_init_noop(void) { return 0; }
void iommu_shutdown_noop(void) { } void iommu_shutdown_noop(void) { }
bool __init bool_x86_init_noop(void) { return false; }
void x86_op_int_noop(int cpu) { }
/* /*
* The platform setup functions are preset with the default functions * The platform setup functions are preset with the default functions
...@@ -81,6 +83,13 @@ struct x86_init_ops x86_init __initdata = { ...@@ -81,6 +83,13 @@ struct x86_init_ops x86_init __initdata = {
.init_irq = x86_default_pci_init_irq, .init_irq = x86_default_pci_init_irq,
.fixup_irqs = x86_default_pci_fixup_irqs, .fixup_irqs = x86_default_pci_fixup_irqs,
}, },
.hyper = {
.init_platform = x86_init_noop,
.guest_late_init = x86_init_noop,
.x2apic_available = bool_x86_init_noop,
.init_mem_mapping = x86_init_noop,
},
}; };
struct x86_cpuinit_ops x86_cpuinit = { struct x86_cpuinit_ops x86_cpuinit = {
...@@ -101,6 +110,7 @@ struct x86_platform_ops x86_platform __ro_after_init = { ...@@ -101,6 +110,7 @@ struct x86_platform_ops x86_platform __ro_after_init = {
.get_nmi_reason = default_get_nmi_reason, .get_nmi_reason = default_get_nmi_reason,
.save_sched_clock_state = tsc_save_sched_clock_state, .save_sched_clock_state = tsc_save_sched_clock_state,
.restore_sched_clock_state = tsc_restore_sched_clock_state, .restore_sched_clock_state = tsc_restore_sched_clock_state,
.hyper.pin_vcpu = x86_op_int_noop,
}; };
EXPORT_SYMBOL_GPL(x86_platform); EXPORT_SYMBOL_GPL(x86_platform);
......
...@@ -671,7 +671,7 @@ void __init init_mem_mapping(void) ...@@ -671,7 +671,7 @@ void __init init_mem_mapping(void)
load_cr3(swapper_pg_dir); load_cr3(swapper_pg_dir);
__flush_tlb_all(); __flush_tlb_all();
hypervisor_init_mem_mapping(); x86_init.hyper.init_mem_mapping();
early_memtest(0, max_pfn_mapped << PAGE_SHIFT); early_memtest(0, max_pfn_mapped << PAGE_SHIFT);
} }
......
#include <linux/acpi.h>
#include <linux/cpu.h> #include <linux/cpu.h>
#include <linux/kexec.h> #include <linux/kexec.h>
#include <linux/memblock.h> #include <linux/memblock.h>
...@@ -188,8 +189,6 @@ static void __init xen_hvm_guest_init(void) ...@@ -188,8 +189,6 @@ static void __init xen_hvm_guest_init(void)
xen_hvm_init_time_ops(); xen_hvm_init_time_ops();
xen_hvm_init_mmu_ops(); xen_hvm_init_mmu_ops();
if (xen_pvh_domain())
machine_ops.emergency_restart = xen_emergency_restart;
#ifdef CONFIG_KEXEC_CORE #ifdef CONFIG_KEXEC_CORE
machine_ops.shutdown = xen_hvm_shutdown; machine_ops.shutdown = xen_hvm_shutdown;
machine_ops.crash_shutdown = xen_hvm_crash_shutdown; machine_ops.crash_shutdown = xen_hvm_crash_shutdown;
...@@ -226,12 +225,33 @@ static uint32_t __init xen_platform_hvm(void) ...@@ -226,12 +225,33 @@ static uint32_t __init xen_platform_hvm(void)
return xen_cpuid_base(); return xen_cpuid_base();
} }
const struct hypervisor_x86 x86_hyper_xen_hvm = { static __init void xen_hvm_guest_late_init(void)
{
#ifdef CONFIG_XEN_PVH
/* Test for PVH domain (PVH boot path taken overrides ACPI flags). */
if (!xen_pvh &&
(x86_platform.legacy.rtc || !x86_platform.legacy.no_vga))
return;
/* PVH detected. */
xen_pvh = true;
/* Make sure we don't fall back to (default) ACPI_IRQ_MODEL_PIC. */
if (!nr_ioapics && acpi_irq_model == ACPI_IRQ_MODEL_PIC)
acpi_irq_model = ACPI_IRQ_MODEL_PLATFORM;
machine_ops.emergency_restart = xen_emergency_restart;
pv_info.name = "Xen PVH";
#endif
}
const __initconst struct hypervisor_x86 x86_hyper_xen_hvm = {
.name = "Xen HVM", .name = "Xen HVM",
.detect = xen_platform_hvm, .detect = xen_platform_hvm,
.init_platform = xen_hvm_guest_init, .type = X86_HYPER_XEN_HVM,
.pin_vcpu = xen_pin_vcpu, .init.init_platform = xen_hvm_guest_init,
.x2apic_available = xen_x2apic_para_available, .init.x2apic_available = xen_x2apic_para_available,
.init_mem_mapping = xen_hvm_init_mem_mapping, .init.init_mem_mapping = xen_hvm_init_mem_mapping,
.init.guest_late_init = xen_hvm_guest_late_init,
.runtime.pin_vcpu = xen_pin_vcpu,
}; };
EXPORT_SYMBOL(x86_hyper_xen_hvm);
...@@ -1459,9 +1459,9 @@ static uint32_t __init xen_platform_pv(void) ...@@ -1459,9 +1459,9 @@ static uint32_t __init xen_platform_pv(void)
return 0; return 0;
} }
const struct hypervisor_x86 x86_hyper_xen_pv = { const __initconst struct hypervisor_x86 x86_hyper_xen_pv = {
.name = "Xen PV", .name = "Xen PV",
.detect = xen_platform_pv, .detect = xen_platform_pv,
.pin_vcpu = xen_pin_vcpu, .type = X86_HYPER_XEN_PV,
.runtime.pin_vcpu = xen_pin_vcpu,
}; };
EXPORT_SYMBOL(x86_hyper_xen_pv);
...@@ -25,13 +25,6 @@ struct boot_params pvh_bootparams __attribute__((section(".data"))); ...@@ -25,13 +25,6 @@ struct boot_params pvh_bootparams __attribute__((section(".data")));
struct hvm_start_info pvh_start_info; struct hvm_start_info pvh_start_info;
unsigned int pvh_start_info_sz = sizeof(pvh_start_info); unsigned int pvh_start_info_sz = sizeof(pvh_start_info);
static void xen_pvh_arch_setup(void)
{
/* Make sure we don't fall back to (default) ACPI_IRQ_MODEL_PIC. */
if (nr_ioapics == 0)
acpi_irq_model = ACPI_IRQ_MODEL_PLATFORM;
}
static void __init init_pvh_bootparams(void) static void __init init_pvh_bootparams(void)
{ {
struct xen_memory_map memmap; struct xen_memory_map memmap;
...@@ -102,6 +95,4 @@ void __init xen_prepare_pvh(void) ...@@ -102,6 +95,4 @@ void __init xen_prepare_pvh(void)
wrmsr_safe(msr, (u32)pfn, (u32)(pfn >> 32)); wrmsr_safe(msr, (u32)pfn, (u32)(pfn >> 32));
init_pvh_bootparams(); init_pvh_bootparams();
x86_init.oem.arch_setup = xen_pvh_arch_setup;
} }
...@@ -1534,7 +1534,7 @@ static int __init hv_acpi_init(void) ...@@ -1534,7 +1534,7 @@ static int __init hv_acpi_init(void)
{ {
int ret, t; int ret, t;
if (x86_hyper != &x86_hyper_ms_hyperv) if (x86_hyper_type != X86_HYPER_MS_HYPERV)
return -ENODEV; return -ENODEV;
init_completion(&probe_event); init_completion(&probe_event);
......
...@@ -316,11 +316,9 @@ static int vmmouse_enable(struct psmouse *psmouse) ...@@ -316,11 +316,9 @@ static int vmmouse_enable(struct psmouse *psmouse)
/* /*
* Array of supported hypervisors. * Array of supported hypervisors.
*/ */
static const struct hypervisor_x86 *vmmouse_supported_hypervisors[] = { static enum x86_hypervisor_type vmmouse_supported_hypervisors[] = {
&x86_hyper_vmware, X86_HYPER_VMWARE,
#ifdef CONFIG_KVM_GUEST X86_HYPER_KVM,
&x86_hyper_kvm,
#endif
}; };
/** /**
...@@ -331,7 +329,7 @@ static bool vmmouse_check_hypervisor(void) ...@@ -331,7 +329,7 @@ static bool vmmouse_check_hypervisor(void)
int i; int i;
for (i = 0; i < ARRAY_SIZE(vmmouse_supported_hypervisors); i++) for (i = 0; i < ARRAY_SIZE(vmmouse_supported_hypervisors); i++)
if (vmmouse_supported_hypervisors[i] == x86_hyper) if (vmmouse_supported_hypervisors[i] == x86_hyper_type)
return true; return true;
return false; return false;
......
...@@ -1271,7 +1271,7 @@ static int __init vmballoon_init(void) ...@@ -1271,7 +1271,7 @@ static int __init vmballoon_init(void)
* Check if we are running on VMware's hypervisor and bail out * Check if we are running on VMware's hypervisor and bail out
* if we are not. * if we are not.
*/ */
if (x86_hyper != &x86_hyper_vmware) if (x86_hyper_type != X86_HYPER_VMWARE)
return -ENODEV; return -ENODEV;
for (is_2m_pages = 0; is_2m_pages < VMW_BALLOON_NUM_PAGE_SIZES; for (is_2m_pages = 0; is_2m_pages < VMW_BALLOON_NUM_PAGE_SIZES;
......
...@@ -7,8 +7,12 @@ ...@@ -7,8 +7,12 @@
* Juergen Gross <jgross@suse.com> * Juergen Gross <jgross@suse.com>
*/ */
#ifdef CONFIG_HYPERVISOR_GUEST #ifdef CONFIG_X86
#include <asm/hypervisor.h> #include <asm/x86_init.h>
static inline void hypervisor_pin_vcpu(int cpu)
{
x86_platform.hyper.pin_vcpu(cpu);
}
#else #else
static inline void hypervisor_pin_vcpu(int cpu) static inline void hypervisor_pin_vcpu(int cpu)
{ {
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment