Commit d34696c2 authored by Linus Torvalds's avatar Linus Torvalds

Merge branch 'for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/s390/linux

Pull s390 fixes from Martin Schwidefsky:
 "Two patches to save some memory if CONFIG_NR_CPUS is large, a changed
  default for the use of compare-and-delay, and a couple of bug fixes"

* 'for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/s390/linux:
  s390/spinlock: disabled compare-and-delay by default
  s390/mm: align 64-bit PIE binaries to 4GB
  s390/cacheinfo: coding style changes
  s390/cacheinfo: fix shared cpu masks
  s390/smp: reduce size of struct pcpu
  s390/topology: convert cpu_topology array to per cpu variable
  s390/topology: delay initialization of topology cpu masks
  s390/vdso: fix clock_gettime for CLOCK_THREAD_CPUTIME_ID, -2 and -3
parents f9677375 61b0b016
...@@ -18,15 +18,15 @@ struct cpu_topology_s390 { ...@@ -18,15 +18,15 @@ struct cpu_topology_s390 {
cpumask_t book_mask; cpumask_t book_mask;
}; };
extern struct cpu_topology_s390 cpu_topology[NR_CPUS]; DECLARE_PER_CPU(struct cpu_topology_s390, cpu_topology);
#define topology_physical_package_id(cpu) (cpu_topology[cpu].socket_id) #define topology_physical_package_id(cpu) (per_cpu(cpu_topology, cpu).socket_id)
#define topology_thread_id(cpu) (cpu_topology[cpu].thread_id) #define topology_thread_id(cpu) (per_cpu(cpu_topology, cpu).thread_id)
#define topology_thread_cpumask(cpu) (&cpu_topology[cpu].thread_mask) #define topology_thread_cpumask(cpu) (&per_cpu(cpu_topology, cpu).thread_mask)
#define topology_core_id(cpu) (cpu_topology[cpu].core_id) #define topology_core_id(cpu) (per_cpu(cpu_topology, cpu).core_id)
#define topology_core_cpumask(cpu) (&cpu_topology[cpu].core_mask) #define topology_core_cpumask(cpu) (&per_cpu(cpu_topology, cpu).core_mask)
#define topology_book_id(cpu) (cpu_topology[cpu].book_id) #define topology_book_id(cpu) (per_cpu(cpu_topology, cpu).book_id)
#define topology_book_cpumask(cpu) (&cpu_topology[cpu].book_mask) #define topology_book_cpumask(cpu) (&per_cpu(cpu_topology, cpu).book_mask)
#define mc_capable() 1 #define mc_capable() 1
...@@ -51,14 +51,6 @@ static inline void topology_expect_change(void) { } ...@@ -51,14 +51,6 @@ static inline void topology_expect_change(void) { }
#define POLARIZATION_VM (2) #define POLARIZATION_VM (2)
#define POLARIZATION_VH (3) #define POLARIZATION_VH (3)
#ifdef CONFIG_SCHED_BOOK
void s390_init_cpu_topology(void);
#else
static inline void s390_init_cpu_topology(void)
{
};
#endif
#include <asm-generic/topology.h> #include <asm-generic/topology.h>
#endif /* _ASM_S390_TOPOLOGY_H */ #endif /* _ASM_S390_TOPOLOGY_H */
...@@ -91,12 +91,9 @@ static inline enum cache_type get_cache_type(struct cache_info *ci, int level) ...@@ -91,12 +91,9 @@ static inline enum cache_type get_cache_type(struct cache_info *ci, int level)
{ {
if (level >= CACHE_MAX_LEVEL) if (level >= CACHE_MAX_LEVEL)
return CACHE_TYPE_NOCACHE; return CACHE_TYPE_NOCACHE;
ci += level; ci += level;
if (ci->scope != CACHE_SCOPE_SHARED && ci->scope != CACHE_SCOPE_PRIVATE) if (ci->scope != CACHE_SCOPE_SHARED && ci->scope != CACHE_SCOPE_PRIVATE)
return CACHE_TYPE_NOCACHE; return CACHE_TYPE_NOCACHE;
return cache_type_map[ci->type]; return cache_type_map[ci->type];
} }
...@@ -111,23 +108,19 @@ static inline unsigned long ecag(int ai, int li, int ti) ...@@ -111,23 +108,19 @@ static inline unsigned long ecag(int ai, int li, int ti)
} }
static void ci_leaf_init(struct cacheinfo *this_leaf, int private, static void ci_leaf_init(struct cacheinfo *this_leaf, int private,
enum cache_type type, unsigned int level) enum cache_type type, unsigned int level, int cpu)
{ {
int ti, num_sets; int ti, num_sets;
int cpu = smp_processor_id();
if (type == CACHE_TYPE_INST) if (type == CACHE_TYPE_INST)
ti = CACHE_TI_INSTRUCTION; ti = CACHE_TI_INSTRUCTION;
else else
ti = CACHE_TI_UNIFIED; ti = CACHE_TI_UNIFIED;
this_leaf->level = level + 1; this_leaf->level = level + 1;
this_leaf->type = type; this_leaf->type = type;
this_leaf->coherency_line_size = ecag(EXTRACT_LINE_SIZE, level, ti); this_leaf->coherency_line_size = ecag(EXTRACT_LINE_SIZE, level, ti);
this_leaf->ways_of_associativity = ecag(EXTRACT_ASSOCIATIVITY, this_leaf->ways_of_associativity = ecag(EXTRACT_ASSOCIATIVITY, level, ti);
level, ti);
this_leaf->size = ecag(EXTRACT_SIZE, level, ti); this_leaf->size = ecag(EXTRACT_SIZE, level, ti);
num_sets = this_leaf->size / this_leaf->coherency_line_size; num_sets = this_leaf->size / this_leaf->coherency_line_size;
num_sets /= this_leaf->ways_of_associativity; num_sets /= this_leaf->ways_of_associativity;
this_leaf->number_of_sets = num_sets; this_leaf->number_of_sets = num_sets;
...@@ -145,7 +138,6 @@ int init_cache_level(unsigned int cpu) ...@@ -145,7 +138,6 @@ int init_cache_level(unsigned int cpu)
if (!this_cpu_ci) if (!this_cpu_ci)
return -EINVAL; return -EINVAL;
ct.raw = ecag(EXTRACT_TOPOLOGY, 0, 0); ct.raw = ecag(EXTRACT_TOPOLOGY, 0, 0);
do { do {
ctype = get_cache_type(&ct.ci[0], level); ctype = get_cache_type(&ct.ci[0], level);
...@@ -154,34 +146,31 @@ int init_cache_level(unsigned int cpu) ...@@ -154,34 +146,31 @@ int init_cache_level(unsigned int cpu)
/* Separate instruction and data caches */ /* Separate instruction and data caches */
leaves += (ctype == CACHE_TYPE_SEPARATE) ? 2 : 1; leaves += (ctype == CACHE_TYPE_SEPARATE) ? 2 : 1;
} while (++level < CACHE_MAX_LEVEL); } while (++level < CACHE_MAX_LEVEL);
this_cpu_ci->num_levels = level; this_cpu_ci->num_levels = level;
this_cpu_ci->num_leaves = leaves; this_cpu_ci->num_leaves = leaves;
return 0; return 0;
} }
int populate_cache_leaves(unsigned int cpu) int populate_cache_leaves(unsigned int cpu)
{ {
struct cpu_cacheinfo *this_cpu_ci = get_cpu_cacheinfo(cpu);
struct cacheinfo *this_leaf = this_cpu_ci->info_list;
unsigned int level, idx, pvt; unsigned int level, idx, pvt;
union cache_topology ct; union cache_topology ct;
enum cache_type ctype; enum cache_type ctype;
struct cpu_cacheinfo *this_cpu_ci = get_cpu_cacheinfo(cpu);
struct cacheinfo *this_leaf = this_cpu_ci->info_list;
ct.raw = ecag(EXTRACT_TOPOLOGY, 0, 0); ct.raw = ecag(EXTRACT_TOPOLOGY, 0, 0);
for (idx = 0, level = 0; level < this_cpu_ci->num_levels && for (idx = 0, level = 0; level < this_cpu_ci->num_levels &&
idx < this_cpu_ci->num_leaves; idx++, level++) { idx < this_cpu_ci->num_leaves; idx++, level++) {
if (!this_leaf) if (!this_leaf)
return -EINVAL; return -EINVAL;
pvt = (ct.ci[level].scope == CACHE_SCOPE_PRIVATE) ? 1 : 0; pvt = (ct.ci[level].scope == CACHE_SCOPE_PRIVATE) ? 1 : 0;
ctype = get_cache_type(&ct.ci[0], level); ctype = get_cache_type(&ct.ci[0], level);
if (ctype == CACHE_TYPE_SEPARATE) { if (ctype == CACHE_TYPE_SEPARATE) {
ci_leaf_init(this_leaf++, pvt, CACHE_TYPE_DATA, level); ci_leaf_init(this_leaf++, pvt, CACHE_TYPE_DATA, level, cpu);
ci_leaf_init(this_leaf++, pvt, CACHE_TYPE_INST, level); ci_leaf_init(this_leaf++, pvt, CACHE_TYPE_INST, level, cpu);
} else { } else {
ci_leaf_init(this_leaf++, pvt, ctype, level); ci_leaf_init(this_leaf++, pvt, ctype, level, cpu);
} }
} }
return 0; return 0;
......
...@@ -393,17 +393,19 @@ static __init void detect_machine_facilities(void) ...@@ -393,17 +393,19 @@ static __init void detect_machine_facilities(void)
S390_lowcore.machine_flags |= MACHINE_FLAG_TLB_LC; S390_lowcore.machine_flags |= MACHINE_FLAG_TLB_LC;
if (test_facility(129)) if (test_facility(129))
S390_lowcore.machine_flags |= MACHINE_FLAG_VX; S390_lowcore.machine_flags |= MACHINE_FLAG_VX;
if (test_facility(128))
S390_lowcore.machine_flags |= MACHINE_FLAG_CAD;
#endif #endif
} }
static int __init nocad_setup(char *str) static int __init cad_setup(char *str)
{ {
S390_lowcore.machine_flags &= ~MACHINE_FLAG_CAD; int val;
get_option(&str, &val);
if (val && test_facility(128))
S390_lowcore.machine_flags |= MACHINE_FLAG_CAD;
return 0; return 0;
} }
early_param("nocad", nocad_setup); early_param("cad", cad_setup);
static int __init cad_init(void) static int __init cad_init(void)
{ {
......
...@@ -909,7 +909,6 @@ void __init setup_arch(char **cmdline_p) ...@@ -909,7 +909,6 @@ void __init setup_arch(char **cmdline_p)
setup_lowcore(); setup_lowcore();
smp_fill_possible_mask(); smp_fill_possible_mask();
cpu_init(); cpu_init();
s390_init_cpu_topology();
/* /*
* Setup capabilities (ELF_HWCAP & ELF_PLATFORM). * Setup capabilities (ELF_HWCAP & ELF_PLATFORM).
......
...@@ -59,14 +59,13 @@ enum { ...@@ -59,14 +59,13 @@ enum {
CPU_STATE_CONFIGURED, CPU_STATE_CONFIGURED,
}; };
static DEFINE_PER_CPU(struct cpu *, cpu_device);
struct pcpu { struct pcpu {
struct cpu *cpu;
struct _lowcore *lowcore; /* lowcore page(s) for the cpu */ struct _lowcore *lowcore; /* lowcore page(s) for the cpu */
unsigned long async_stack; /* async stack for the cpu */
unsigned long panic_stack; /* panic stack for the cpu */
unsigned long ec_mask; /* bit mask for ec_xxx functions */ unsigned long ec_mask; /* bit mask for ec_xxx functions */
int state; /* physical cpu state */ signed char state; /* physical cpu state */
int polarization; /* physical polarization */ signed char polarization; /* physical polarization */
u16 address; /* physical cpu address */ u16 address; /* physical cpu address */
}; };
...@@ -173,25 +172,30 @@ static void pcpu_ec_call(struct pcpu *pcpu, int ec_bit) ...@@ -173,25 +172,30 @@ static void pcpu_ec_call(struct pcpu *pcpu, int ec_bit)
pcpu_sigp_retry(pcpu, order, 0); pcpu_sigp_retry(pcpu, order, 0);
} }
#define ASYNC_FRAME_OFFSET (ASYNC_SIZE - STACK_FRAME_OVERHEAD - __PT_SIZE)
#define PANIC_FRAME_OFFSET (PAGE_SIZE - STACK_FRAME_OVERHEAD - __PT_SIZE)
static int pcpu_alloc_lowcore(struct pcpu *pcpu, int cpu) static int pcpu_alloc_lowcore(struct pcpu *pcpu, int cpu)
{ {
unsigned long async_stack, panic_stack;
struct _lowcore *lc; struct _lowcore *lc;
if (pcpu != &pcpu_devices[0]) { if (pcpu != &pcpu_devices[0]) {
pcpu->lowcore = (struct _lowcore *) pcpu->lowcore = (struct _lowcore *)
__get_free_pages(GFP_KERNEL | GFP_DMA, LC_ORDER); __get_free_pages(GFP_KERNEL | GFP_DMA, LC_ORDER);
pcpu->async_stack = __get_free_pages(GFP_KERNEL, ASYNC_ORDER); async_stack = __get_free_pages(GFP_KERNEL, ASYNC_ORDER);
pcpu->panic_stack = __get_free_page(GFP_KERNEL); panic_stack = __get_free_page(GFP_KERNEL);
if (!pcpu->lowcore || !pcpu->panic_stack || !pcpu->async_stack) if (!pcpu->lowcore || !panic_stack || !async_stack)
goto out; goto out;
} else {
async_stack = pcpu->lowcore->async_stack - ASYNC_FRAME_OFFSET;
panic_stack = pcpu->lowcore->panic_stack - PANIC_FRAME_OFFSET;
} }
lc = pcpu->lowcore; lc = pcpu->lowcore;
memcpy(lc, &S390_lowcore, 512); memcpy(lc, &S390_lowcore, 512);
memset((char *) lc + 512, 0, sizeof(*lc) - 512); memset((char *) lc + 512, 0, sizeof(*lc) - 512);
lc->async_stack = pcpu->async_stack + ASYNC_SIZE lc->async_stack = async_stack + ASYNC_FRAME_OFFSET;
- STACK_FRAME_OVERHEAD - sizeof(struct pt_regs); lc->panic_stack = panic_stack + PANIC_FRAME_OFFSET;
lc->panic_stack = pcpu->panic_stack + PAGE_SIZE
- STACK_FRAME_OVERHEAD - sizeof(struct pt_regs);
lc->cpu_nr = cpu; lc->cpu_nr = cpu;
lc->spinlock_lockval = arch_spin_lockval(cpu); lc->spinlock_lockval = arch_spin_lockval(cpu);
#ifndef CONFIG_64BIT #ifndef CONFIG_64BIT
...@@ -212,8 +216,8 @@ static int pcpu_alloc_lowcore(struct pcpu *pcpu, int cpu) ...@@ -212,8 +216,8 @@ static int pcpu_alloc_lowcore(struct pcpu *pcpu, int cpu)
return 0; return 0;
out: out:
if (pcpu != &pcpu_devices[0]) { if (pcpu != &pcpu_devices[0]) {
free_page(pcpu->panic_stack); free_page(panic_stack);
free_pages(pcpu->async_stack, ASYNC_ORDER); free_pages(async_stack, ASYNC_ORDER);
free_pages((unsigned long) pcpu->lowcore, LC_ORDER); free_pages((unsigned long) pcpu->lowcore, LC_ORDER);
} }
return -ENOMEM; return -ENOMEM;
...@@ -235,11 +239,11 @@ static void pcpu_free_lowcore(struct pcpu *pcpu) ...@@ -235,11 +239,11 @@ static void pcpu_free_lowcore(struct pcpu *pcpu)
#else #else
vdso_free_per_cpu(pcpu->lowcore); vdso_free_per_cpu(pcpu->lowcore);
#endif #endif
if (pcpu != &pcpu_devices[0]) { if (pcpu == &pcpu_devices[0])
free_page(pcpu->panic_stack); return;
free_pages(pcpu->async_stack, ASYNC_ORDER); free_page(pcpu->lowcore->panic_stack-PANIC_FRAME_OFFSET);
free_pages((unsigned long) pcpu->lowcore, LC_ORDER); free_pages(pcpu->lowcore->async_stack-ASYNC_FRAME_OFFSET, ASYNC_ORDER);
} free_pages((unsigned long) pcpu->lowcore, LC_ORDER);
} }
#endif /* CONFIG_HOTPLUG_CPU */ #endif /* CONFIG_HOTPLUG_CPU */
...@@ -366,7 +370,8 @@ void smp_call_online_cpu(void (*func)(void *), void *data) ...@@ -366,7 +370,8 @@ void smp_call_online_cpu(void (*func)(void *), void *data)
void smp_call_ipl_cpu(void (*func)(void *), void *data) void smp_call_ipl_cpu(void (*func)(void *), void *data)
{ {
pcpu_delegate(&pcpu_devices[0], func, data, pcpu_delegate(&pcpu_devices[0], func, data,
pcpu_devices->panic_stack + PAGE_SIZE); pcpu_devices->lowcore->panic_stack -
PANIC_FRAME_OFFSET + PAGE_SIZE);
} }
int smp_find_processor_id(u16 address) int smp_find_processor_id(u16 address)
...@@ -935,10 +940,6 @@ void __init smp_prepare_boot_cpu(void) ...@@ -935,10 +940,6 @@ void __init smp_prepare_boot_cpu(void)
pcpu->state = CPU_STATE_CONFIGURED; pcpu->state = CPU_STATE_CONFIGURED;
pcpu->address = stap(); pcpu->address = stap();
pcpu->lowcore = (struct _lowcore *)(unsigned long) store_prefix(); pcpu->lowcore = (struct _lowcore *)(unsigned long) store_prefix();
pcpu->async_stack = S390_lowcore.async_stack - ASYNC_SIZE
+ STACK_FRAME_OVERHEAD + sizeof(struct pt_regs);
pcpu->panic_stack = S390_lowcore.panic_stack - PAGE_SIZE
+ STACK_FRAME_OVERHEAD + sizeof(struct pt_regs);
S390_lowcore.percpu_offset = __per_cpu_offset[0]; S390_lowcore.percpu_offset = __per_cpu_offset[0];
smp_cpu_set_polarization(0, POLARIZATION_UNKNOWN); smp_cpu_set_polarization(0, POLARIZATION_UNKNOWN);
set_cpu_present(0, true); set_cpu_present(0, true);
...@@ -1078,8 +1079,7 @@ static int smp_cpu_notify(struct notifier_block *self, unsigned long action, ...@@ -1078,8 +1079,7 @@ static int smp_cpu_notify(struct notifier_block *self, unsigned long action,
void *hcpu) void *hcpu)
{ {
unsigned int cpu = (unsigned int)(long)hcpu; unsigned int cpu = (unsigned int)(long)hcpu;
struct cpu *c = pcpu_devices[cpu].cpu; struct device *s = &per_cpu(cpu_device, cpu)->dev;
struct device *s = &c->dev;
int err = 0; int err = 0;
switch (action & ~CPU_TASKS_FROZEN) { switch (action & ~CPU_TASKS_FROZEN) {
...@@ -1102,7 +1102,7 @@ static int smp_add_present_cpu(int cpu) ...@@ -1102,7 +1102,7 @@ static int smp_add_present_cpu(int cpu)
c = kzalloc(sizeof(*c), GFP_KERNEL); c = kzalloc(sizeof(*c), GFP_KERNEL);
if (!c) if (!c)
return -ENOMEM; return -ENOMEM;
pcpu_devices[cpu].cpu = c; per_cpu(cpu_device, cpu) = c;
s = &c->dev; s = &c->dev;
c->hotpluggable = 1; c->hotpluggable = 1;
rc = register_cpu(c, cpu); rc = register_cpu(c, cpu);
......
...@@ -7,14 +7,14 @@ ...@@ -7,14 +7,14 @@
#define pr_fmt(fmt) KMSG_COMPONENT ": " fmt #define pr_fmt(fmt) KMSG_COMPONENT ": " fmt
#include <linux/workqueue.h> #include <linux/workqueue.h>
#include <linux/bootmem.h>
#include <linux/cpuset.h> #include <linux/cpuset.h>
#include <linux/device.h> #include <linux/device.h>
#include <linux/export.h> #include <linux/export.h>
#include <linux/kernel.h> #include <linux/kernel.h>
#include <linux/sched.h> #include <linux/sched.h>
#include <linux/init.h>
#include <linux/delay.h> #include <linux/delay.h>
#include <linux/init.h>
#include <linux/slab.h>
#include <linux/cpu.h> #include <linux/cpu.h>
#include <linux/smp.h> #include <linux/smp.h>
#include <linux/mm.h> #include <linux/mm.h>
...@@ -42,8 +42,8 @@ static DEFINE_SPINLOCK(topology_lock); ...@@ -42,8 +42,8 @@ static DEFINE_SPINLOCK(topology_lock);
static struct mask_info socket_info; static struct mask_info socket_info;
static struct mask_info book_info; static struct mask_info book_info;
struct cpu_topology_s390 cpu_topology[NR_CPUS]; DEFINE_PER_CPU(struct cpu_topology_s390, cpu_topology);
EXPORT_SYMBOL_GPL(cpu_topology); EXPORT_PER_CPU_SYMBOL_GPL(cpu_topology);
static cpumask_t cpu_group_map(struct mask_info *info, unsigned int cpu) static cpumask_t cpu_group_map(struct mask_info *info, unsigned int cpu)
{ {
...@@ -90,15 +90,15 @@ static struct mask_info *add_cpus_to_mask(struct topology_core *tl_core, ...@@ -90,15 +90,15 @@ static struct mask_info *add_cpus_to_mask(struct topology_core *tl_core,
if (lcpu < 0) if (lcpu < 0)
continue; continue;
for (i = 0; i <= smp_cpu_mtid; i++) { for (i = 0; i <= smp_cpu_mtid; i++) {
cpu_topology[lcpu + i].book_id = book->id; per_cpu(cpu_topology, lcpu + i).book_id = book->id;
cpu_topology[lcpu + i].core_id = rcore; per_cpu(cpu_topology, lcpu + i).core_id = rcore;
cpu_topology[lcpu + i].thread_id = lcpu + i; per_cpu(cpu_topology, lcpu + i).thread_id = lcpu + i;
cpumask_set_cpu(lcpu + i, &book->mask); cpumask_set_cpu(lcpu + i, &book->mask);
cpumask_set_cpu(lcpu + i, &socket->mask); cpumask_set_cpu(lcpu + i, &socket->mask);
if (one_socket_per_cpu) if (one_socket_per_cpu)
cpu_topology[lcpu + i].socket_id = rcore; per_cpu(cpu_topology, lcpu + i).socket_id = rcore;
else else
cpu_topology[lcpu + i].socket_id = socket->id; per_cpu(cpu_topology, lcpu + i).socket_id = socket->id;
smp_cpu_set_polarization(lcpu + i, tl_core->pp); smp_cpu_set_polarization(lcpu + i, tl_core->pp);
} }
if (one_socket_per_cpu) if (one_socket_per_cpu)
...@@ -249,14 +249,14 @@ static void update_cpu_masks(void) ...@@ -249,14 +249,14 @@ static void update_cpu_masks(void)
spin_lock_irqsave(&topology_lock, flags); spin_lock_irqsave(&topology_lock, flags);
for_each_possible_cpu(cpu) { for_each_possible_cpu(cpu) {
cpu_topology[cpu].thread_mask = cpu_thread_map(cpu); per_cpu(cpu_topology, cpu).thread_mask = cpu_thread_map(cpu);
cpu_topology[cpu].core_mask = cpu_group_map(&socket_info, cpu); per_cpu(cpu_topology, cpu).core_mask = cpu_group_map(&socket_info, cpu);
cpu_topology[cpu].book_mask = cpu_group_map(&book_info, cpu); per_cpu(cpu_topology, cpu).book_mask = cpu_group_map(&book_info, cpu);
if (!MACHINE_HAS_TOPOLOGY) { if (!MACHINE_HAS_TOPOLOGY) {
cpu_topology[cpu].thread_id = cpu; per_cpu(cpu_topology, cpu).thread_id = cpu;
cpu_topology[cpu].core_id = cpu; per_cpu(cpu_topology, cpu).core_id = cpu;
cpu_topology[cpu].socket_id = cpu; per_cpu(cpu_topology, cpu).socket_id = cpu;
cpu_topology[cpu].book_id = cpu; per_cpu(cpu_topology, cpu).book_id = cpu;
} }
} }
spin_unlock_irqrestore(&topology_lock, flags); spin_unlock_irqrestore(&topology_lock, flags);
...@@ -334,50 +334,6 @@ void topology_expect_change(void) ...@@ -334,50 +334,6 @@ void topology_expect_change(void)
set_topology_timer(); set_topology_timer();
} }
static int __init early_parse_topology(char *p)
{
if (strncmp(p, "off", 3))
return 0;
topology_enabled = 0;
return 0;
}
early_param("topology", early_parse_topology);
static void __init alloc_masks(struct sysinfo_15_1_x *info,
struct mask_info *mask, int offset)
{
int i, nr_masks;
nr_masks = info->mag[TOPOLOGY_NR_MAG - offset];
for (i = 0; i < info->mnest - offset; i++)
nr_masks *= info->mag[TOPOLOGY_NR_MAG - offset - 1 - i];
nr_masks = max(nr_masks, 1);
for (i = 0; i < nr_masks; i++) {
mask->next = alloc_bootmem_align(
roundup_pow_of_two(sizeof(struct mask_info)),
roundup_pow_of_two(sizeof(struct mask_info)));
mask = mask->next;
}
}
void __init s390_init_cpu_topology(void)
{
struct sysinfo_15_1_x *info;
int i;
if (!MACHINE_HAS_TOPOLOGY)
return;
tl_info = alloc_bootmem_pages(PAGE_SIZE);
info = tl_info;
store_topology(info);
pr_info("The CPU configuration topology of the machine is:");
for (i = 0; i < TOPOLOGY_NR_MAG; i++)
printk(KERN_CONT " %d", info->mag[i]);
printk(KERN_CONT " / %d\n", info->mnest);
alloc_masks(info, &socket_info, 1);
alloc_masks(info, &book_info, 2);
}
static int cpu_management; static int cpu_management;
static ssize_t dispatching_show(struct device *dev, static ssize_t dispatching_show(struct device *dev,
...@@ -467,20 +423,29 @@ int topology_cpu_init(struct cpu *cpu) ...@@ -467,20 +423,29 @@ int topology_cpu_init(struct cpu *cpu)
const struct cpumask *cpu_thread_mask(int cpu) const struct cpumask *cpu_thread_mask(int cpu)
{ {
return &cpu_topology[cpu].thread_mask; return &per_cpu(cpu_topology, cpu).thread_mask;
} }
const struct cpumask *cpu_coregroup_mask(int cpu) const struct cpumask *cpu_coregroup_mask(int cpu)
{ {
return &cpu_topology[cpu].core_mask; return &per_cpu(cpu_topology, cpu).core_mask;
} }
static const struct cpumask *cpu_book_mask(int cpu) static const struct cpumask *cpu_book_mask(int cpu)
{ {
return &cpu_topology[cpu].book_mask; return &per_cpu(cpu_topology, cpu).book_mask;
} }
static int __init early_parse_topology(char *p)
{
if (strncmp(p, "off", 3))
return 0;
topology_enabled = 0;
return 0;
}
early_param("topology", early_parse_topology);
static struct sched_domain_topology_level s390_topology[] = { static struct sched_domain_topology_level s390_topology[] = {
{ cpu_thread_mask, cpu_smt_flags, SD_INIT_NAME(SMT) }, { cpu_thread_mask, cpu_smt_flags, SD_INIT_NAME(SMT) },
{ cpu_coregroup_mask, cpu_core_flags, SD_INIT_NAME(MC) }, { cpu_coregroup_mask, cpu_core_flags, SD_INIT_NAME(MC) },
...@@ -489,6 +454,42 @@ static struct sched_domain_topology_level s390_topology[] = { ...@@ -489,6 +454,42 @@ static struct sched_domain_topology_level s390_topology[] = {
{ NULL, }, { NULL, },
}; };
static void __init alloc_masks(struct sysinfo_15_1_x *info,
struct mask_info *mask, int offset)
{
int i, nr_masks;
nr_masks = info->mag[TOPOLOGY_NR_MAG - offset];
for (i = 0; i < info->mnest - offset; i++)
nr_masks *= info->mag[TOPOLOGY_NR_MAG - offset - 1 - i];
nr_masks = max(nr_masks, 1);
for (i = 0; i < nr_masks; i++) {
mask->next = kzalloc(sizeof(*mask->next), GFP_KERNEL);
mask = mask->next;
}
}
static int __init s390_topology_init(void)
{
struct sysinfo_15_1_x *info;
int i;
if (!MACHINE_HAS_TOPOLOGY)
return 0;
tl_info = (struct sysinfo_15_1_x *)__get_free_page(GFP_KERNEL);
info = tl_info;
store_topology(info);
pr_info("The CPU configuration topology of the machine is:");
for (i = 0; i < TOPOLOGY_NR_MAG; i++)
printk(KERN_CONT " %d", info->mag[i]);
printk(KERN_CONT " / %d\n", info->mnest);
alloc_masks(info, &socket_info, 1);
alloc_masks(info, &book_info, 2);
set_sched_topology(s390_topology);
return 0;
}
early_initcall(s390_topology_init);
static int __init topology_init(void) static int __init topology_init(void)
{ {
if (MACHINE_HAS_TOPOLOGY) if (MACHINE_HAS_TOPOLOGY)
...@@ -498,10 +499,3 @@ static int __init topology_init(void) ...@@ -498,10 +499,3 @@ static int __init topology_init(void)
return device_create_file(cpu_subsys.dev_root, &dev_attr_dispatching); return device_create_file(cpu_subsys.dev_root, &dev_attr_dispatching);
} }
device_initcall(topology_init); device_initcall(topology_init);
static int __init early_topology_init(void)
{
set_sched_topology(s390_topology);
return 0;
}
early_initcall(early_topology_init);
...@@ -25,9 +25,7 @@ __kernel_clock_gettime: ...@@ -25,9 +25,7 @@ __kernel_clock_gettime:
je 4f je 4f
cghi %r2,__CLOCK_REALTIME cghi %r2,__CLOCK_REALTIME
je 5f je 5f
cghi %r2,__CLOCK_THREAD_CPUTIME_ID cghi %r2,-3 /* Per-thread CPUCLOCK with PID=0, VIRT=1 */
je 9f
cghi %r2,-2 /* Per-thread CPUCLOCK with PID=0, VIRT=1 */
je 9f je 9f
cghi %r2,__CLOCK_MONOTONIC_COARSE cghi %r2,__CLOCK_MONOTONIC_COARSE
je 3f je 3f
...@@ -106,7 +104,7 @@ __kernel_clock_gettime: ...@@ -106,7 +104,7 @@ __kernel_clock_gettime:
aghi %r15,16 aghi %r15,16
br %r14 br %r14
/* CLOCK_THREAD_CPUTIME_ID for this thread */ /* CPUCLOCK_VIRT for this thread */
9: icm %r0,15,__VDSO_ECTG_OK(%r5) 9: icm %r0,15,__VDSO_ECTG_OK(%r5)
jz 12f jz 12f
ear %r2,%a4 ear %r2,%a4
......
...@@ -183,7 +183,10 @@ unsigned long randomize_et_dyn(void) ...@@ -183,7 +183,10 @@ unsigned long randomize_et_dyn(void)
{ {
unsigned long base; unsigned long base;
base = (STACK_TOP / 3 * 2) & (~mmap_align_mask << PAGE_SHIFT); base = STACK_TOP / 3 * 2;
if (!is_32bit_task())
/* Align to 4GB */
base &= ~((1UL << 32) - 1);
return base + mmap_rnd(); return base + mmap_rnd();
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment