s390: unify stack size definitions
authorVasily Gorbik <gor@linux.ibm.com>
Tue, 18 Sep 2018 16:23:40 +0000 (18:23 +0200)
committerMartin Schwidefsky <schwidefsky@de.ibm.com>
Tue, 9 Oct 2018 09:20:58 +0000 (11:20 +0200)
Remove STACK_ORDER and STACK_SIZE in favour of identical THREAD_SIZE_ORDER
and THREAD_SIZE definitions. THREAD_SIZE and THREAD_SIZE_ORDER naming is
misleading since it is used as general kernel stack size information. But
both those definitions are used in the common code and throughout
architectures specific code, so changing the naming is problematic.

Reviewed-by: Martin Schwidefsky <schwidefsky@de.ibm.com>
Signed-off-by: Vasily Gorbik <gor@linux.ibm.com>
Signed-off-by: Martin Schwidefsky <schwidefsky@de.ibm.com>
arch/s390/include/asm/processor.h
arch/s390/include/asm/thread_info.h
arch/s390/kernel/dumpstack.c
arch/s390/kernel/irq.c
arch/s390/kernel/setup.c
arch/s390/kernel/smp.c

index 3c1e723..43494a0 100644 (file)
@@ -162,14 +162,6 @@ struct thread_struct {
 
 typedef struct thread_struct thread_struct;
 
-/*
- * General size of a stack
- */
-#define STACK_ORDER 2
-#define STACK_SIZE (PAGE_SIZE << STACK_ORDER)
-#define STACK_INIT_OFFSET \
-       (STACK_SIZE - STACK_FRAME_OVERHEAD - sizeof(struct pt_regs))
-
 /*
  * Stack layout of a C stack frame.
  */
index 3fa2fea..1bbbaf6 100644 (file)
@@ -11,7 +11,7 @@
 #include <linux/const.h>
 
 /*
- * Size of kernel stack for each process
+ * General size of kernel stacks
  */
 #define THREAD_SIZE_ORDER 2
 #define THREAD_SIZE (PAGE_SIZE << THREAD_SIZE_ORDER)
@@ -21,6 +21,9 @@
 #include <asm/page.h>
 #include <asm/processor.h>
 
+#define STACK_INIT_OFFSET \
+       (THREAD_SIZE - STACK_FRAME_OVERHEAD - sizeof(struct pt_regs))
+
 /*
  * low level task data that entry.S needs immediate access to
  * - this struct should fit entirely inside of one cache line
index 301b945..ef85a00 100644 (file)
@@ -77,11 +77,11 @@ void dump_trace(dump_trace_func_t func, void *data, struct task_struct *task,
        frame_size = STACK_FRAME_OVERHEAD + sizeof(struct pt_regs);
 #ifdef CONFIG_CHECK_STACK
        sp = __dump_trace(func, data, sp,
-                         S390_lowcore.nodat_stack + frame_size - STACK_SIZE,
+                         S390_lowcore.nodat_stack + frame_size - THREAD_SIZE,
                          S390_lowcore.nodat_stack + frame_size);
 #endif
        sp = __dump_trace(func, data, sp,
-                         S390_lowcore.async_stack + frame_size - STACK_SIZE,
+                         S390_lowcore.async_stack + frame_size - THREAD_SIZE,
                          S390_lowcore.async_stack + frame_size);
        task = task ?: current;
        __dump_trace(func, data, sp,
index b2bc0eb..0e8d68b 100644 (file)
@@ -171,7 +171,7 @@ void do_softirq_own_stack(void)
        old = current_stack_pointer();
        /* Check against async. stack address range. */
        new = S390_lowcore.async_stack;
-       if (((new - old) >> (PAGE_SHIFT + STACK_ORDER)) != 0) {
+       if (((new - old) >> (PAGE_SHIFT + THREAD_SIZE_ORDER)) != 0) {
                CALL_ON_STACK(__do_softirq, new, 0);
        } else {
                /* We are already on the async stack. */
index eca51c4..67fa7cb 100644 (file)
@@ -308,13 +308,13 @@ unsigned long stack_alloc(void)
 {
 #ifdef CONFIG_VMAP_STACK
        return (unsigned long)
-               __vmalloc_node_range(STACK_SIZE, STACK_SIZE,
+               __vmalloc_node_range(THREAD_SIZE, THREAD_SIZE,
                                     VMALLOC_START, VMALLOC_END,
                                     THREADINFO_GFP,
                                     PAGE_KERNEL, 0, NUMA_NO_NODE,
                                     __builtin_return_address(0));
 #else
-       return __get_free_pages(GFP_KERNEL, STACK_ORDER);
+       return __get_free_pages(GFP_KERNEL, THREAD_SIZE_ORDER);
 #endif
 }
 
@@ -323,7 +323,7 @@ void stack_free(unsigned long stack)
 #ifdef CONFIG_VMAP_STACK
        vfree((void *) stack);
 #else
-       free_pages(stack, STACK_ORDER);
+       free_pages(stack, THREAD_SIZE_ORDER);
 #endif
 }
 
@@ -331,7 +331,7 @@ int __init arch_early_irq_init(void)
 {
        unsigned long stack;
 
-       stack = __get_free_pages(GFP_KERNEL, STACK_ORDER);
+       stack = __get_free_pages(GFP_KERNEL, THREAD_SIZE_ORDER);
        if (!stack)
                panic("Couldn't allocate async stack");
        S390_lowcore.async_stack = stack + STACK_INIT_OFFSET;
@@ -347,7 +347,7 @@ static int __init async_stack_realloc(void)
        if (!new)
                panic("Couldn't allocate async stack");
        S390_lowcore.async_stack = new + STACK_INIT_OFFSET;
-       free_pages(old, STACK_ORDER);
+       free_pages(old, THREAD_SIZE_ORDER);
        return 0;
 }
 early_initcall(async_stack_realloc);
@@ -428,7 +428,7 @@ static void __init setup_lowcore(void)
         * Allocate the global restart stack which is the same for
         * all CPUs in cast *one* of them does a PSW restart.
         */
-       restart_stack = memblock_virt_alloc(STACK_SIZE, STACK_SIZE);
+       restart_stack = memblock_virt_alloc(THREAD_SIZE, THREAD_SIZE);
        restart_stack += STACK_INIT_OFFSET;
 
        /*
index fccdb96..032d98b 100644 (file)
@@ -194,7 +194,7 @@ static int pcpu_alloc_lowcore(struct pcpu *pcpu, int cpu)
        if (pcpu != &pcpu_devices[0]) {
                pcpu->lowcore = (struct lowcore *)
                        __get_free_pages(GFP_KERNEL | GFP_DMA, LC_ORDER);
-               nodat_stack = __get_free_pages(GFP_KERNEL, STACK_ORDER);
+               nodat_stack = __get_free_pages(GFP_KERNEL, THREAD_SIZE_ORDER);
                if (!pcpu->lowcore || !nodat_stack)
                        goto out;
        } else {
@@ -226,7 +226,7 @@ out_async:
        stack_free(async_stack);
 out:
        if (pcpu != &pcpu_devices[0]) {
-               free_pages(nodat_stack, STACK_ORDER);
+               free_pages(nodat_stack, THREAD_SIZE_ORDER);
                free_pages((unsigned long) pcpu->lowcore, LC_ORDER);
        }
        return -ENOMEM;
@@ -249,7 +249,7 @@ static void pcpu_free_lowcore(struct pcpu *pcpu)
        stack_free(async_stack);
        if (pcpu == &pcpu_devices[0])
                return;
-       free_pages(nodat_stack, STACK_ORDER);
+       free_pages(nodat_stack, THREAD_SIZE_ORDER);
        free_pages(lowcore, LC_ORDER);
 }