arm64: initialize per-cpu offsets earlier
authorMark Rutland <mark.rutland@arm.com>
Mon, 5 Oct 2020 16:43:03 +0000 (17:43 +0100)
committerWill Deacon <will@kernel.org>
Mon, 5 Oct 2020 17:54:49 +0000 (18:54 +0100)
commit353e228eb355be5a65a3c0996c774a0f46737fda
tree1bd5619d6b4765e8dd452f4144e20e0bfe04dc94
parent4dafc08d0ba4768e8540f49ab40c3ea26e40d554
arm64: initialize per-cpu offsets earlier

The current initialization of the per-cpu offset register is difficult
to follow and this initialization is not always early enough for
upcoming instrumentation with KCSAN, where the instrumentation callbacks
use the per-cpu offset.

To make it possible to support KCSAN, and to simplify reasoning about
early bringup code, let's initialize the per-cpu offset earlier, before
we run any C code that may consume it. To do so, this patch adds a new
init_this_cpu_offset() helper that's called before the usual
primary/secondary start functions. For consistency, this is also used to
re-initialize the per-cpu offset after the runtime per-cpu areas have
been allocated (which can change CPU0's offset).

So that init_this_cpu_offset() isn't subject to any instrumentation that
might consume the per-cpu offset, it is marked with noinstr, preventing
instrumentation.

Signed-off-by: Mark Rutland <mark.rutland@arm.com>
Cc: Catalin Marinas <catalin.marinas@arm.com>
Cc: James Morse <james.morse@arm.com>
Cc: Will Deacon <will@kernel.org>
Link: https://lore.kernel.org/r/20201005164303.21389-1-mark.rutland@arm.com
Signed-off-by: Will Deacon <will@kernel.org>
arch/arm64/include/asm/cpu.h
arch/arm64/kernel/head.S
arch/arm64/kernel/setup.c
arch/arm64/kernel/smp.c