diff options
author | Aaron Durbin <adurbin@chromium.org> | 2018-05-03 13:49:41 -0600 |
---|---|---|
committer | Patrick Georgi <pgeorgi@google.com> | 2018-05-14 08:39:18 +0000 |
commit | 223fb436fe89a265f4ddbe7f8fe38a35b47a8253 (patch) | |
tree | e4c4b0d95dadf15da2a8946ae50930c637d3f8c6 /src | |
parent | e6cc21e262aec4846831dd1da57d34cc8fc9099f (diff) | |
download | coreboot-223fb436fe89a265f4ddbe7f8fe38a35b47a8253.tar.xz |
cpu/x86/mp: pass pointers to structures for AP callbacks
In order to extend the MP callback infrastructure prepare for
easier changes by making the AP callback get signalled by a
single pointer to a local variable on the signaller's stack.
When the APs see the callback they will copy the structure
to a local variable and then set the acknowledgement by
clearing out the slot.
The reading and writing to the slots were implemented using inline
assembly which forces a memory access and a compiler barrier.
BUG=b:74436746
Change-Id: Ia46133a49c03ce3ce0e73ae3d30547316c7ec43c
Signed-off-by: Aaron Durbin <adurbin@chromium.org>
Reviewed-on: https://review.coreboot.org/26043
Reviewed-by: Subrata Banik <subrata.banik@intel.com>
Reviewed-by: Nico Huber <nico.h@gmx.de>
Tested-by: build bot (Jenkins) <no-reply@coreboot.org>
Diffstat (limited to 'src')
-rw-r--r-- | src/cpu/x86/mp_init.c | 51 |
1 files changed, 35 insertions, 16 deletions
diff --git a/src/cpu/x86/mp_init.c b/src/cpu/x86/mp_init.c index 88a8f36bad..f7cf0b418d 100644 --- a/src/cpu/x86/mp_init.c +++ b/src/cpu/x86/mp_init.c @@ -40,7 +40,9 @@ #define MAX_APIC_IDS 256 -typedef void (*mp_callback_t)(void); +struct mp_callback { + void (*func)(void); +}; /* * A mp_flight_record details a sequence of calls for the APs to perform @@ -58,8 +60,8 @@ typedef void (*mp_callback_t)(void); struct mp_flight_record { atomic_t barrier; atomic_t cpus_entered; - mp_callback_t ap_call; - mp_callback_t bsp_call; + void (*ap_call)(void); + void (*bsp_call)(void); } __aligned(CACHELINE_SIZE); #define _MP_FLIGHT_RECORD(barrier_, ap_func_, bsp_func_) \ @@ -851,19 +853,30 @@ static void trigger_smm_relocation(void) mp_state.ops.per_cpu_smm_trigger(); } -static mp_callback_t ap_callbacks[CONFIG_MAX_CPUS]; +static struct mp_callback *ap_callbacks[CONFIG_MAX_CPUS]; -static mp_callback_t read_callback(mp_callback_t *slot) +static struct mp_callback *read_callback(struct mp_callback **slot) { - return *(volatile mp_callback_t *)slot; + struct mp_callback *ret; + + asm volatile ("mov %1, %0\n" + : "=r" (ret) + : "m" (*slot) + : "memory" + ); + return ret; } -static void store_callback(mp_callback_t *slot, mp_callback_t value) +static void store_callback(struct mp_callback **slot, struct mp_callback *val) { - *(volatile mp_callback_t *)slot = value; + asm volatile ("mov %1, %0\n" + : "=m" (*slot) + : "r" (val) + : "memory" + ); } -static int run_ap_work(mp_callback_t func, long expire_us) +static int run_ap_work(struct mp_callback *val, long expire_us) { int i; int cpus_accepted; @@ -879,7 +892,7 @@ static int run_ap_work(mp_callback_t func, long expire_us) for (i = 0; i < ARRAY_SIZE(ap_callbacks); i++) { if (cur_cpu == i) continue; - store_callback(&ap_callbacks[i], func); + store_callback(&ap_callbacks[i], val); } mfence(); @@ -908,28 +921,34 @@ static int run_ap_work(mp_callback_t func, long expire_us) static void ap_wait_for_instruction(void) { - int cur_cpu = cpu_index(); + struct mp_callback lcb; + struct mp_callback **per_cpu_slot; if (!IS_ENABLED(CONFIG_PARALLEL_MP_AP_WORK)) return; + per_cpu_slot = &ap_callbacks[cpu_index()]; + while (1) { - mp_callback_t func = read_callback(&ap_callbacks[cur_cpu]); + struct mp_callback *cb = read_callback(per_cpu_slot); - if (func == NULL) { + if (cb == NULL) { asm ("pause"); continue; } - store_callback(&ap_callbacks[cur_cpu], NULL); + /* Copy to local variable before signalling consumption. */ + memcpy(&lcb, cb, sizeof(lcb)); mfence(); - func(); + store_callback(per_cpu_slot, NULL); + lcb.func(); } } int mp_run_on_aps(void (*func)(void), long expire_us) { - return run_ap_work(func, expire_us); + struct mp_callback lcb = { .func = func }; + return run_ap_work(&lcb, expire_us); } int mp_run_on_all_cpus(void (*func)(void), long expire_us) |