mirror of
https://github.com/AsahiLinux/u-boot
synced 2024-11-18 02:38:56 +00:00
401d1c4f5d
Move this out of the common header and include it only where needed. In a number of cases this requires adding "struct udevice;" to avoid adding another large header or in other cases replacing / adding missing header files that had been pulled in, very indirectly. Finally, we have a few cases where we did not need to include <asm/global_data.h> at all, so remove that include. Signed-off-by: Simon Glass <sjg@chromium.org> Signed-off-by: Tom Rini <trini@konsulko.com>
195 lines
5.6 KiB
C
195 lines
5.6 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
/*
|
|
* Copyright (c) 2015 Google, Inc
|
|
*
|
|
* Taken from coreboot file of the same name
|
|
*/
|
|
|
|
#ifndef _X86_MP_H_
|
|
#define _X86_MP_H_
|
|
|
|
#include <asm/atomic.h>
|
|
#include <asm/cache.h>
|
|
|
|
struct udevice;
|
|
|
|
enum {
|
|
/* Indicates that the function should run on all CPUs */
|
|
MP_SELECT_ALL = -1,
|
|
|
|
/* Run on boot CPUs */
|
|
MP_SELECT_BSP = -2,
|
|
|
|
/* Run on non-boot CPUs */
|
|
MP_SELECT_APS = -3,
|
|
};
|
|
|
|
typedef int (*mp_callback_t)(struct udevice *cpu, void *arg);
|
|
|
|
/*
|
|
* A mp_flight_record details a sequence of calls for the APs to perform
|
|
* along with the BSP to coordinate sequencing. Each flight record either
|
|
* provides a barrier for each AP before calling the callback or the APs
|
|
* are allowed to perform the callback without waiting. Regardless, each
|
|
* record has the cpus_entered field incremented for each record. When
|
|
* the BSP observes that the cpus_entered matches the number of APs
|
|
* the bsp_call is called with bsp_arg and upon returning releases the
|
|
* barrier allowing the APs to make further progress.
|
|
*
|
|
* Note that ap_call() and bsp_call() can be NULL. In the NULL case the
|
|
* callback will just not be called.
|
|
*
|
|
* @barrier: Ensures that the BSP and AP don't run the flight record at the same
|
|
* time
|
|
* @cpus_entered: Counts the number of APs that have run this record
|
|
* @ap_call: Function for the APs to call
|
|
* @ap_arg: Argument to pass to @ap_call
|
|
* @bsp_call: Function for the BSP to call
|
|
* @bsp_arg: Argument to pass to @bsp_call
|
|
*/
|
|
struct mp_flight_record {
|
|
atomic_t barrier;
|
|
atomic_t cpus_entered;
|
|
mp_callback_t ap_call;
|
|
void *ap_arg;
|
|
mp_callback_t bsp_call;
|
|
void *bsp_arg;
|
|
} __attribute__((aligned(ARCH_DMA_MINALIGN)));
|
|
|
|
#define MP_FLIGHT_RECORD(barrier_, ap_func_, ap_arg_, bsp_func_, bsp_arg_) \
|
|
{ \
|
|
.barrier = ATOMIC_INIT(barrier_), \
|
|
.cpus_entered = ATOMIC_INIT(0), \
|
|
.ap_call = ap_func_, \
|
|
.ap_arg = ap_arg_, \
|
|
.bsp_call = bsp_func_, \
|
|
.bsp_arg = bsp_arg_, \
|
|
}
|
|
|
|
#define MP_FR_BLOCK_APS(ap_func, ap_arg, bsp_func, bsp_arg) \
|
|
MP_FLIGHT_RECORD(0, ap_func, ap_arg, bsp_func, bsp_arg)
|
|
|
|
#define MP_FR_NOBLOCK_APS(ap_func, ap_arg, bsp_func, bsp_arg) \
|
|
MP_FLIGHT_RECORD(1, ap_func, ap_arg, bsp_func, bsp_arg)
|
|
|
|
/*
|
|
* mp_init() will set up the SIPI vector and bring up the APs according to
|
|
* mp_params. Each flight record will be executed according to the plan. Note
|
|
* that the MP infrastructure uses SMM default area without saving it. It's
|
|
* up to the chipset or mainboard to either e820 reserve this area or save this
|
|
* region prior to calling mp_init() and restoring it after mp_init returns.
|
|
*
|
|
* At the time mp_init() is called the MTRR MSRs are mirrored into APs then
|
|
* caching is enabled before running the flight plan.
|
|
*
|
|
* The MP init has the following properties:
|
|
* 1. APs are brought up in parallel.
|
|
* 2. The ordering of cpu number and APIC ids is not deterministic.
|
|
* Therefore, one cannot rely on this property or the order of devices in
|
|
* the device tree unless the chipset or mainboard know the APIC ids
|
|
* a priori.
|
|
*
|
|
* mp_init() returns < 0 on error, 0 on success.
|
|
*/
|
|
int mp_init(void);
|
|
|
|
/**
|
|
* x86_mp_init() - Set up additional CPUs
|
|
*
|
|
* @returns < 0 on error, 0 on success.
|
|
*/
|
|
int x86_mp_init(void);
|
|
|
|
/**
|
|
* mp_run_func() - Function to call on the AP
|
|
*
|
|
* @arg: Argument to pass
|
|
*/
|
|
typedef void (*mp_run_func)(void *arg);
|
|
|
|
#if CONFIG_IS_ENABLED(SMP) && !CONFIG_IS_ENABLED(X86_64)
|
|
/**
|
|
* mp_run_on_cpus() - Run a function on one or all CPUs
|
|
*
|
|
* This does not return until all CPUs have completed the work
|
|
*
|
|
* Running on anything other than the boot CPU is only supported if
|
|
* CONFIG_SMP_AP_WORK is enabled
|
|
*
|
|
* @cpu_select: CPU to run on (its dev_seq() value), or MP_SELECT_ALL for
|
|
* all, or MP_SELECT_BSP for BSP
|
|
* @func: Function to run
|
|
* @arg: Argument to pass to the function
|
|
* @return 0 on success, -ve on error
|
|
*/
|
|
int mp_run_on_cpus(int cpu_select, mp_run_func func, void *arg);
|
|
|
|
/**
|
|
* mp_park_aps() - Park the APs ready for the OS
|
|
*
|
|
* This halts all CPUs except the main one, ready for the OS to use them
|
|
*
|
|
* @return 0 if OK, -ve on error
|
|
*/
|
|
int mp_park_aps(void);
|
|
|
|
/**
|
|
* mp_first_cpu() - Get the first CPU to process, from a selection
|
|
*
|
|
* This is used to iterate through selected CPUs. Call this function first, then
|
|
* call mp_next_cpu() repeatedly (with the same @cpu_select) until it returns
|
|
* -EFBIG.
|
|
*
|
|
* @cpu_select: Selected CPUs (either a CPU number or MP_SELECT_...)
|
|
* @return next CPU number to run on (e.g. 0)
|
|
*/
|
|
int mp_first_cpu(int cpu_select);
|
|
|
|
/**
|
|
* mp_next_cpu() - Get the next CPU to process, from a selection
|
|
*
|
|
* This is used to iterate through selected CPUs. After first calling
|
|
* mp_first_cpu() once, call this function repeatedly until it returns -EFBIG.
|
|
*
|
|
* The value of @cpu_select must be the same for all calls and must match the
|
|
* value passed to mp_first_cpu(), otherwise the behaviour is undefined.
|
|
*
|
|
* @cpu_select: Selected CPUs (either a CPU number or MP_SELECT_...)
|
|
* @prev_cpu: Previous value returned by mp_first_cpu()/mp_next_cpu()
|
|
* @return next CPU number to run on (e.g. 0)
|
|
*/
|
|
int mp_next_cpu(int cpu_select, int prev_cpu);
|
|
#else
|
|
static inline int mp_run_on_cpus(int cpu_select, mp_run_func func, void *arg)
|
|
{
|
|
/* There is only one CPU, so just call the function here */
|
|
func(arg);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static inline int mp_park_aps(void)
|
|
{
|
|
/* No APs to park */
|
|
|
|
return 0;
|
|
}
|
|
|
|
static inline int mp_first_cpu(int cpu_select)
|
|
{
|
|
/* We cannot run on any APs, nor a selected CPU */
|
|
return cpu_select == MP_SELECT_APS ? -EFBIG : MP_SELECT_BSP;
|
|
}
|
|
|
|
static inline int mp_next_cpu(int cpu_select, int prev_cpu)
|
|
{
|
|
/*
|
|
* When MP is not enabled, there is only one CPU and we did it in
|
|
* mp_first_cpu()
|
|
*/
|
|
return -EFBIG;
|
|
}
|
|
|
|
#endif
|
|
|
|
#endif /* _X86_MP_H_ */
|