summaryrefslogtreecommitdiff
path: root/UefiCpuPkg/PiSmmCpuDxeSmm/MpService.c
diff options
context:
space:
mode:
Diffstat (limited to 'UefiCpuPkg/PiSmmCpuDxeSmm/MpService.c')
-rw-r--r--UefiCpuPkg/PiSmmCpuDxeSmm/MpService.c1410
1 files changed, 0 insertions, 1410 deletions
diff --git a/UefiCpuPkg/PiSmmCpuDxeSmm/MpService.c b/UefiCpuPkg/PiSmmCpuDxeSmm/MpService.c
deleted file mode 100644
index f14b471bcf..0000000000
--- a/UefiCpuPkg/PiSmmCpuDxeSmm/MpService.c
+++ /dev/null
@@ -1,1410 +0,0 @@
-/** @file
-SMM MP service implementation
-
-Copyright (c) 2009 - 2016, Intel Corporation. All rights reserved.<BR>
-This program and the accompanying materials
-are licensed and made available under the terms and conditions of the BSD License
-which accompanies this distribution. The full text of the license may be found at
-http://opensource.org/licenses/bsd-license.php
-
-THE PROGRAM IS DISTRIBUTED UNDER THE BSD LICENSE ON AN "AS IS" BASIS,
-WITHOUT WARRANTIES OR REPRESENTATIONS OF ANY KIND, EITHER EXPRESS OR IMPLIED.
-
-**/
-
-#include "PiSmmCpuDxeSmm.h"
-
-//
-// Slots for all MTRR( FIXED MTRR + VARIABLE MTRR + MTRR_LIB_IA32_MTRR_DEF_TYPE)
-//
-UINT64 gSmiMtrrs[MTRR_NUMBER_OF_FIXED_MTRR + 2 * MTRR_NUMBER_OF_VARIABLE_MTRR + 1];
-UINT64 gPhyMask;
-SMM_DISPATCHER_MP_SYNC_DATA *mSmmMpSyncData = NULL;
-UINTN mSmmMpSyncDataSize;
-SMM_CPU_SEMAPHORES mSmmCpuSemaphores;
-UINTN mSemaphoreSize;
-SPIN_LOCK *mPFLock = NULL;
-
-/**
- Performs an atomic compare exchange operation to get semaphore.
- The compare exchange operation must be performed using
- MP safe mechanisms.
-
- @param Sem IN: 32-bit unsigned integer
- OUT: original integer - 1
- @return Original integer - 1
-
-**/
-UINT32
-WaitForSemaphore (
- IN OUT volatile UINT32 *Sem
- )
-{
- UINT32 Value;
-
- do {
- Value = *Sem;
- } while (Value == 0 ||
- InterlockedCompareExchange32 (
- (UINT32*)Sem,
- Value,
- Value - 1
- ) != Value);
- return Value - 1;
-}
-
-
-/**
- Performs an atomic compare exchange operation to release semaphore.
- The compare exchange operation must be performed using
- MP safe mechanisms.
-
- @param Sem IN: 32-bit unsigned integer
- OUT: original integer + 1
- @return Original integer + 1
-
-**/
-UINT32
-ReleaseSemaphore (
- IN OUT volatile UINT32 *Sem
- )
-{
- UINT32 Value;
-
- do {
- Value = *Sem;
- } while (Value + 1 != 0 &&
- InterlockedCompareExchange32 (
- (UINT32*)Sem,
- Value,
- Value + 1
- ) != Value);
- return Value + 1;
-}
-
-/**
- Performs an atomic compare exchange operation to lock semaphore.
- The compare exchange operation must be performed using
- MP safe mechanisms.
-
- @param Sem IN: 32-bit unsigned integer
- OUT: -1
- @return Original integer
-
-**/
-UINT32
-LockdownSemaphore (
- IN OUT volatile UINT32 *Sem
- )
-{
- UINT32 Value;
-
- do {
- Value = *Sem;
- } while (InterlockedCompareExchange32 (
- (UINT32*)Sem,
- Value, (UINT32)-1
- ) != Value);
- return Value;
-}
-
-/**
- Wait all APs to performs an atomic compare exchange operation to release semaphore.
-
- @param NumberOfAPs AP number
-
-**/
-VOID
-WaitForAllAPs (
- IN UINTN NumberOfAPs
- )
-{
- UINTN BspIndex;
-
- BspIndex = mSmmMpSyncData->BspIndex;
- while (NumberOfAPs-- > 0) {
- WaitForSemaphore (mSmmMpSyncData->CpuData[BspIndex].Run);
- }
-}
-
-/**
- Performs an atomic compare exchange operation to release semaphore
- for each AP.
-
-**/
-VOID
-ReleaseAllAPs (
- VOID
- )
-{
- UINTN Index;
- UINTN BspIndex;
-
- BspIndex = mSmmMpSyncData->BspIndex;
- for (Index = mMaxNumberOfCpus; Index-- > 0;) {
- if (Index != BspIndex && *(mSmmMpSyncData->CpuData[Index].Present)) {
- ReleaseSemaphore (mSmmMpSyncData->CpuData[Index].Run);
- }
- }
-}
-
-/**
- Checks if all CPUs (with certain exceptions) have checked in for this SMI run
-
- @param Exceptions CPU Arrival exception flags.
-
- @retval TRUE if all CPUs the have checked in.
- @retval FALSE if at least one Normal AP hasn't checked in.
-
-**/
-BOOLEAN
-AllCpusInSmmWithExceptions (
- SMM_CPU_ARRIVAL_EXCEPTIONS Exceptions
- )
-{
- UINTN Index;
- SMM_CPU_DATA_BLOCK *CpuData;
- EFI_PROCESSOR_INFORMATION *ProcessorInfo;
-
- ASSERT (*mSmmMpSyncData->Counter <= mNumberOfCpus);
-
- if (*mSmmMpSyncData->Counter == mNumberOfCpus) {
- return TRUE;
- }
-
- CpuData = mSmmMpSyncData->CpuData;
- ProcessorInfo = gSmmCpuPrivate->ProcessorInfo;
- for (Index = mMaxNumberOfCpus; Index-- > 0;) {
- if (!(*(CpuData[Index].Present)) && ProcessorInfo[Index].ProcessorId != INVALID_APIC_ID) {
- if (((Exceptions & ARRIVAL_EXCEPTION_DELAYED) != 0) && SmmCpuFeaturesGetSmmRegister (Index, SmmRegSmmDelayed) != 0) {
- continue;
- }
- if (((Exceptions & ARRIVAL_EXCEPTION_BLOCKED) != 0) && SmmCpuFeaturesGetSmmRegister (Index, SmmRegSmmBlocked) != 0) {
- continue;
- }
- if (((Exceptions & ARRIVAL_EXCEPTION_SMI_DISABLED) != 0) && SmmCpuFeaturesGetSmmRegister (Index, SmmRegSmmEnable) != 0) {
- continue;
- }
- return FALSE;
- }
- }
-
-
- return TRUE;
-}
-
-
-/**
- Given timeout constraint, wait for all APs to arrive, and insure when this function returns, no AP will execute normal mode code before
- entering SMM, except SMI disabled APs.
-
-**/
-VOID
-SmmWaitForApArrival (
- VOID
- )
-{
- UINT64 Timer;
- UINTN Index;
-
- ASSERT (*mSmmMpSyncData->Counter <= mNumberOfCpus);
-
- //
- // Platform implementor should choose a timeout value appropriately:
- // - The timeout value should balance the SMM time constrains and the likelihood that delayed CPUs are excluded in the SMM run. Note
- // the SMI Handlers must ALWAYS take into account the cases that not all APs are available in an SMI run.
- // - The timeout value must, in the case of 2nd timeout, be at least long enough to give time for all APs to receive the SMI IPI
- // and either enter SMM or buffer the SMI, to insure there is no CPU running normal mode code when SMI handling starts. This will
- // be TRUE even if a blocked CPU is brought out of the blocked state by a normal mode CPU (before the normal mode CPU received the
- // SMI IPI), because with a buffered SMI, and CPU will enter SMM immediately after it is brought out of the blocked state.
- // - The timeout value must be longer than longest possible IO operation in the system
- //
-
- //
- // Sync with APs 1st timeout
- //
- for (Timer = StartSyncTimer ();
- !IsSyncTimerTimeout (Timer) &&
- !AllCpusInSmmWithExceptions (ARRIVAL_EXCEPTION_BLOCKED | ARRIVAL_EXCEPTION_SMI_DISABLED );
- ) {
- CpuPause ();
- }
-
- //
- // Not all APs have arrived, so we need 2nd round of timeout. IPIs should be sent to ALL none present APs,
- // because:
- // a) Delayed AP may have just come out of the delayed state. Blocked AP may have just been brought out of blocked state by some AP running
- // normal mode code. These APs need to be guaranteed to have an SMI pending to insure that once they are out of delayed / blocked state, they
- // enter SMI immediately without executing instructions in normal mode. Note traditional flow requires there are no APs doing normal mode
- // work while SMI handling is on-going.
- // b) As a consequence of SMI IPI sending, (spurious) SMI may occur after this SMM run.
- // c) ** NOTE **: Use SMI disabling feature VERY CAREFULLY (if at all) for traditional flow, because a processor in SMI-disabled state
- // will execute normal mode code, which breaks the traditional SMI handlers' assumption that no APs are doing normal
- // mode work while SMI handling is on-going.
- // d) We don't add code to check SMI disabling status to skip sending IPI to SMI disabled APs, because:
- // - In traditional flow, SMI disabling is discouraged.
- // - In relaxed flow, CheckApArrival() will check SMI disabling status before calling this function.
- // In both cases, adding SMI-disabling checking code increases overhead.
- //
- if (*mSmmMpSyncData->Counter < mNumberOfCpus) {
- //
- // Send SMI IPIs to bring outside processors in
- //
- for (Index = mMaxNumberOfCpus; Index-- > 0;) {
- if (!(*(mSmmMpSyncData->CpuData[Index].Present)) && gSmmCpuPrivate->ProcessorInfo[Index].ProcessorId != INVALID_APIC_ID) {
- SendSmiIpi ((UINT32)gSmmCpuPrivate->ProcessorInfo[Index].ProcessorId);
- }
- }
-
- //
- // Sync with APs 2nd timeout.
- //
- for (Timer = StartSyncTimer ();
- !IsSyncTimerTimeout (Timer) &&
- !AllCpusInSmmWithExceptions (ARRIVAL_EXCEPTION_BLOCKED | ARRIVAL_EXCEPTION_SMI_DISABLED );
- ) {
- CpuPause ();
- }
- }
-
- return;
-}
-
-
-/**
- Replace OS MTRR's with SMI MTRR's.
-
- @param CpuIndex Processor Index
-
-**/
-VOID
-ReplaceOSMtrrs (
- IN UINTN CpuIndex
- )
-{
- PROCESSOR_SMM_DESCRIPTOR *Psd;
- UINT64 *SmiMtrrs;
- MTRR_SETTINGS *BiosMtrr;
-
- Psd = (PROCESSOR_SMM_DESCRIPTOR*)(mCpuHotPlugData.SmBase[CpuIndex] + SMM_PSD_OFFSET);
- SmiMtrrs = (UINT64*)(UINTN)Psd->MtrrBaseMaskPtr;
-
- SmmCpuFeaturesDisableSmrr ();
-
- //
- // Replace all MTRRs registers
- //
- BiosMtrr = (MTRR_SETTINGS*)SmiMtrrs;
- MtrrSetAllMtrrs(BiosMtrr);
-}
-
-/**
- SMI handler for BSP.
-
- @param CpuIndex BSP processor Index
- @param SyncMode SMM MP sync mode
-
-**/
-VOID
-BSPHandler (
- IN UINTN CpuIndex,
- IN SMM_CPU_SYNC_MODE SyncMode
- )
-{
- UINTN Index;
- MTRR_SETTINGS Mtrrs;
- UINTN ApCount;
- BOOLEAN ClearTopLevelSmiResult;
- UINTN PresentCount;
-
- ASSERT (CpuIndex == mSmmMpSyncData->BspIndex);
- ApCount = 0;
-
- //
- // Flag BSP's presence
- //
- *mSmmMpSyncData->InsideSmm = TRUE;
-
- //
- // Initialize Debug Agent to start source level debug in BSP handler
- //
- InitializeDebugAgent (DEBUG_AGENT_INIT_ENTER_SMI, NULL, NULL);
-
- //
- // Mark this processor's presence
- //
- *(mSmmMpSyncData->CpuData[CpuIndex].Present) = TRUE;
-
- //
- // Clear platform top level SMI status bit before calling SMI handlers. If
- // we cleared it after SMI handlers are run, we would miss the SMI that
- // occurs after SMI handlers are done and before SMI status bit is cleared.
- //
- ClearTopLevelSmiResult = ClearTopLevelSmiStatus();
- ASSERT (ClearTopLevelSmiResult == TRUE);
-
- //
- // Set running processor index
- //
- gSmmCpuPrivate->SmmCoreEntryContext.CurrentlyExecutingCpu = CpuIndex;
-
- //
- // If Traditional Sync Mode or need to configure MTRRs: gather all available APs.
- //
- if (SyncMode == SmmCpuSyncModeTradition || SmmCpuFeaturesNeedConfigureMtrrs()) {
-
- //
- // Wait for APs to arrive
- //
- SmmWaitForApArrival();
-
- //
- // Lock the counter down and retrieve the number of APs
- //
- *mSmmMpSyncData->AllCpusInSync = TRUE;
- ApCount = LockdownSemaphore (mSmmMpSyncData->Counter) - 1;
-
- //
- // Wait for all APs to get ready for programming MTRRs
- //
- WaitForAllAPs (ApCount);
-
- if (SmmCpuFeaturesNeedConfigureMtrrs()) {
- //
- // Signal all APs it's time for backup MTRRs
- //
- ReleaseAllAPs ();
-
- //
- // WaitForSemaphore() may wait for ever if an AP happens to enter SMM at
- // exactly this point. Please make sure PcdCpuSmmMaxSyncLoops has been set
- // to a large enough value to avoid this situation.
- // Note: For HT capable CPUs, threads within a core share the same set of MTRRs.
- // We do the backup first and then set MTRR to avoid race condition for threads
- // in the same core.
- //
- MtrrGetAllMtrrs(&Mtrrs);
-
- //
- // Wait for all APs to complete their MTRR saving
- //
- WaitForAllAPs (ApCount);
-
- //
- // Let all processors program SMM MTRRs together
- //
- ReleaseAllAPs ();
-
- //
- // WaitForSemaphore() may wait for ever if an AP happens to enter SMM at
- // exactly this point. Please make sure PcdCpuSmmMaxSyncLoops has been set
- // to a large enough value to avoid this situation.
- //
- ReplaceOSMtrrs (CpuIndex);
-
- //
- // Wait for all APs to complete their MTRR programming
- //
- WaitForAllAPs (ApCount);
- }
- }
-
- //
- // The BUSY lock is initialized to Acquired state
- //
- AcquireSpinLockOrFail (mSmmMpSyncData->CpuData[CpuIndex].Busy);
-
- //
- // Perform the pre tasks
- //
- PerformPreTasks ();
-
- //
- // Invoke SMM Foundation EntryPoint with the processor information context.
- //
- gSmmCpuPrivate->SmmCoreEntry (&gSmmCpuPrivate->SmmCoreEntryContext);
-
- //
- // Make sure all APs have completed their pending none-block tasks
- //
- for (Index = mMaxNumberOfCpus; Index-- > 0;) {
- if (Index != CpuIndex && *(mSmmMpSyncData->CpuData[Index].Present)) {
- AcquireSpinLock (mSmmMpSyncData->CpuData[Index].Busy);
- ReleaseSpinLock (mSmmMpSyncData->CpuData[Index].Busy);
- }
- }
-
- //
- // Perform the remaining tasks
- //
- PerformRemainingTasks ();
-
- //
- // If Relaxed-AP Sync Mode: gather all available APs after BSP SMM handlers are done, and
- // make those APs to exit SMI synchronously. APs which arrive later will be excluded and
- // will run through freely.
- //
- if (SyncMode != SmmCpuSyncModeTradition && !SmmCpuFeaturesNeedConfigureMtrrs()) {
-
- //
- // Lock the counter down and retrieve the number of APs
- //
- *mSmmMpSyncData->AllCpusInSync = TRUE;
- ApCount = LockdownSemaphore (mSmmMpSyncData->Counter) - 1;
- //
- // Make sure all APs have their Present flag set
- //
- while (TRUE) {
- PresentCount = 0;
- for (Index = mMaxNumberOfCpus; Index-- > 0;) {
- if (*(mSmmMpSyncData->CpuData[Index].Present)) {
- PresentCount ++;
- }
- }
- if (PresentCount > ApCount) {
- break;
- }
- }
- }
-
- //
- // Notify all APs to exit
- //
- *mSmmMpSyncData->InsideSmm = FALSE;
- ReleaseAllAPs ();
-
- //
- // Wait for all APs to complete their pending tasks
- //
- WaitForAllAPs (ApCount);
-
- if (SmmCpuFeaturesNeedConfigureMtrrs()) {
- //
- // Signal APs to restore MTRRs
- //
- ReleaseAllAPs ();
-
- //
- // Restore OS MTRRs
- //
- SmmCpuFeaturesReenableSmrr ();
- MtrrSetAllMtrrs(&Mtrrs);
-
- //
- // Wait for all APs to complete MTRR programming
- //
- WaitForAllAPs (ApCount);
- }
-
- //
- // Stop source level debug in BSP handler, the code below will not be
- // debugged.
- //
- InitializeDebugAgent (DEBUG_AGENT_INIT_EXIT_SMI, NULL, NULL);
-
- //
- // Signal APs to Reset states/semaphore for this processor
- //
- ReleaseAllAPs ();
-
- //
- // Perform pending operations for hot-plug
- //
- SmmCpuUpdate ();
-
- //
- // Clear the Present flag of BSP
- //
- *(mSmmMpSyncData->CpuData[CpuIndex].Present) = FALSE;
-
- //
- // Gather APs to exit SMM synchronously. Note the Present flag is cleared by now but
- // WaitForAllAps does not depend on the Present flag.
- //
- WaitForAllAPs (ApCount);
-
- //
- // Reset BspIndex to -1, meaning BSP has not been elected.
- //
- if (FeaturePcdGet (PcdCpuSmmEnableBspElection)) {
- mSmmMpSyncData->BspIndex = (UINT32)-1;
- }
-
- //
- // Allow APs to check in from this point on
- //
- *mSmmMpSyncData->Counter = 0;
- *mSmmMpSyncData->AllCpusInSync = FALSE;
-}
-
-/**
- SMI handler for AP.
-
- @param CpuIndex AP processor Index.
- @param ValidSmi Indicates that current SMI is a valid SMI or not.
- @param SyncMode SMM MP sync mode.
-
-**/
-VOID
-APHandler (
- IN UINTN CpuIndex,
- IN BOOLEAN ValidSmi,
- IN SMM_CPU_SYNC_MODE SyncMode
- )
-{
- UINT64 Timer;
- UINTN BspIndex;
- MTRR_SETTINGS Mtrrs;
-
- //
- // Timeout BSP
- //
- for (Timer = StartSyncTimer ();
- !IsSyncTimerTimeout (Timer) &&
- !(*mSmmMpSyncData->InsideSmm);
- ) {
- CpuPause ();
- }
-
- if (!(*mSmmMpSyncData->InsideSmm)) {
- //
- // BSP timeout in the first round
- //
- if (mSmmMpSyncData->BspIndex != -1) {
- //
- // BSP Index is known
- //
- BspIndex = mSmmMpSyncData->BspIndex;
- ASSERT (CpuIndex != BspIndex);
-
- //
- // Send SMI IPI to bring BSP in
- //
- SendSmiIpi ((UINT32)gSmmCpuPrivate->ProcessorInfo[BspIndex].ProcessorId);
-
- //
- // Now clock BSP for the 2nd time
- //
- for (Timer = StartSyncTimer ();
- !IsSyncTimerTimeout (Timer) &&
- !(*mSmmMpSyncData->InsideSmm);
- ) {
- CpuPause ();
- }
-
- if (!(*mSmmMpSyncData->InsideSmm)) {
- //
- // Give up since BSP is unable to enter SMM
- // and signal the completion of this AP
- WaitForSemaphore (mSmmMpSyncData->Counter);
- return;
- }
- } else {
- //
- // Don't know BSP index. Give up without sending IPI to BSP.
- //
- WaitForSemaphore (mSmmMpSyncData->Counter);
- return;
- }
- }
-
- //
- // BSP is available
- //
- BspIndex = mSmmMpSyncData->BspIndex;
- ASSERT (CpuIndex != BspIndex);
-
- //
- // Mark this processor's presence
- //
- *(mSmmMpSyncData->CpuData[CpuIndex].Present) = TRUE;
-
- if (SyncMode == SmmCpuSyncModeTradition || SmmCpuFeaturesNeedConfigureMtrrs()) {
- //
- // Notify BSP of arrival at this point
- //
- ReleaseSemaphore (mSmmMpSyncData->CpuData[BspIndex].Run);
- }
-
- if (SmmCpuFeaturesNeedConfigureMtrrs()) {
- //
- // Wait for the signal from BSP to backup MTRRs
- //
- WaitForSemaphore (mSmmMpSyncData->CpuData[CpuIndex].Run);
-
- //
- // Backup OS MTRRs
- //
- MtrrGetAllMtrrs(&Mtrrs);
-
- //
- // Signal BSP the completion of this AP
- //
- ReleaseSemaphore (mSmmMpSyncData->CpuData[BspIndex].Run);
-
- //
- // Wait for BSP's signal to program MTRRs
- //
- WaitForSemaphore (mSmmMpSyncData->CpuData[CpuIndex].Run);
-
- //
- // Replace OS MTRRs with SMI MTRRs
- //
- ReplaceOSMtrrs (CpuIndex);
-
- //
- // Signal BSP the completion of this AP
- //
- ReleaseSemaphore (mSmmMpSyncData->CpuData[BspIndex].Run);
- }
-
- while (TRUE) {
- //
- // Wait for something to happen
- //
- WaitForSemaphore (mSmmMpSyncData->CpuData[CpuIndex].Run);
-
- //
- // Check if BSP wants to exit SMM
- //
- if (!(*mSmmMpSyncData->InsideSmm)) {
- break;
- }
-
- //
- // BUSY should be acquired by SmmStartupThisAp()
- //
- ASSERT (
- !AcquireSpinLockOrFail (mSmmMpSyncData->CpuData[CpuIndex].Busy)
- );
-
- //
- // Invoke the scheduled procedure
- //
- (*mSmmMpSyncData->CpuData[CpuIndex].Procedure) (
- (VOID*)mSmmMpSyncData->CpuData[CpuIndex].Parameter
- );
-
- //
- // Release BUSY
- //
- ReleaseSpinLock (mSmmMpSyncData->CpuData[CpuIndex].Busy);
- }
-
- if (SmmCpuFeaturesNeedConfigureMtrrs()) {
- //
- // Notify BSP the readiness of this AP to program MTRRs
- //
- ReleaseSemaphore (mSmmMpSyncData->CpuData[BspIndex].Run);
-
- //
- // Wait for the signal from BSP to program MTRRs
- //
- WaitForSemaphore (mSmmMpSyncData->CpuData[CpuIndex].Run);
-
- //
- // Restore OS MTRRs
- //
- SmmCpuFeaturesReenableSmrr ();
- MtrrSetAllMtrrs(&Mtrrs);
- }
-
- //
- // Notify BSP the readiness of this AP to Reset states/semaphore for this processor
- //
- ReleaseSemaphore (mSmmMpSyncData->CpuData[BspIndex].Run);
-
- //
- // Wait for the signal from BSP to Reset states/semaphore for this processor
- //
- WaitForSemaphore (mSmmMpSyncData->CpuData[CpuIndex].Run);
-
- //
- // Reset states/semaphore for this processor
- //
- *(mSmmMpSyncData->CpuData[CpuIndex].Present) = FALSE;
-
- //
- // Notify BSP the readiness of this AP to exit SMM
- //
- ReleaseSemaphore (mSmmMpSyncData->CpuData[BspIndex].Run);
-
-}
-
-/**
- Create 4G PageTable in SMRAM.
-
- @param ExtraPages Additional page numbers besides for 4G memory
- @param Is32BitPageTable Whether the page table is 32-bit PAE
- @return PageTable Address
-
-**/
-UINT32
-Gen4GPageTable (
- IN UINTN ExtraPages,
- IN BOOLEAN Is32BitPageTable
- )
-{
- VOID *PageTable;
- UINTN Index;
- UINT64 *Pte;
- UINTN PagesNeeded;
- UINTN Low2MBoundary;
- UINTN High2MBoundary;
- UINTN Pages;
- UINTN GuardPage;
- UINT64 *Pdpte;
- UINTN PageIndex;
- UINTN PageAddress;
-
- Low2MBoundary = 0;
- High2MBoundary = 0;
- PagesNeeded = 0;
- if (FeaturePcdGet (PcdCpuSmmStackGuard)) {
- //
- // Add one more page for known good stack, then find the lower 2MB aligned address.
- //
- Low2MBoundary = (mSmmStackArrayBase + EFI_PAGE_SIZE) & ~(SIZE_2MB-1);
- //
- // Add two more pages for known good stack and stack guard page,
- // then find the lower 2MB aligned address.
- //
- High2MBoundary = (mSmmStackArrayEnd - mSmmStackSize + EFI_PAGE_SIZE * 2) & ~(SIZE_2MB-1);
- PagesNeeded = ((High2MBoundary - Low2MBoundary) / SIZE_2MB) + 1;
- }
- //
- // Allocate the page table
- //
- PageTable = AllocatePageTableMemory (ExtraPages + 5 + PagesNeeded);
- ASSERT (PageTable != NULL);
-
- PageTable = (VOID *)((UINTN)PageTable + EFI_PAGES_TO_SIZE (ExtraPages));
- Pte = (UINT64*)PageTable;
-
- //
- // Zero out all page table entries first
- //
- ZeroMem (Pte, EFI_PAGES_TO_SIZE (1));
-
- //
- // Set Page Directory Pointers
- //
- for (Index = 0; Index < 4; Index++) {
- Pte[Index] = (UINTN)PageTable + EFI_PAGE_SIZE * (Index + 1) + (Is32BitPageTable ? IA32_PAE_PDPTE_ATTRIBUTE_BITS : PAGE_ATTRIBUTE_BITS);
- }
- Pte += EFI_PAGE_SIZE / sizeof (*Pte);
-
- //
- // Fill in Page Directory Entries
- //
- for (Index = 0; Index < EFI_PAGE_SIZE * 4 / sizeof (*Pte); Index++) {
- Pte[Index] = (Index << 21) | IA32_PG_PS | PAGE_ATTRIBUTE_BITS;
- }
-
- if (FeaturePcdGet (PcdCpuSmmStackGuard)) {
- Pages = (UINTN)PageTable + EFI_PAGES_TO_SIZE (5);
- GuardPage = mSmmStackArrayBase + EFI_PAGE_SIZE;
- Pdpte = (UINT64*)PageTable;
- for (PageIndex = Low2MBoundary; PageIndex <= High2MBoundary; PageIndex += SIZE_2MB) {
- Pte = (UINT64*)(UINTN)(Pdpte[BitFieldRead32 ((UINT32)PageIndex, 30, 31)] & ~(EFI_PAGE_SIZE - 1));
- Pte[BitFieldRead32 ((UINT32)PageIndex, 21, 29)] = (UINT64)Pages | PAGE_ATTRIBUTE_BITS;
- //
- // Fill in Page Table Entries
- //
- Pte = (UINT64*)Pages;
- PageAddress = PageIndex;
- for (Index = 0; Index < EFI_PAGE_SIZE / sizeof (*Pte); Index++) {
- if (PageAddress == GuardPage) {
- //
- // Mark the guard page as non-present
- //
- Pte[Index] = PageAddress;
- GuardPage += mSmmStackSize;
- if (GuardPage > mSmmStackArrayEnd) {
- GuardPage = 0;
- }
- } else {
- Pte[Index] = PageAddress | PAGE_ATTRIBUTE_BITS;
- }
- PageAddress+= EFI_PAGE_SIZE;
- }
- Pages += EFI_PAGE_SIZE;
- }
- }
-
- return (UINT32)(UINTN)PageTable;
-}
-
-/**
- Set memory cache ability.
-
- @param PageTable PageTable Address
- @param Address Memory Address to change cache ability
- @param Cacheability Cache ability to set
-
-**/
-VOID
-SetCacheability (
- IN UINT64 *PageTable,
- IN UINTN Address,
- IN UINT8 Cacheability
- )
-{
- UINTN PTIndex;
- VOID *NewPageTableAddress;
- UINT64 *NewPageTable;
- UINTN Index;
-
- ASSERT ((Address & EFI_PAGE_MASK) == 0);
-
- if (sizeof (UINTN) == sizeof (UINT64)) {
- PTIndex = (UINTN)RShiftU64 (Address, 39) & 0x1ff;
- ASSERT (PageTable[PTIndex] & IA32_PG_P);
- PageTable = (UINT64*)(UINTN)(PageTable[PTIndex] & gPhyMask);
- }
-
- PTIndex = (UINTN)RShiftU64 (Address, 30) & 0x1ff;
- ASSERT (PageTable[PTIndex] & IA32_PG_P);
- PageTable = (UINT64*)(UINTN)(PageTable[PTIndex] & gPhyMask);
-
- //
- // A perfect implementation should check the original cacheability with the
- // one being set, and break a 2M page entry into pieces only when they
- // disagreed.
- //
- PTIndex = (UINTN)RShiftU64 (Address, 21) & 0x1ff;
- if ((PageTable[PTIndex] & IA32_PG_PS) != 0) {
- //
- // Allocate a page from SMRAM
- //
- NewPageTableAddress = AllocatePageTableMemory (1);
- ASSERT (NewPageTableAddress != NULL);
-
- NewPageTable = (UINT64 *)NewPageTableAddress;
-
- for (Index = 0; Index < 0x200; Index++) {
- NewPageTable[Index] = PageTable[PTIndex];
- if ((NewPageTable[Index] & IA32_PG_PAT_2M) != 0) {
- NewPageTable[Index] &= ~((UINT64)IA32_PG_PAT_2M);
- NewPageTable[Index] |= (UINT64)IA32_PG_PAT_4K;
- }
- NewPageTable[Index] |= (UINT64)(Index << EFI_PAGE_SHIFT);
- }
-
- PageTable[PTIndex] = ((UINTN)NewPageTableAddress & gPhyMask) | PAGE_ATTRIBUTE_BITS;
- }
-
- ASSERT (PageTable[PTIndex] & IA32_PG_P);
- PageTable = (UINT64*)(UINTN)(PageTable[PTIndex] & gPhyMask);
-
- PTIndex = (UINTN)RShiftU64 (Address, 12) & 0x1ff;
- ASSERT (PageTable[PTIndex] & IA32_PG_P);
- PageTable[PTIndex] &= ~((UINT64)((IA32_PG_PAT_4K | IA32_PG_CD | IA32_PG_WT)));
- PageTable[PTIndex] |= (UINT64)Cacheability;
-}
-
-
-/**
- Schedule a procedure to run on the specified CPU.
-
- @param Procedure The address of the procedure to run
- @param CpuIndex Target CPU Index
- @param ProcArguments The parameter to pass to the procedure
-
- @retval EFI_INVALID_PARAMETER CpuNumber not valid
- @retval EFI_INVALID_PARAMETER CpuNumber specifying BSP
- @retval EFI_INVALID_PARAMETER The AP specified by CpuNumber did not enter SMM
- @retval EFI_INVALID_PARAMETER The AP specified by CpuNumber is busy
- @retval EFI_SUCCESS The procedure has been successfully scheduled
-
-**/
-EFI_STATUS
-EFIAPI
-SmmStartupThisAp (
- IN EFI_AP_PROCEDURE Procedure,
- IN UINTN CpuIndex,
- IN OUT VOID *ProcArguments OPTIONAL
- )
-{
- if (CpuIndex >= gSmmCpuPrivate->SmmCoreEntryContext.NumberOfCpus ||
- CpuIndex == gSmmCpuPrivate->SmmCoreEntryContext.CurrentlyExecutingCpu ||
- !(*(mSmmMpSyncData->CpuData[CpuIndex].Present)) ||
- gSmmCpuPrivate->Operation[CpuIndex] == SmmCpuRemove ||
- !AcquireSpinLockOrFail (mSmmMpSyncData->CpuData[CpuIndex].Busy)) {
- return EFI_INVALID_PARAMETER;
- }
-
- mSmmMpSyncData->CpuData[CpuIndex].Procedure = Procedure;
- mSmmMpSyncData->CpuData[CpuIndex].Parameter = ProcArguments;
- ReleaseSemaphore (mSmmMpSyncData->CpuData[CpuIndex].Run);
-
- if (FeaturePcdGet (PcdCpuSmmBlockStartupThisAp)) {
- AcquireSpinLock (mSmmMpSyncData->CpuData[CpuIndex].Busy);
- ReleaseSpinLock (mSmmMpSyncData->CpuData[CpuIndex].Busy);
- }
- return EFI_SUCCESS;
-}
-
-/**
- This function sets DR6 & DR7 according to SMM save state, before running SMM C code.
- They are useful when you want to enable hardware breakpoints in SMM without entry SMM mode.
-
- NOTE: It might not be appreciated in runtime since it might
- conflict with OS debugging facilities. Turn them off in RELEASE.
-
- @param CpuIndex CPU Index
-
-**/
-VOID
-EFIAPI
-CpuSmmDebugEntry (
- IN UINTN CpuIndex
- )
-{
- SMRAM_SAVE_STATE_MAP *CpuSaveState;
-
- if (FeaturePcdGet (PcdCpuSmmDebug)) {
- CpuSaveState = (SMRAM_SAVE_STATE_MAP *)gSmmCpuPrivate->CpuSaveState[CpuIndex];
- if (mSmmSaveStateRegisterLma == EFI_SMM_SAVE_STATE_REGISTER_LMA_32BIT) {
- AsmWriteDr6 (CpuSaveState->x86._DR6);
- AsmWriteDr7 (CpuSaveState->x86._DR7);
- } else {
- AsmWriteDr6 ((UINTN)CpuSaveState->x64._DR6);
- AsmWriteDr7 ((UINTN)CpuSaveState->x64._DR7);
- }
- }
-}
-
-/**
- This function restores DR6 & DR7 to SMM save state.
-
- NOTE: It might not be appreciated in runtime since it might
- conflict with OS debugging facilities. Turn them off in RELEASE.
-
- @param CpuIndex CPU Index
-
-**/
-VOID
-EFIAPI
-CpuSmmDebugExit (
- IN UINTN CpuIndex
- )
-{
- SMRAM_SAVE_STATE_MAP *CpuSaveState;
-
- if (FeaturePcdGet (PcdCpuSmmDebug)) {
- CpuSaveState = (SMRAM_SAVE_STATE_MAP *)gSmmCpuPrivate->CpuSaveState[CpuIndex];
- if (mSmmSaveStateRegisterLma == EFI_SMM_SAVE_STATE_REGISTER_LMA_32BIT) {
- CpuSaveState->x86._DR7 = (UINT32)AsmReadDr7 ();
- CpuSaveState->x86._DR6 = (UINT32)AsmReadDr6 ();
- } else {
- CpuSaveState->x64._DR7 = AsmReadDr7 ();
- CpuSaveState->x64._DR6 = AsmReadDr6 ();
- }
- }
-}
-
-/**
- C function for SMI entry, each processor comes here upon SMI trigger.
-
- @param CpuIndex CPU Index
-
-**/
-VOID
-EFIAPI
-SmiRendezvous (
- IN UINTN CpuIndex
- )
-{
- EFI_STATUS Status;
- BOOLEAN ValidSmi;
- BOOLEAN IsBsp;
- BOOLEAN BspInProgress;
- UINTN Index;
- UINTN Cr2;
- BOOLEAN XdDisableFlag;
-
- //
- // Save Cr2 because Page Fault exception in SMM may override its value
- //
- Cr2 = AsmReadCr2 ();
-
- //
- // Perform CPU specific entry hooks
- //
- SmmCpuFeaturesRendezvousEntry (CpuIndex);
-
- //
- // Determine if this is a valid SMI
- //
- ValidSmi = PlatformValidSmi();
-
- //
- // Determine if BSP has been already in progress. Note this must be checked after
- // ValidSmi because BSP may clear a valid SMI source after checking in.
- //
- BspInProgress = *mSmmMpSyncData->InsideSmm;
-
- if (!BspInProgress && !ValidSmi) {
- //
- // If we reach here, it means when we sampled the ValidSmi flag, SMI status had not
- // been cleared by BSP in a new SMI run (so we have a truly invalid SMI), or SMI
- // status had been cleared by BSP and an existing SMI run has almost ended. (Note
- // we sampled ValidSmi flag BEFORE judging BSP-in-progress status.) In both cases, there
- // is nothing we need to do.
- //
- goto Exit;
- } else {
- //
- // Signal presence of this processor
- //
- if (ReleaseSemaphore (mSmmMpSyncData->Counter) == 0) {
- //
- // BSP has already ended the synchronization, so QUIT!!!
- //
-
- //
- // Wait for BSP's signal to finish SMI
- //
- while (*mSmmMpSyncData->AllCpusInSync) {
- CpuPause ();
- }
- goto Exit;
- } else {
-
- //
- // The BUSY lock is initialized to Released state.
- // This needs to be done early enough to be ready for BSP's SmmStartupThisAp() call.
- // E.g., with Relaxed AP flow, SmmStartupThisAp() may be called immediately
- // after AP's present flag is detected.
- //
- InitializeSpinLock (mSmmMpSyncData->CpuData[CpuIndex].Busy);
- }
-
- //
- // Try to enable XD
- //
- XdDisableFlag = FALSE;
- if (mXdSupported) {
- if ((AsmReadMsr64 (MSR_IA32_MISC_ENABLE) & B_XD_DISABLE_BIT) != 0) {
- XdDisableFlag = TRUE;
- AsmMsrAnd64 (MSR_IA32_MISC_ENABLE, ~B_XD_DISABLE_BIT);
- }
- ActivateXd ();
- }
-
- if (FeaturePcdGet (PcdCpuSmmProfileEnable)) {
- ActivateSmmProfile (CpuIndex);
- }
-
- if (BspInProgress) {
- //
- // BSP has been elected. Follow AP path, regardless of ValidSmi flag
- // as BSP may have cleared the SMI status
- //
- APHandler (CpuIndex, ValidSmi, mSmmMpSyncData->EffectiveSyncMode);
- } else {
- //
- // We have a valid SMI
- //
-
- //
- // Elect BSP
- //
- IsBsp = FALSE;
- if (FeaturePcdGet (PcdCpuSmmEnableBspElection)) {
- if (!mSmmMpSyncData->SwitchBsp || mSmmMpSyncData->CandidateBsp[CpuIndex]) {
- //
- // Call platform hook to do BSP election
- //
- Status = PlatformSmmBspElection (&IsBsp);
- if (EFI_SUCCESS == Status) {
- //
- // Platform hook determines successfully
- //
- if (IsBsp) {
- mSmmMpSyncData->BspIndex = (UINT32)CpuIndex;
- }
- } else {
- //
- // Platform hook fails to determine, use default BSP election method
- //
- InterlockedCompareExchange32 (
- (UINT32*)&mSmmMpSyncData->BspIndex,
- (UINT32)-1,
- (UINT32)CpuIndex
- );
- }
- }
- }
-
- //
- // "mSmmMpSyncData->BspIndex == CpuIndex" means this is the BSP
- //
- if (mSmmMpSyncData->BspIndex == CpuIndex) {
-
- //
- // Clear last request for SwitchBsp.
- //
- if (mSmmMpSyncData->SwitchBsp) {
- mSmmMpSyncData->SwitchBsp = FALSE;
- for (Index = 0; Index < mMaxNumberOfCpus; Index++) {
- mSmmMpSyncData->CandidateBsp[Index] = FALSE;
- }
- }
-
- if (FeaturePcdGet (PcdCpuSmmProfileEnable)) {
- SmmProfileRecordSmiNum ();
- }
-
- //
- // BSP Handler is always called with a ValidSmi == TRUE
- //
- BSPHandler (CpuIndex, mSmmMpSyncData->EffectiveSyncMode);
- } else {
- APHandler (CpuIndex, ValidSmi, mSmmMpSyncData->EffectiveSyncMode);
- }
- }
-
- ASSERT (*mSmmMpSyncData->CpuData[CpuIndex].Run == 0);
-
- //
- // Wait for BSP's signal to exit SMI
- //
- while (*mSmmMpSyncData->AllCpusInSync) {
- CpuPause ();
- }
-
- //
- // Restore XD
- //
- if (XdDisableFlag) {
- AsmMsrOr64 (MSR_IA32_MISC_ENABLE, B_XD_DISABLE_BIT);
- }
- }
-
-Exit:
- SmmCpuFeaturesRendezvousExit (CpuIndex);
- //
- // Restore Cr2
- //
- AsmWriteCr2 (Cr2);
-}
-
-/**
- Allocate buffer for all semaphores and spin locks.
-
-**/
-VOID
-InitializeSmmCpuSemaphores (
- VOID
- )
-{
- UINTN ProcessorCount;
- UINTN TotalSize;
- UINTN GlobalSemaphoresSize;
- UINTN CpuSemaphoresSize;
- UINTN MsrSemahporeSize;
- UINTN SemaphoreSize;
- UINTN Pages;
- UINTN *SemaphoreBlock;
- UINTN SemaphoreAddr;
-
- SemaphoreSize = GetSpinLockProperties ();
- ProcessorCount = gSmmCpuPrivate->SmmCoreEntryContext.NumberOfCpus;
- GlobalSemaphoresSize = (sizeof (SMM_CPU_SEMAPHORE_GLOBAL) / sizeof (VOID *)) * SemaphoreSize;
- CpuSemaphoresSize = (sizeof (SMM_CPU_SEMAPHORE_CPU) / sizeof (VOID *)) * ProcessorCount * SemaphoreSize;
- MsrSemahporeSize = MSR_SPIN_LOCK_INIT_NUM * SemaphoreSize;
- TotalSize = GlobalSemaphoresSize + CpuSemaphoresSize + MsrSemahporeSize;
- DEBUG((EFI_D_INFO, "One Semaphore Size = 0x%x\n", SemaphoreSize));
- DEBUG((EFI_D_INFO, "Total Semaphores Size = 0x%x\n", TotalSize));
- Pages = EFI_SIZE_TO_PAGES (TotalSize);
- SemaphoreBlock = AllocatePages (Pages);
- ASSERT (SemaphoreBlock != NULL);
- ZeroMem (SemaphoreBlock, TotalSize);
-
- SemaphoreAddr = (UINTN)SemaphoreBlock;
- mSmmCpuSemaphores.SemaphoreGlobal.Counter = (UINT32 *)SemaphoreAddr;
- SemaphoreAddr += SemaphoreSize;
- mSmmCpuSemaphores.SemaphoreGlobal.InsideSmm = (BOOLEAN *)SemaphoreAddr;
- SemaphoreAddr += SemaphoreSize;
- mSmmCpuSemaphores.SemaphoreGlobal.AllCpusInSync = (BOOLEAN *)SemaphoreAddr;
- SemaphoreAddr += SemaphoreSize;
- mSmmCpuSemaphores.SemaphoreGlobal.PFLock = (SPIN_LOCK *)SemaphoreAddr;
- SemaphoreAddr += SemaphoreSize;
- mSmmCpuSemaphores.SemaphoreGlobal.CodeAccessCheckLock
- = (SPIN_LOCK *)SemaphoreAddr;
- SemaphoreAddr = (UINTN)SemaphoreBlock + GlobalSemaphoresSize;
- mSmmCpuSemaphores.SemaphoreCpu.Busy = (SPIN_LOCK *)SemaphoreAddr;
- SemaphoreAddr += ProcessorCount * SemaphoreSize;
- mSmmCpuSemaphores.SemaphoreCpu.Run = (UINT32 *)SemaphoreAddr;
- SemaphoreAddr += ProcessorCount * SemaphoreSize;
- mSmmCpuSemaphores.SemaphoreCpu.Present = (BOOLEAN *)SemaphoreAddr;
-
- SemaphoreAddr = (UINTN)SemaphoreBlock + GlobalSemaphoresSize + CpuSemaphoresSize;
- mSmmCpuSemaphores.SemaphoreMsr.Msr = (SPIN_LOCK *)SemaphoreAddr;
- mSmmCpuSemaphores.SemaphoreMsr.AvailableCounter =
- ((UINTN)SemaphoreBlock + Pages * SIZE_4KB - SemaphoreAddr) / SemaphoreSize;
- ASSERT (mSmmCpuSemaphores.SemaphoreMsr.AvailableCounter >= MSR_SPIN_LOCK_INIT_NUM);
-
- mPFLock = mSmmCpuSemaphores.SemaphoreGlobal.PFLock;
- mConfigSmmCodeAccessCheckLock = mSmmCpuSemaphores.SemaphoreGlobal.CodeAccessCheckLock;
-
- mSemaphoreSize = SemaphoreSize;
-}
-
-/**
- Initialize un-cacheable data.
-
-**/
-VOID
-EFIAPI
-InitializeMpSyncData (
- VOID
- )
-{
- UINTN CpuIndex;
-
- if (mSmmMpSyncData != NULL) {
- //
- // mSmmMpSyncDataSize includes one structure of SMM_DISPATCHER_MP_SYNC_DATA, one
- // CpuData array of SMM_CPU_DATA_BLOCK and one CandidateBsp array of BOOLEAN.
- //
- ZeroMem (mSmmMpSyncData, mSmmMpSyncDataSize);
- mSmmMpSyncData->CpuData = (SMM_CPU_DATA_BLOCK *)((UINT8 *)mSmmMpSyncData + sizeof (SMM_DISPATCHER_MP_SYNC_DATA));
- mSmmMpSyncData->CandidateBsp = (BOOLEAN *)(mSmmMpSyncData->CpuData + gSmmCpuPrivate->SmmCoreEntryContext.NumberOfCpus);
- if (FeaturePcdGet (PcdCpuSmmEnableBspElection)) {
- //
- // Enable BSP election by setting BspIndex to -1
- //
- mSmmMpSyncData->BspIndex = (UINT32)-1;
- }
- mSmmMpSyncData->EffectiveSyncMode = (SMM_CPU_SYNC_MODE) PcdGet8 (PcdCpuSmmSyncMode);
-
- mSmmMpSyncData->Counter = mSmmCpuSemaphores.SemaphoreGlobal.Counter;
- mSmmMpSyncData->InsideSmm = mSmmCpuSemaphores.SemaphoreGlobal.InsideSmm;
- mSmmMpSyncData->AllCpusInSync = mSmmCpuSemaphores.SemaphoreGlobal.AllCpusInSync;
- ASSERT (mSmmMpSyncData->Counter != NULL && mSmmMpSyncData->InsideSmm != NULL &&
- mSmmMpSyncData->AllCpusInSync != NULL);
- *mSmmMpSyncData->Counter = 0;
- *mSmmMpSyncData->InsideSmm = FALSE;
- *mSmmMpSyncData->AllCpusInSync = FALSE;
-
- for (CpuIndex = 0; CpuIndex < gSmmCpuPrivate->SmmCoreEntryContext.NumberOfCpus; CpuIndex ++) {
- mSmmMpSyncData->CpuData[CpuIndex].Busy =
- (SPIN_LOCK *)((UINTN)mSmmCpuSemaphores.SemaphoreCpu.Busy + mSemaphoreSize * CpuIndex);
- mSmmMpSyncData->CpuData[CpuIndex].Run =
- (UINT32 *)((UINTN)mSmmCpuSemaphores.SemaphoreCpu.Run + mSemaphoreSize * CpuIndex);
- mSmmMpSyncData->CpuData[CpuIndex].Present =
- (BOOLEAN *)((UINTN)mSmmCpuSemaphores.SemaphoreCpu.Present + mSemaphoreSize * CpuIndex);
- }
- }
-}
-
-/**
- Initialize global data for MP synchronization.
-
- @param Stacks Base address of SMI stack buffer for all processors.
- @param StackSize Stack size for each processor in SMM.
-
-**/
-UINT32
-InitializeMpServiceData (
- IN VOID *Stacks,
- IN UINTN StackSize
- )
-{
- UINT32 Cr3;
- UINTN Index;
- MTRR_SETTINGS *Mtrr;
- PROCESSOR_SMM_DESCRIPTOR *Psd;
- UINT8 *GdtTssTables;
- UINTN GdtTableStepSize;
-
- //
- // Allocate memory for all locks and semaphores
- //
- InitializeSmmCpuSemaphores ();
-
- //
- // Initialize mSmmMpSyncData
- //
- mSmmMpSyncDataSize = sizeof (SMM_DISPATCHER_MP_SYNC_DATA) +
- (sizeof (SMM_CPU_DATA_BLOCK) + sizeof (BOOLEAN)) * gSmmCpuPrivate->SmmCoreEntryContext.NumberOfCpus;
- mSmmMpSyncData = (SMM_DISPATCHER_MP_SYNC_DATA*) AllocatePages (EFI_SIZE_TO_PAGES (mSmmMpSyncDataSize));
- ASSERT (mSmmMpSyncData != NULL);
- InitializeMpSyncData ();
-
- //
- // Initialize physical address mask
- // NOTE: Physical memory above virtual address limit is not supported !!!
- //
- AsmCpuid (0x80000008, (UINT32*)&Index, NULL, NULL, NULL);
- gPhyMask = LShiftU64 (1, (UINT8)Index) - 1;
- gPhyMask &= (1ull << 48) - EFI_PAGE_SIZE;
-
- //
- // Create page tables
- //
- Cr3 = SmmInitPageTable ();
-
- GdtTssTables = InitGdt (Cr3, &GdtTableStepSize);
-
- //
- // Initialize PROCESSOR_SMM_DESCRIPTOR for each CPU
- //
- for (Index = 0; Index < mMaxNumberOfCpus; Index++) {
- Psd = (PROCESSOR_SMM_DESCRIPTOR *)(VOID *)(UINTN)(mCpuHotPlugData.SmBase[Index] + SMM_PSD_OFFSET);
- CopyMem (Psd, &gcPsd, sizeof (gcPsd));
- Psd->SmmGdtPtr = (UINT64)(UINTN)(GdtTssTables + GdtTableStepSize * Index);
- Psd->SmmGdtSize = gcSmiGdtr.Limit + 1;
-
- //
- // Install SMI handler
- //
- InstallSmiHandler (
- Index,
- (UINT32)mCpuHotPlugData.SmBase[Index],
- (VOID*)((UINTN)Stacks + (StackSize * Index)),
- StackSize,
- (UINTN)Psd->SmmGdtPtr,
- Psd->SmmGdtSize,
- gcSmiIdtr.Base,
- gcSmiIdtr.Limit + 1,
- Cr3
- );
- }
-
- //
- // Record current MTRR settings
- //
- ZeroMem(gSmiMtrrs, sizeof (gSmiMtrrs));
- Mtrr = (MTRR_SETTINGS*)gSmiMtrrs;
- MtrrGetAllMtrrs (Mtrr);
-
- return Cr3;
-}
-
-/**
-
- Register the SMM Foundation entry point.
-
- @param This Pointer to EFI_SMM_CONFIGURATION_PROTOCOL instance
- @param SmmEntryPoint SMM Foundation EntryPoint
-
- @retval EFI_SUCCESS Successfully to register SMM foundation entry point
-
-**/
-EFI_STATUS
-EFIAPI
-RegisterSmmEntry (
- IN CONST EFI_SMM_CONFIGURATION_PROTOCOL *This,
- IN EFI_SMM_ENTRY_POINT SmmEntryPoint
- )
-{
- //
- // Record SMM Foundation EntryPoint, later invoke it on SMI entry vector.
- //
- gSmmCpuPrivate->SmmCoreEntry = SmmEntryPoint;
- return EFI_SUCCESS;
-}