## @file # This file is used to parse meta files # # Copyright (c) 2008 - 2015, Intel Corporation. All rights reserved.
# This program and the accompanying materials # are licensed and made available under the terms and conditions of the BSD License # which accompanies this distribution. The full text of the license may be found at # http://opensource.org/licenses/bsd-license.php # # THE PROGRAM IS DISTRIBUTED UNDER THE BSD LICENSE ON AN "AS IS" BASIS, # WITHOUT WARRANTIES OR REPRESENTATIONS OF ANY KIND, EITHER EXPRESS OR IMPLIED. # ## # Import Modules # import Common.LongFilePathOs as os import re import time import copy import Common.EdkLogger as EdkLogger import Common.GlobalData as GlobalData import EccGlobalData import EccToolError from CommonDataClass.DataClass import * from Common.DataType import * from Common.String import * from Common.Misc import GuidStructureStringToGuidString, CheckPcdDatum, PathClass, AnalyzePcdData from Common.Expression import * from CommonDataClass.Exceptions import * from MetaFileTable import MetaFileStorage from GenFds.FdfParser import FdfParser from Common.LongFilePathSupport import OpenLongFilePath as open from Common.LongFilePathSupport import CodecOpenLongFilePath ## A decorator used to parse macro definition def ParseMacro(Parser): def MacroParser(self): Match = gMacroDefPattern.match(self._CurrentLine) if not Match: # Not 'DEFINE/EDK_GLOBAL' statement, call decorated method Parser(self) return TokenList = GetSplitValueList(self._CurrentLine[Match.end(1):], TAB_EQUAL_SPLIT, 1) # Syntax check if not TokenList[0]: EdkLogger.error('Parser', FORMAT_INVALID, "No macro name given", ExtraData=self._CurrentLine, File=self.MetaFile, Line=self._LineIndex+1) if len(TokenList) < 2: TokenList.append('') Type = Match.group(1) Name, Value = TokenList # Global macros can be only defined via environment variable if Name in GlobalData.gGlobalDefines: EdkLogger.error('Parser', FORMAT_INVALID, "%s can only be defined via environment variable" % Name, ExtraData=self._CurrentLine, File=self.MetaFile, Line=self._LineIndex+1) # Only upper case letters, digit and '_' are allowed if not gMacroNamePattern.match(Name): EdkLogger.error('Parser', FORMAT_INVALID, "The macro name must be in the pattern [A-Z][A-Z0-9_]*", ExtraData=self._CurrentLine, File=self.MetaFile, Line=self._LineIndex+1) Value = ReplaceMacro(Value, self._Macros) self._ItemType = MODEL_META_DATA_DEFINE # DEFINE defined macros if Type == TAB_DSC_DEFINES_DEFINE: if type(self) == DecParser: if MODEL_META_DATA_HEADER in self._SectionType: self._FileLocalMacros[Name] = Value else: for Scope in self._Scope: self._SectionsMacroDict.setdefault((Scope[2], Scope[0], Scope[1]), {})[Name] = Value elif self._SectionType == MODEL_META_DATA_HEADER: self._FileLocalMacros[Name] = Value else: SectionDictKey = self._SectionType, self._Scope[0][0], self._Scope[0][1] if SectionDictKey not in self._SectionsMacroDict: self._SectionsMacroDict[SectionDictKey] = {} SectionLocalMacros = self._SectionsMacroDict[SectionDictKey] SectionLocalMacros[Name] = Value # EDK_GLOBAL defined macros elif type(self) != DscParser: EdkLogger.error('Parser', FORMAT_INVALID, "EDK_GLOBAL can only be used in .dsc file", ExtraData=self._CurrentLine, File=self.MetaFile, Line=self._LineIndex+1) elif self._SectionType != MODEL_META_DATA_HEADER: EdkLogger.error('Parser', FORMAT_INVALID, "EDK_GLOBAL can only be used under [Defines] section", ExtraData=self._CurrentLine, File=self.MetaFile, Line=self._LineIndex+1) elif (Name in self._FileLocalMacros) and (self._FileLocalMacros[Name] != Value): EdkLogger.error('Parser', FORMAT_INVALID, "EDK_GLOBAL defined a macro with the same name and different value as one defined by 'DEFINE'", ExtraData=self._CurrentLine, File=self.MetaFile, Line=self._LineIndex+1) self._ValueList = [Type, Name, Value] return MacroParser ## Base class of parser # # This class is used for derivation purpose. The specific parser for one kind # type file must derive this class and implement some public interfaces. # # @param FilePath The path of platform description file # @param FileType The raw data of DSC file # @param Table Database used to retrieve module/package information # @param Macros Macros used for replacement in file # @param Owner Owner ID (for sub-section parsing) # @param From ID from which the data comes (for !INCLUDE directive) # class MetaFileParser(object): # data type (file content) for specific file type DataType = {} # Parser objects used to implement singleton MetaFiles = {} ## Factory method # # One file, one parser object. This factory method makes sure that there's # only one object constructed for one meta file. # # @param Class class object of real AutoGen class # (InfParser, DecParser or DscParser) # @param FilePath The path of meta file # @param *args The specific class related parameters # @param **kwargs The specific class related dict parameters # def __new__(Class, FilePath, *args, **kwargs): if FilePath in Class.MetaFiles: return Class.MetaFiles[FilePath] else: ParserObject = super(MetaFileParser, Class).__new__(Class) Class.MetaFiles[FilePath] = ParserObject return ParserObject ## Constructor of MetaFileParser # # Initialize object of MetaFileParser # # @param FilePath The path of platform description file # @param FileType The raw data of DSC file # @param Table Database used to retrieve module/package information # @param Macros Macros used for replacement in file # @param Owner Owner ID (for sub-section parsing) # @param From ID from which the data comes (for !INCLUDE directive) # def __init__(self, FilePath, FileType, Table, Owner=-1, From=-1): self._Table = Table self._RawTable = Table self._FileType = FileType self.MetaFile = FilePath self._Defines = {} self._FileLocalMacros = {} self._SectionsMacroDict = {} # for recursive parsing self._Owner = [Owner] self._From = From # parsr status for parsing self._ValueList = ['', '', '', '', ''] self._Scope = [] self._LineIndex = 0 self._CurrentLine = '' self._SectionType = MODEL_UNKNOWN self._SectionName = '' self._InSubsection = False self._SubsectionType = MODEL_UNKNOWN self._SubsectionName = '' self._ItemType = MODEL_UNKNOWN self._LastItem = -1 self._Enabled = 0 self._Finished = False self._PostProcessed = False # Different version of meta-file has different way to parse. self._Version = 0 # UNI object and extra UNI object self._UniObj = None self._UniExtraObj = None ## Store the parsed data in table def _Store(self, *Args): return self._Table.Insert(*Args) ## Virtual method for starting parse def Start(self): raise NotImplementedError ## Notify a post-process is needed def DoPostProcess(self): self._PostProcessed = False ## Set parsing complete flag in both class and table def _Done(self): self._Finished = True ## Do not set end flag when processing included files if self._From == -1: self._Table.SetEndFlag() def _PostProcess(self): self._PostProcessed = True ## Get the parse complete flag def _GetFinished(self): return self._Finished ## Set the complete flag def _SetFinished(self, Value): self._Finished = Value ## Use [] style to query data in table, just for readability # # DataInfo = [data_type, scope1(arch), scope2(platform/moduletype)] # def __getitem__(self, DataInfo): if type(DataInfo) != type(()): DataInfo = (DataInfo,) # Parse the file first, if necessary if not self._Finished: if self._RawTable.IsIntegrity(): self._Finished = True else: self._Table = self._RawTable self._PostProcessed = False self.Start() # No specific ARCH or Platform given, use raw data if self._RawTable and (len(DataInfo) == 1 or DataInfo[1] == None): return self._RawTable.Query(*DataInfo) # Do post-process if necessary if not self._PostProcessed: self._PostProcess() return self._Table.Query(*DataInfo) ## Data parser for the common format in different type of file # # The common format in the meatfile is like # # xxx1 | xxx2 | xxx3 # @ParseMacro def _CommonParser(self): TokenList = GetSplitValueList(self._CurrentLine, TAB_VALUE_SPLIT) self._ValueList[0:len(TokenList)] = TokenList ## Data parser for the format in which there's path # # Only path can have macro used. So we need to replace them before use. # @ParseMacro def _PathParser(self): TokenList = GetSplitValueList(self._CurrentLine, TAB_VALUE_SPLIT) self._ValueList[0:len(TokenList)] = TokenList # Don't do macro replacement for dsc file at this point if type(self) != DscParser: Macros = self._Macros self._ValueList = [ReplaceMacro(Value, Macros) for Value in self._ValueList] ## Skip unsupported data def _Skip(self): if self._SectionName == TAB_USER_EXTENSIONS.upper() and self._CurrentLine.upper().endswith('.UNI'): if EccGlobalData.gConfig.UniCheckHelpInfo == '1' or EccGlobalData.gConfig.UniCheckAll == '1' or EccGlobalData.gConfig.CheckAll == '1': ExtraUni = self._CurrentLine.strip() ExtraUniFile = os.path.join(os.path.dirname(self.MetaFile), ExtraUni) IsModuleUni = self.MetaFile.upper().endswith('.INF') self._UniExtraObj = UniParser(ExtraUniFile, IsExtraUni=True, IsModuleUni=IsModuleUni) self._UniExtraObj.Start() else: EdkLogger.warn("Parser", "Unrecognized content", File=self.MetaFile, Line=self._LineIndex + 1, ExtraData=self._CurrentLine); self._ValueList[0:1] = [self._CurrentLine] ## Section header parser # # The section header is always in following format: # # [section_name.arch<.platform|module_type>] # def _SectionHeaderParser(self): self._Scope = [] self._SectionName = '' ArchList = set() for Item in GetSplitValueList(self._CurrentLine[1:-1], TAB_COMMA_SPLIT): if Item == '': continue ItemList = GetSplitValueList(Item, TAB_SPLIT) # different section should not mix in one section if self._SectionName != '' and self._SectionName != ItemList[0].upper(): EdkLogger.error('Parser', FORMAT_INVALID, "Different section names in the same section", File=self.MetaFile, Line=self._LineIndex+1, ExtraData=self._CurrentLine) self._SectionName = ItemList[0].upper() if self._SectionName in self.DataType: self._SectionType = self.DataType[self._SectionName] else: self._SectionType = MODEL_UNKNOWN EdkLogger.warn("Parser", "Unrecognized section", File=self.MetaFile, Line=self._LineIndex+1, ExtraData=self._CurrentLine) # S1 is always Arch if len(ItemList) > 1: S1 = ItemList[1].upper() else: S1 = 'COMMON' ArchList.add(S1) # S2 may be Platform or ModuleType if len(ItemList) > 2: S2 = ItemList[2].upper() else: S2 = 'COMMON' self._Scope.append([S1, S2]) # 'COMMON' must not be used with specific ARCHs at the same section if 'COMMON' in ArchList and len(ArchList) > 1: EdkLogger.error('Parser', FORMAT_INVALID, "'common' ARCH must not be used with specific ARCHs", File=self.MetaFile, Line=self._LineIndex+1, ExtraData=self._CurrentLine) # If the section information is needed later, it should be stored in database self._ValueList[0] = self._SectionName ## [defines] section parser @ParseMacro def _DefineParser(self): TokenList = GetSplitValueList(self._CurrentLine, TAB_EQUAL_SPLIT, 1) self._ValueList[1:len(TokenList)] = TokenList if not self._ValueList[1]: EdkLogger.error('Parser', FORMAT_INVALID, "No name specified", ExtraData=self._CurrentLine, File=self.MetaFile, Line=self._LineIndex+1) if not self._ValueList[2]: EdkLogger.error('Parser', FORMAT_INVALID, "No value specified", ExtraData=self._CurrentLine, File=self.MetaFile, Line=self._LineIndex+1) self._ValueList = [ReplaceMacro(Value, self._Macros) for Value in self._ValueList] Name, Value = self._ValueList[1], self._ValueList[2] # Sometimes, we need to make differences between EDK and EDK2 modules if Name == 'INF_VERSION': try: self._Version = int(Value, 0) except: EdkLogger.error('Parser', FORMAT_INVALID, "Invalid version number", ExtraData=self._CurrentLine, File=self.MetaFile, Line=self._LineIndex+1) elif Name == 'MODULE_UNI_FILE': UniFile = os.path.join(os.path.dirname(self.MetaFile), Value) if os.path.exists(UniFile): self._UniObj = UniParser(UniFile, IsExtraUni=False, IsModuleUni=True) self._UniObj.Start() else: EdkLogger.error('Parser', FILE_NOT_FOUND, "Module UNI file %s is missing." % Value, ExtraData=self._CurrentLine, File=self.MetaFile, Line=self._LineIndex+1, RaiseError=False) elif Name == 'PACKAGE_UNI_FILE': UniFile = os.path.join(os.path.dirname(self.MetaFile), Value) if os.path.exists(UniFile): self._UniObj = UniParser(UniFile, IsExtraUni=False, IsModuleUni=False) if type(self) == InfParser and self._Version < 0x00010005: # EDK module allows using defines as macros self._FileLocalMacros[Name] = Value self._Defines[Name] = Value ## [BuildOptions] section parser @ParseMacro def _BuildOptionParser(self): TokenList = GetSplitValueList(self._CurrentLine, TAB_EQUAL_SPLIT, 1) TokenList2 = GetSplitValueList(TokenList[0], ':', 1) if len(TokenList2) == 2: self._ValueList[0] = TokenList2[0] # toolchain family self._ValueList[1] = TokenList2[1] # keys else: self._ValueList[1] = TokenList[0] if len(TokenList) == 2 and type(self) != DscParser: # value self._ValueList[2] = ReplaceMacro(TokenList[1], self._Macros) if self._ValueList[1].count('_') != 4: EdkLogger.error( 'Parser', FORMAT_INVALID, "'%s' must be in format of ____FLAGS" % self._ValueList[1], ExtraData=self._CurrentLine, File=self.MetaFile, Line=self._LineIndex+1 ) def _GetMacros(self): Macros = {} Macros.update(self._FileLocalMacros) Macros.update(self._GetApplicableSectionMacro()) return Macros ## Get section Macros that are applicable to current line, which may come from other sections ## that share the same name while scope is wider def _GetApplicableSectionMacro(self): Macros = {} for Scope1, Scope2 in [("COMMON", "COMMON"), ("COMMON", self._Scope[0][1]), (self._Scope[0][0], "COMMON"), (self._Scope[0][0], self._Scope[0][1])]: if (self._SectionType, Scope1, Scope2) in self._SectionsMacroDict: Macros.update(self._SectionsMacroDict[(self._SectionType, Scope1, Scope2)]) return Macros _SectionParser = {} Finished = property(_GetFinished, _SetFinished) _Macros = property(_GetMacros) ## INF file parser class # # @param FilePath The path of platform description file # @param FileType The raw data of DSC file # @param Table Database used to retrieve module/package information # @param Macros Macros used for replacement in file # class InfParser(MetaFileParser): # INF file supported data types (one type per section) DataType = { TAB_UNKNOWN.upper() : MODEL_UNKNOWN, TAB_INF_DEFINES.upper() : MODEL_META_DATA_HEADER, TAB_DSC_DEFINES_DEFINE : MODEL_META_DATA_DEFINE, TAB_BUILD_OPTIONS.upper() : MODEL_META_DATA_BUILD_OPTION, TAB_INCLUDES.upper() : MODEL_EFI_INCLUDE, TAB_LIBRARIES.upper() : MODEL_EFI_LIBRARY_INSTANCE, TAB_LIBRARY_CLASSES.upper() : MODEL_EFI_LIBRARY_CLASS, TAB_PACKAGES.upper() : MODEL_META_DATA_PACKAGE, TAB_NMAKE.upper() : MODEL_META_DATA_NMAKE, TAB_INF_FIXED_PCD.upper() : MODEL_PCD_FIXED_AT_BUILD, TAB_INF_PATCH_PCD.upper() : MODEL_PCD_PATCHABLE_IN_MODULE, TAB_INF_FEATURE_PCD.upper() : MODEL_PCD_FEATURE_FLAG, TAB_INF_PCD_EX.upper() : MODEL_PCD_DYNAMIC_EX, TAB_INF_PCD.upper() : MODEL_PCD_DYNAMIC, TAB_SOURCES.upper() : MODEL_EFI_SOURCE_FILE, TAB_GUIDS.upper() : MODEL_EFI_GUID, TAB_PROTOCOLS.upper() : MODEL_EFI_PROTOCOL, TAB_PPIS.upper() : MODEL_EFI_PPI, TAB_DEPEX.upper() : MODEL_EFI_DEPEX, TAB_BINARIES.upper() : MODEL_EFI_BINARY_FILE, TAB_USER_EXTENSIONS.upper() : MODEL_META_DATA_USER_EXTENSION } ## Constructor of InfParser # # Initialize object of InfParser # # @param FilePath The path of module description file # @param FileType The raw data of DSC file # @param Table Database used to retrieve module/package information # @param Macros Macros used for replacement in file # def __init__(self, FilePath, FileType, Table): # prevent re-initialization if hasattr(self, "_Table"): return MetaFileParser.__init__(self, FilePath, FileType, Table) self.TblFile = EccGlobalData.gDb.TblFile self.FileID = -1 ## Parser starter def Start(self): NmakeLine = '' Content = '' Usage = '' try: Content = open(str(self.MetaFile), 'r').readlines() except: EdkLogger.error("Parser", FILE_READ_FAILURE, ExtraData=self.MetaFile) # # Insert a record for file # Filename = NormPath(self.MetaFile) FileID = self.TblFile.GetFileId(Filename) if FileID: self.FileID = FileID else: self.FileID = self.TblFile.InsertFile(Filename, MODEL_FILE_INF) # parse the file line by line IsFindBlockComment = False for Index in range(0, len(Content)): if self._SectionType in [MODEL_EFI_GUID, MODEL_EFI_PROTOCOL, MODEL_EFI_PPI, MODEL_PCD_FIXED_AT_BUILD, MODEL_PCD_PATCHABLE_IN_MODULE, MODEL_PCD_FEATURE_FLAG, MODEL_PCD_DYNAMIC_EX, MODEL_PCD_DYNAMIC]: Line = Content[Index].strip() if Line.startswith(TAB_SPECIAL_COMMENT): Usage += ' ' + Line[Line.find(TAB_SPECIAL_COMMENT):] continue elif Line.startswith(TAB_COMMENT_SPLIT): continue elif Line.find(TAB_COMMENT_SPLIT) > 0: Usage += ' ' + Line[Line.find(TAB_COMMENT_SPLIT):] Line = Line[:Line.find(TAB_COMMENT_SPLIT)] else: # skip empty, commented, block commented lines Line = CleanString(Content[Index], AllowCppStyleComment=True) Usage = '' NextLine = '' if Index + 1 < len(Content): NextLine = CleanString(Content[Index + 1]) if Line == '': continue if Line.find(DataType.TAB_COMMENT_EDK_START) > -1: IsFindBlockComment = True continue if Line.find(DataType.TAB_COMMENT_EDK_END) > -1: IsFindBlockComment = False continue if IsFindBlockComment: continue self._LineIndex = Index self._CurrentLine = Line # section header if Line[0] == TAB_SECTION_START and Line[-1] == TAB_SECTION_END: self._SectionHeaderParser() # Check invalid sections if self._Version < 0x00010005: if self._SectionType in [MODEL_META_DATA_BUILD_OPTION, MODEL_EFI_LIBRARY_CLASS, MODEL_META_DATA_PACKAGE, MODEL_PCD_FIXED_AT_BUILD, MODEL_PCD_PATCHABLE_IN_MODULE, MODEL_PCD_FEATURE_FLAG, MODEL_PCD_DYNAMIC_EX, MODEL_PCD_DYNAMIC, MODEL_EFI_GUID, MODEL_EFI_PROTOCOL, MODEL_EFI_PPI, MODEL_META_DATA_USER_EXTENSION]: EdkLogger.error('Parser', FORMAT_INVALID, "Section [%s] is not allowed in inf file without version" % (self._SectionName), ExtraData=self._CurrentLine, File=self.MetaFile, Line=self._LineIndex+1) elif self._SectionType in [MODEL_EFI_INCLUDE, MODEL_EFI_LIBRARY_INSTANCE, MODEL_META_DATA_NMAKE]: EdkLogger.error('Parser', FORMAT_INVALID, "Section [%s] is not allowed in inf file with version 0x%08x" % (self._SectionName, self._Version), ExtraData=self._CurrentLine, File=self.MetaFile, Line=self._LineIndex+1) continue # merge two lines specified by '\' in section NMAKE elif self._SectionType == MODEL_META_DATA_NMAKE: if Line[-1] == '\\': if NextLine == '': self._CurrentLine = NmakeLine + Line[0:-1] NmakeLine = '' else: if NextLine[0] == TAB_SECTION_START and NextLine[-1] == TAB_SECTION_END: self._CurrentLine = NmakeLine + Line[0:-1] NmakeLine = '' else: NmakeLine = NmakeLine + ' ' + Line[0:-1] continue else: self._CurrentLine = NmakeLine + Line NmakeLine = '' # section content self._ValueList = ['','',''] # parse current line, result will be put in self._ValueList self._SectionParser[self._SectionType](self) if self._ValueList == None or self._ItemType == MODEL_META_DATA_DEFINE: self._ItemType = -1 continue # # Model, Value1, Value2, Value3, Arch, Platform, BelongsToItem=-1, # LineBegin=-1, ColumnBegin=-1, LineEnd=-1, ColumnEnd=-1, Enabled=-1 # self._ValueList[0] = self._ValueList[0].replace('/', '\\') Usage = Usage.strip() for Arch, Platform in self._Scope: self._Store(self._SectionType, self._ValueList[0], self._ValueList[1], self._ValueList[2], Arch, Platform, self._Owner[-1], self.FileID, self._LineIndex+1, -1, self._LineIndex+1, -1, 0, Usage ) Usage = '' if IsFindBlockComment: EdkLogger.error("Parser", FORMAT_INVALID, "Open block comments (starting with /*) are expected to end with */", File=self.MetaFile) self._Done() ## Data parser for the format in which there's path # # Only path can have macro used. So we need to replace them before use. # def _IncludeParser(self): TokenList = GetSplitValueList(self._CurrentLine, TAB_VALUE_SPLIT) self._ValueList[0:len(TokenList)] = TokenList Macros = self._Macros if Macros: for Index in range(0, len(self._ValueList)): Value = self._ValueList[Index] if not Value: continue if Value.upper().find('$(EFI_SOURCE)\Edk'.upper()) > -1 or Value.upper().find('$(EFI_SOURCE)/Edk'.upper()) > -1: Value = '$(EDK_SOURCE)' + Value[17:] if Value.find('$(EFI_SOURCE)') > -1 or Value.find('$(EDK_SOURCE)') > -1: pass elif Value.startswith('.'): pass elif Value.startswith('$('): pass else: Value = '$(EFI_SOURCE)/' + Value self._ValueList[Index] = ReplaceMacro(Value, Macros) ## Parse [Sources] section # # Only path can have macro used. So we need to replace them before use. # @ParseMacro def _SourceFileParser(self): TokenList = GetSplitValueList(self._CurrentLine, TAB_VALUE_SPLIT) self._ValueList[0:len(TokenList)] = TokenList Macros = self._Macros # For Acpi tables, remove macro like ' TABLE_NAME=Sata1' if 'COMPONENT_TYPE' in Macros: if self._Defines['COMPONENT_TYPE'].upper() == 'ACPITABLE': self._ValueList[0] = GetSplitValueList(self._ValueList[0], ' ', 1)[0] if self._Defines['BASE_NAME'] == 'Microcode': pass self._ValueList = [ReplaceMacro(Value, Macros) for Value in self._ValueList] ## Parse [Binaries] section # # Only path can have macro used. So we need to replace them before use. # @ParseMacro def _BinaryFileParser(self): TokenList = GetSplitValueList(self._CurrentLine, TAB_VALUE_SPLIT, 2) if len(TokenList) < 2: EdkLogger.error('Parser', FORMAT_INVALID, "No file type or path specified", ExtraData=self._CurrentLine + " ( | [| ])", File=self.MetaFile, Line=self._LineIndex+1) if not TokenList[0]: EdkLogger.error('Parser', FORMAT_INVALID, "No file type specified", ExtraData=self._CurrentLine + " ( | [| ])", File=self.MetaFile, Line=self._LineIndex+1) if not TokenList[1]: EdkLogger.error('Parser', FORMAT_INVALID, "No file path specified", ExtraData=self._CurrentLine + " ( | [| ])", File=self.MetaFile, Line=self._LineIndex+1) self._ValueList[0:len(TokenList)] = TokenList self._ValueList[1] = ReplaceMacro(self._ValueList[1], self._Macros) ## [nmake] section parser (Edk.x style only) def _NmakeParser(self): TokenList = GetSplitValueList(self._CurrentLine, TAB_EQUAL_SPLIT, 1) self._ValueList[0:len(TokenList)] = TokenList # remove macros self._ValueList[1] = ReplaceMacro(self._ValueList[1], self._Macros) # remove self-reference in macro setting #self._ValueList[1] = ReplaceMacro(self._ValueList[1], {self._ValueList[0]:''}) ## [FixedPcd], [FeaturePcd], [PatchPcd], [Pcd] and [PcdEx] sections parser @ParseMacro def _PcdParser(self): TokenList = GetSplitValueList(self._CurrentLine, TAB_VALUE_SPLIT, 1) ValueList = GetSplitValueList(TokenList[0], TAB_SPLIT) if len(ValueList) != 2: EdkLogger.error('Parser', FORMAT_INVALID, "Illegal token space GUID and PCD name format", ExtraData=self._CurrentLine + " (.)", File=self.MetaFile, Line=self._LineIndex+1) self._ValueList[0:1] = ValueList if len(TokenList) > 1: self._ValueList[2] = TokenList[1] if self._ValueList[0] == '' or self._ValueList[1] == '': EdkLogger.error('Parser', FORMAT_INVALID, "No token space GUID or PCD name specified", ExtraData=self._CurrentLine + " (.)", File=self.MetaFile, Line=self._LineIndex+1) # if value are 'True', 'true', 'TRUE' or 'False', 'false', 'FALSE', replace with integer 1 or 0. if self._ValueList[2] != '': InfPcdValueList = GetSplitValueList(TokenList[1], TAB_VALUE_SPLIT, 1) if InfPcdValueList[0] in ['True', 'true', 'TRUE']: self._ValueList[2] = TokenList[1].replace(InfPcdValueList[0], '1', 1); elif InfPcdValueList[0] in ['False', 'false', 'FALSE']: self._ValueList[2] = TokenList[1].replace(InfPcdValueList[0], '0', 1); ## [depex] section parser @ParseMacro def _DepexParser(self): self._ValueList[0:1] = [self._CurrentLine] _SectionParser = { MODEL_UNKNOWN : MetaFileParser._Skip, MODEL_META_DATA_HEADER : MetaFileParser._DefineParser, MODEL_META_DATA_BUILD_OPTION : MetaFileParser._BuildOptionParser, MODEL_EFI_INCLUDE : _IncludeParser, # for Edk.x modules MODEL_EFI_LIBRARY_INSTANCE : MetaFileParser._CommonParser, # for Edk.x modules MODEL_EFI_LIBRARY_CLASS : MetaFileParser._PathParser, MODEL_META_DATA_PACKAGE : MetaFileParser._PathParser, MODEL_META_DATA_NMAKE : _NmakeParser, # for Edk.x modules MODEL_PCD_FIXED_AT_BUILD : _PcdParser, MODEL_PCD_PATCHABLE_IN_MODULE : _PcdParser, MODEL_PCD_FEATURE_FLAG : _PcdParser, MODEL_PCD_DYNAMIC_EX : _PcdParser, MODEL_PCD_DYNAMIC : _PcdParser, MODEL_EFI_SOURCE_FILE : _SourceFileParser, MODEL_EFI_GUID : MetaFileParser._CommonParser, MODEL_EFI_PROTOCOL : MetaFileParser._CommonParser, MODEL_EFI_PPI : MetaFileParser._CommonParser, MODEL_EFI_DEPEX : _DepexParser, MODEL_EFI_BINARY_FILE : _BinaryFileParser, MODEL_META_DATA_USER_EXTENSION : MetaFileParser._Skip, } ## DSC file parser class # # @param FilePath The path of platform description file # @param FileType The raw data of DSC file # @param Table Database used to retrieve module/package information # @param Macros Macros used for replacement in file # @param Owner Owner ID (for sub-section parsing) # @param From ID from which the data comes (for !INCLUDE directive) # class DscParser(MetaFileParser): # DSC file supported data types (one type per section) DataType = { TAB_SKUIDS.upper() : MODEL_EFI_SKU_ID, TAB_LIBRARIES.upper() : MODEL_EFI_LIBRARY_INSTANCE, TAB_LIBRARY_CLASSES.upper() : MODEL_EFI_LIBRARY_CLASS, TAB_BUILD_OPTIONS.upper() : MODEL_META_DATA_BUILD_OPTION, TAB_PCDS_FIXED_AT_BUILD_NULL.upper() : MODEL_PCD_FIXED_AT_BUILD, TAB_PCDS_PATCHABLE_IN_MODULE_NULL.upper() : MODEL_PCD_PATCHABLE_IN_MODULE, TAB_PCDS_FEATURE_FLAG_NULL.upper() : MODEL_PCD_FEATURE_FLAG, TAB_PCDS_DYNAMIC_DEFAULT_NULL.upper() : MODEL_PCD_DYNAMIC_DEFAULT, TAB_PCDS_DYNAMIC_HII_NULL.upper() : MODEL_PCD_DYNAMIC_HII, TAB_PCDS_DYNAMIC_VPD_NULL.upper() : MODEL_PCD_DYNAMIC_VPD, TAB_PCDS_DYNAMIC_EX_DEFAULT_NULL.upper() : MODEL_PCD_DYNAMIC_EX_DEFAULT, TAB_PCDS_DYNAMIC_EX_HII_NULL.upper() : MODEL_PCD_DYNAMIC_EX_HII, TAB_PCDS_DYNAMIC_EX_VPD_NULL.upper() : MODEL_PCD_DYNAMIC_EX_VPD, TAB_COMPONENTS.upper() : MODEL_META_DATA_COMPONENT, TAB_COMPONENTS_SOURCE_OVERRIDE_PATH.upper() : MODEL_META_DATA_COMPONENT_SOURCE_OVERRIDE_PATH, TAB_DSC_DEFINES.upper() : MODEL_META_DATA_HEADER, TAB_DSC_DEFINES_DEFINE : MODEL_META_DATA_DEFINE, TAB_DSC_DEFINES_EDKGLOBAL : MODEL_META_DATA_GLOBAL_DEFINE, TAB_INCLUDE.upper() : MODEL_META_DATA_INCLUDE, TAB_IF.upper() : MODEL_META_DATA_CONDITIONAL_STATEMENT_IF, TAB_IF_DEF.upper() : MODEL_META_DATA_CONDITIONAL_STATEMENT_IFDEF, TAB_IF_N_DEF.upper() : MODEL_META_DATA_CONDITIONAL_STATEMENT_IFNDEF, TAB_ELSE_IF.upper() : MODEL_META_DATA_CONDITIONAL_STATEMENT_ELSEIF, TAB_ELSE.upper() : MODEL_META_DATA_CONDITIONAL_STATEMENT_ELSE, TAB_END_IF.upper() : MODEL_META_DATA_CONDITIONAL_STATEMENT_ENDIF, TAB_ERROR.upper() : MODEL_META_DATA_CONDITIONAL_STATEMENT_ERROR, } # Valid names in define section DefineKeywords = [ "DSC_SPECIFICATION", "PLATFORM_NAME", "PLATFORM_GUID", "PLATFORM_VERSION", "SKUID_IDENTIFIER", "PCD_INFO_GENERATION", "SUPPORTED_ARCHITECTURES", "BUILD_TARGETS", "OUTPUT_DIRECTORY", "FLASH_DEFINITION", "BUILD_NUMBER", "RFC_LANGUAGES", "ISO_LANGUAGES", "TIME_STAMP_FILE", "VPD_TOOL_GUID", "FIX_LOAD_TOP_MEMORY_ADDRESS" ] SubSectionDefineKeywords = [ "FILE_GUID" ] SymbolPattern = ValueExpression.SymbolPattern ## Constructor of DscParser # # Initialize object of DscParser # # @param FilePath The path of platform description file # @param FileType The raw data of DSC file # @param Table Database used to retrieve module/package information # @param Macros Macros used for replacement in file # @param Owner Owner ID (for sub-section parsing) # @param From ID from which the data comes (for !INCLUDE directive) # def __init__(self, FilePath, FileType, Table, Owner=-1, From=-1): # prevent re-initialization if hasattr(self, "_Table"): return MetaFileParser.__init__(self, FilePath, FileType, Table, Owner, From) self._Version = 0x00010005 # Only EDK2 dsc file is supported # to store conditional directive evaluation result self._DirectiveStack = [] self._DirectiveEvalStack = [] self._Enabled = 1 # Final valid replacable symbols self._Symbols = {} # # Map the ID between the original table and new table to track # the owner item # self._IdMapping = {-1:-1} self.TblFile = EccGlobalData.gDb.TblFile self.FileID = -1 ## Parser starter def Start(self): Content = '' try: Content = open(str(self.MetaFile.Path), 'r').readlines() except: EdkLogger.error("Parser", FILE_READ_FAILURE, ExtraData=self.MetaFile) # # Insert a record for file # Filename = NormPath(self.MetaFile.Path) FileID = self.TblFile.GetFileId(Filename) if FileID: self.FileID = FileID else: self.FileID = self.TblFile.InsertFile(Filename, MODEL_FILE_DSC) for Index in range(0, len(Content)): Line = CleanString(Content[Index]) # skip empty line if Line == '': continue self._CurrentLine = Line self._LineIndex = Index if self._InSubsection and self._Owner[-1] == -1: self._Owner.append(self._LastItem) # section header if Line[0] == TAB_SECTION_START and Line[-1] == TAB_SECTION_END: self._SectionType = MODEL_META_DATA_SECTION_HEADER # subsection ending elif Line[0] == '}' and self._InSubsection: self._InSubsection = False self._SubsectionType = MODEL_UNKNOWN self._SubsectionName = '' self._Owner[-1] = -1 continue # subsection header elif Line[0] == TAB_OPTION_START and Line[-1] == TAB_OPTION_END: self._SubsectionType = MODEL_META_DATA_SUBSECTION_HEADER # directive line elif Line[0] == '!': self._DirectiveParser() continue if self._InSubsection: SectionType = self._SubsectionType else: SectionType = self._SectionType self._ItemType = SectionType self._ValueList = ['', '', ''] self._SectionParser[SectionType](self) if self._ValueList == None: continue # # Model, Value1, Value2, Value3, Arch, ModuleType, BelongsToItem=-1, BelongsToFile=-1, # LineBegin=-1, ColumnBegin=-1, LineEnd=-1, ColumnEnd=-1, Enabled=-1 # for Arch, ModuleType in self._Scope: self._LastItem = self._Store( self._ItemType, self._ValueList[0], self._ValueList[1], self._ValueList[2], Arch, ModuleType, self._Owner[-1], self.FileID, self._From, self._LineIndex+1, -1, self._LineIndex+1, -1, self._Enabled ) if self._DirectiveStack: Type, Line, Text = self._DirectiveStack[-1] EdkLogger.error('Parser', FORMAT_INVALID, "No matching '!endif' found", ExtraData=Text, File=self.MetaFile, Line=Line) self._Done() ## parser def _SubsectionHeaderParser(self): self._SubsectionName = self._CurrentLine[1:-1].upper() if self._SubsectionName in self.DataType: self._SubsectionType = self.DataType[self._SubsectionName] else: self._SubsectionType = MODEL_UNKNOWN EdkLogger.warn("Parser", "Unrecognized sub-section", File=self.MetaFile, Line=self._LineIndex+1, ExtraData=self._CurrentLine) self._ValueList[0] = self._SubsectionName ## Directive statement parser def _DirectiveParser(self): self._ValueList = ['','',''] TokenList = GetSplitValueList(self._CurrentLine, ' ', 1) self._ValueList[0:len(TokenList)] = TokenList # Syntax check DirectiveName = self._ValueList[0].upper() if DirectiveName not in self.DataType: EdkLogger.error("Parser", FORMAT_INVALID, "Unknown directive [%s]" % DirectiveName, File=self.MetaFile, Line=self._LineIndex+1) if DirectiveName in ['!IF', '!IFDEF', '!INCLUDE', '!IFNDEF', '!ELSEIF'] and self._ValueList[1] == '': EdkLogger.error("Parser", FORMAT_INVALID, "Missing expression", File=self.MetaFile, Line=self._LineIndex+1, ExtraData=self._CurrentLine) ItemType = self.DataType[DirectiveName] if ItemType == MODEL_META_DATA_CONDITIONAL_STATEMENT_ENDIF: # Remove all directives between !if and !endif, including themselves while self._DirectiveStack: # Remove any !else or !elseif DirectiveInfo = self._DirectiveStack.pop() if DirectiveInfo[0] in [MODEL_META_DATA_CONDITIONAL_STATEMENT_IF, MODEL_META_DATA_CONDITIONAL_STATEMENT_IFDEF, MODEL_META_DATA_CONDITIONAL_STATEMENT_IFNDEF]: break else: EdkLogger.error("Parser", FORMAT_INVALID, "Redundant '!endif'", File=self.MetaFile, Line=self._LineIndex+1, ExtraData=self._CurrentLine) elif ItemType != MODEL_META_DATA_INCLUDE: # Break if there's a !else is followed by a !elseif if ItemType == MODEL_META_DATA_CONDITIONAL_STATEMENT_ELSEIF and \ self._DirectiveStack and \ self._DirectiveStack[-1][0] == MODEL_META_DATA_CONDITIONAL_STATEMENT_ELSE: EdkLogger.error("Parser", FORMAT_INVALID, "'!elseif' after '!else'", File=self.MetaFile, Line=self._LineIndex+1, ExtraData=self._CurrentLine) self._DirectiveStack.append((ItemType, self._LineIndex+1, self._CurrentLine)) elif self._From > 0: EdkLogger.error('Parser', FORMAT_INVALID, "No '!include' allowed in included file", ExtraData=self._CurrentLine, File=self.MetaFile, Line=self._LineIndex+1) # # Model, Value1, Value2, Value3, Arch, ModuleType, BelongsToItem=-1, BelongsToFile=-1, # LineBegin=-1, ColumnBegin=-1, LineEnd=-1, ColumnEnd=-1, Enabled=-1 # self._LastItem = self._Store( ItemType, self._ValueList[0], self._ValueList[1], self._ValueList[2], 'COMMON', 'COMMON', self._Owner[-1], self.FileID, self._From, self._LineIndex+1, -1, self._LineIndex+1, -1, 0 ) ## [defines] section parser @ParseMacro def _DefineParser(self): TokenList = GetSplitValueList(self._CurrentLine, TAB_EQUAL_SPLIT, 1) self._ValueList[1:len(TokenList)] = TokenList # Syntax check if not self._ValueList[1]: EdkLogger.error('Parser', FORMAT_INVALID, "No name specified", ExtraData=self._CurrentLine, File=self.MetaFile, Line=self._LineIndex+1) if not self._ValueList[2]: EdkLogger.error('Parser', FORMAT_INVALID, "No value specified", ExtraData=self._CurrentLine, File=self.MetaFile, Line=self._LineIndex+1) if (not self._ValueList[1] in self.DefineKeywords and (self._InSubsection and self._ValueList[1] not in self.SubSectionDefineKeywords)): EdkLogger.error('Parser', FORMAT_INVALID, "Unknown keyword found: %s. " "If this is a macro you must " "add it as a DEFINE in the DSC" % self._ValueList[1], ExtraData=self._CurrentLine, File=self.MetaFile, Line=self._LineIndex+1) self._Defines[self._ValueList[1]] = self._ValueList[2] self._ItemType = self.DataType[TAB_DSC_DEFINES.upper()] @ParseMacro def _SkuIdParser(self): TokenList = GetSplitValueList(self._CurrentLine, TAB_VALUE_SPLIT) if len(TokenList) != 2: EdkLogger.error('Parser', FORMAT_INVALID, "Correct format is '|'", ExtraData=self._CurrentLine, File=self.MetaFile, Line=self._LineIndex+1) self._ValueList[0:len(TokenList)] = TokenList ## Parse Edk style of library modules def _LibraryInstanceParser(self): self._ValueList[0] = self._CurrentLine ## PCD sections parser # # [PcdsFixedAtBuild] # [PcdsPatchableInModule] # [PcdsFeatureFlag] # [PcdsDynamicEx # [PcdsDynamicExDefault] # [PcdsDynamicExVpd] # [PcdsDynamicExHii] # [PcdsDynamic] # [PcdsDynamicDefault] # [PcdsDynamicVpd] # [PcdsDynamicHii] # @ParseMacro def _PcdParser(self): TokenList = GetSplitValueList(self._CurrentLine, TAB_VALUE_SPLIT, 1) self._ValueList[0:1] = GetSplitValueList(TokenList[0], TAB_SPLIT) if len(TokenList) == 2: self._ValueList[2] = TokenList[1] if self._ValueList[0] == '' or self._ValueList[1] == '': EdkLogger.error('Parser', FORMAT_INVALID, "No token space GUID or PCD name specified", ExtraData=self._CurrentLine + " (.|)", File=self.MetaFile, Line=self._LineIndex+1) if self._ValueList[2] == '': EdkLogger.error('Parser', FORMAT_INVALID, "No PCD value given", ExtraData=self._CurrentLine + " (.|)", File=self.MetaFile, Line=self._LineIndex+1) # if value are 'True', 'true', 'TRUE' or 'False', 'false', 'FALSE', replace with integer 1 or 0. DscPcdValueList = GetSplitValueList(TokenList[1], TAB_VALUE_SPLIT, 1) if DscPcdValueList[0] in ['True', 'true', 'TRUE']: self._ValueList[2] = TokenList[1].replace(DscPcdValueList[0], '1', 1); elif DscPcdValueList[0] in ['False', 'false', 'FALSE']: self._ValueList[2] = TokenList[1].replace(DscPcdValueList[0], '0', 1); ## [components] section parser @ParseMacro def _ComponentParser(self): if self._CurrentLine[-1] == '{': self._ValueList[0] = self._CurrentLine[0:-1].strip() self._InSubsection = True else: self._ValueList[0] = self._CurrentLine ## [LibraryClasses] section @ParseMacro def _LibraryClassParser(self): TokenList = GetSplitValueList(self._CurrentLine, TAB_VALUE_SPLIT) if len(TokenList) < 2: EdkLogger.error('Parser', FORMAT_INVALID, "No library class or instance specified", ExtraData=self._CurrentLine + " (|)", File=self.MetaFile, Line=self._LineIndex+1) if TokenList[0] == '': EdkLogger.error('Parser', FORMAT_INVALID, "No library class specified", ExtraData=self._CurrentLine + " (|)", File=self.MetaFile, Line=self._LineIndex+1) if TokenList[1] == '': EdkLogger.error('Parser', FORMAT_INVALID, "No library instance specified", ExtraData=self._CurrentLine + " (|)", File=self.MetaFile, Line=self._LineIndex+1) self._ValueList[0:len(TokenList)] = TokenList def _CompponentSourceOverridePathParser(self): self._ValueList[0] = self._CurrentLine ## [BuildOptions] section parser @ParseMacro def _BuildOptionParser(self): TokenList = GetSplitValueList(self._CurrentLine, TAB_EQUAL_SPLIT, 1) TokenList2 = GetSplitValueList(TokenList[0], ':', 1) if len(TokenList2) == 2: self._ValueList[0] = TokenList2[0] # toolchain family self._ValueList[1] = TokenList2[1] # keys else: self._ValueList[1] = TokenList[0] if len(TokenList) == 2: # value self._ValueList[2] = TokenList[1] if self._ValueList[1].count('_') != 4: EdkLogger.error( 'Parser', FORMAT_INVALID, "'%s' must be in format of ____FLAGS" % self._ValueList[1], ExtraData=self._CurrentLine, File=self.MetaFile, Line=self._LineIndex+1 ) ## Override parent's method since we'll do all macro replacements in parser def _GetMacros(self): Macros = dict( [('ARCH','IA32'), ('FAMILY','MSFT'),('TOOL_CHAIN_TAG','VS2008x86'),('TARGET','DEBUG')]) Macros.update(self._FileLocalMacros) Macros.update(self._GetApplicableSectionMacro()) Macros.update(GlobalData.gEdkGlobal) Macros.update(GlobalData.gPlatformDefines) Macros.update(GlobalData.gCommandLineDefines) # PCD cannot be referenced in macro definition if self._ItemType not in [MODEL_META_DATA_DEFINE, MODEL_META_DATA_GLOBAL_DEFINE]: Macros.update(self._Symbols) return Macros def _PostProcess(self): Processer = { MODEL_META_DATA_SECTION_HEADER : self.__ProcessSectionHeader, MODEL_META_DATA_SUBSECTION_HEADER : self.__ProcessSubsectionHeader, MODEL_META_DATA_HEADER : self.__ProcessDefine, MODEL_META_DATA_DEFINE : self.__ProcessDefine, MODEL_META_DATA_GLOBAL_DEFINE : self.__ProcessDefine, MODEL_META_DATA_INCLUDE : self.__ProcessDirective, MODEL_META_DATA_CONDITIONAL_STATEMENT_IF : self.__ProcessDirective, MODEL_META_DATA_CONDITIONAL_STATEMENT_ELSE : self.__ProcessDirective, MODEL_META_DATA_CONDITIONAL_STATEMENT_IFDEF : self.__ProcessDirective, MODEL_META_DATA_CONDITIONAL_STATEMENT_IFNDEF : self.__ProcessDirective, MODEL_META_DATA_CONDITIONAL_STATEMENT_ENDIF : self.__ProcessDirective, MODEL_META_DATA_CONDITIONAL_STATEMENT_ELSEIF : self.__ProcessDirective, MODEL_EFI_SKU_ID : self.__ProcessSkuId, MODEL_EFI_LIBRARY_INSTANCE : self.__ProcessLibraryInstance, MODEL_EFI_LIBRARY_CLASS : self.__ProcessLibraryClass, MODEL_PCD_FIXED_AT_BUILD : self.__ProcessPcd, MODEL_PCD_PATCHABLE_IN_MODULE : self.__ProcessPcd, MODEL_PCD_FEATURE_FLAG : self.__ProcessPcd, MODEL_PCD_DYNAMIC_DEFAULT : self.__ProcessPcd, MODEL_PCD_DYNAMIC_HII : self.__ProcessPcd, MODEL_PCD_DYNAMIC_VPD : self.__ProcessPcd, MODEL_PCD_DYNAMIC_EX_DEFAULT : self.__ProcessPcd, MODEL_PCD_DYNAMIC_EX_HII : self.__ProcessPcd, MODEL_PCD_DYNAMIC_EX_VPD : self.__ProcessPcd, MODEL_META_DATA_COMPONENT : self.__ProcessComponent, MODEL_META_DATA_COMPONENT_SOURCE_OVERRIDE_PATH : self.__ProcessSourceOverridePath, MODEL_META_DATA_BUILD_OPTION : self.__ProcessBuildOption, MODEL_UNKNOWN : self._Skip, MODEL_META_DATA_USER_EXTENSION : self._Skip, MODEL_META_DATA_CONDITIONAL_STATEMENT_ERROR : self._Skip, } self._RawTable = self._Table self._Table = MetaFileStorage(self._RawTable.Cur, self.MetaFile, MODEL_FILE_DSC, True) self._DirectiveStack = [] self._DirectiveEvalStack = [] self._FileWithError = self.MetaFile self._FileLocalMacros = {} self._SectionsMacroDict = {} GlobalData.gPlatformDefines = {} # Get all macro and PCD which has straitforward value self.__RetrievePcdValue() self._Content = self._RawTable.GetAll() self._ContentIndex = 0 while self._ContentIndex < len(self._Content) : Id, self._ItemType, V1, V2, V3, S1, S2, Owner, BelongsToFile, self._From, \ LineStart, ColStart, LineEnd, ColEnd, Enabled = self._Content[self._ContentIndex] if self._From < 0: self._FileWithError = self.MetaFile self._ContentIndex += 1 self._Scope = [[S1, S2]] self._LineIndex = LineStart - 1 self._ValueList = [V1, V2, V3] try: Processer[self._ItemType]() except EvaluationException, Excpt: # # Only catch expression evaluation error here. We need to report # the precise number of line on which the error occurred # pass # EdkLogger.error('Parser', FORMAT_INVALID, "Invalid expression: %s" % str(Excpt), # File=self._FileWithError, ExtraData=' '.join(self._ValueList), # Line=self._LineIndex+1) except MacroException, Excpt: EdkLogger.error('Parser', FORMAT_INVALID, str(Excpt), File=self._FileWithError, ExtraData=' '.join(self._ValueList), Line=self._LineIndex+1) if self._ValueList == None: continue NewOwner = self._IdMapping.get(Owner, -1) self._Enabled = int((not self._DirectiveEvalStack) or (False not in self._DirectiveEvalStack)) self._LastItem = self._Store( self._ItemType, self._ValueList[0], self._ValueList[1], self._ValueList[2], S1, S2, NewOwner, BelongsToFile, self._From, self._LineIndex+1, -1, self._LineIndex+1, -1, self._Enabled ) self._IdMapping[Id] = self._LastItem RecordList = self._Table.GetAll() self._RawTable.Drop() self._Table.Drop() for Record in RecordList: EccGlobalData.gDb.TblDsc.Insert(Record[1],Record[2],Record[3],Record[4],Record[5],Record[6],Record[7],Record[8],Record[9],Record[10],Record[11],Record[12],Record[13],Record[14]) GlobalData.gPlatformDefines.update(self._FileLocalMacros) self._PostProcessed = True self._Content = None def __ProcessSectionHeader(self): self._SectionName = self._ValueList[0] if self._SectionName in self.DataType: self._SectionType = self.DataType[self._SectionName] else: self._SectionType = MODEL_UNKNOWN def __ProcessSubsectionHeader(self): self._SubsectionName = self._ValueList[0] if self._SubsectionName in self.DataType: self._SubsectionType = self.DataType[self._SubsectionName] else: self._SubsectionType = MODEL_UNKNOWN def __RetrievePcdValue(self): Records = self._RawTable.Query(MODEL_PCD_FEATURE_FLAG, BelongsToItem=-1.0) for TokenSpaceGuid,PcdName,Value,Dummy2,Dummy3,ID,Line in Records: Value, DatumType, MaxDatumSize = AnalyzePcdData(Value) # Only use PCD whose value is straitforward (no macro and PCD) if self.SymbolPattern.findall(Value): continue Name = TokenSpaceGuid + '.' + PcdName # Don't use PCD with different values. if Name in self._Symbols and self._Symbols[Name] != Value: self._Symbols.pop(Name) continue self._Symbols[Name] = Value Records = self._RawTable.Query(MODEL_PCD_FIXED_AT_BUILD, BelongsToItem=-1.0) for TokenSpaceGuid,PcdName,Value,Dummy2,Dummy3,ID,Line in Records: Value, DatumType, MaxDatumSize = AnalyzePcdData(Value) # Only use PCD whose value is straitforward (no macro and PCD) if self.SymbolPattern.findall(Value): continue Name = TokenSpaceGuid+'.'+PcdName # Don't use PCD with different values. if Name in self._Symbols and self._Symbols[Name] != Value: self._Symbols.pop(Name) continue self._Symbols[Name] = Value def __ProcessDefine(self): if not self._Enabled: return Type, Name, Value = self._ValueList Value = ReplaceMacro(Value, self._Macros, False) if self._ItemType == MODEL_META_DATA_DEFINE: if self._SectionType == MODEL_META_DATA_HEADER: self._FileLocalMacros[Name] = Value else: SectionDictKey = self._SectionType, self._Scope[0][0], self._Scope[0][1] if SectionDictKey not in self._SectionsMacroDict: self._SectionsMacroDict[SectionDictKey] = {} SectionLocalMacros = self._SectionsMacroDict[SectionDictKey] SectionLocalMacros[Name] = Value elif self._ItemType == MODEL_META_DATA_GLOBAL_DEFINE: GlobalData.gEdkGlobal[Name] = Value # # Keyword in [Defines] section can be used as Macros # if (self._ItemType == MODEL_META_DATA_HEADER) and (self._SectionType == MODEL_META_DATA_HEADER): self._FileLocalMacros[Name] = Value self._ValueList = [Type, Name, Value] def __ProcessDirective(self): Result = None if self._ItemType in [MODEL_META_DATA_CONDITIONAL_STATEMENT_IF, MODEL_META_DATA_CONDITIONAL_STATEMENT_ELSEIF]: Macros = self._Macros Macros.update(GlobalData.gGlobalDefines) try: Result = ValueExpression(self._ValueList[1], Macros)() except SymbolNotFound, Exc: EdkLogger.debug(EdkLogger.DEBUG_5, str(Exc), self._ValueList[1]) Result = False except WrnExpression, Excpt: # # Catch expression evaluation warning here. We need to report # the precise number of line and return the evaluation result # EdkLogger.warn('Parser', "Suspicious expression: %s" % str(Excpt), File=self._FileWithError, ExtraData=' '.join(self._ValueList), Line=self._LineIndex+1) Result = Excpt.result except BadExpression, Exc: EdkLogger.debug(EdkLogger.DEBUG_5, str(Exc), self._ValueList[1]) Result = False if self._ItemType in [MODEL_META_DATA_CONDITIONAL_STATEMENT_IF, MODEL_META_DATA_CONDITIONAL_STATEMENT_IFDEF, MODEL_META_DATA_CONDITIONAL_STATEMENT_IFNDEF]: self._DirectiveStack.append(self._ItemType) if self._ItemType == MODEL_META_DATA_CONDITIONAL_STATEMENT_IF: Result = bool(Result) else: Macro = self._ValueList[1] Macro = Macro[2:-1] if (Macro.startswith("$(") and Macro.endswith(")")) else Macro Result = Macro in self._Macros if self._ItemType == MODEL_META_DATA_CONDITIONAL_STATEMENT_IFNDEF: Result = not Result self._DirectiveEvalStack.append(Result) elif self._ItemType == MODEL_META_DATA_CONDITIONAL_STATEMENT_ELSEIF: self._DirectiveStack.append(self._ItemType) self._DirectiveEvalStack[-1] = not self._DirectiveEvalStack[-1] self._DirectiveEvalStack.append(bool(Result)) elif self._ItemType == MODEL_META_DATA_CONDITIONAL_STATEMENT_ELSE: self._DirectiveStack[-1] = self._ItemType self._DirectiveEvalStack[-1] = not self._DirectiveEvalStack[-1] elif self._ItemType == MODEL_META_DATA_CONDITIONAL_STATEMENT_ENDIF: # Back to the nearest !if/!ifdef/!ifndef while self._DirectiveStack: self._DirectiveEvalStack.pop() Directive = self._DirectiveStack.pop() if Directive in [MODEL_META_DATA_CONDITIONAL_STATEMENT_IF, MODEL_META_DATA_CONDITIONAL_STATEMENT_IFDEF, MODEL_META_DATA_CONDITIONAL_STATEMENT_ELSE, MODEL_META_DATA_CONDITIONAL_STATEMENT_IFNDEF]: break elif self._ItemType == MODEL_META_DATA_INCLUDE: # The included file must be relative to workspace or same directory as DSC file __IncludeMacros = {} # # Allow using system environment variables in path after !include # __IncludeMacros['WORKSPACE'] = GlobalData.gGlobalDefines['WORKSPACE'] if "ECP_SOURCE" in GlobalData.gGlobalDefines.keys(): __IncludeMacros['ECP_SOURCE'] = GlobalData.gGlobalDefines['ECP_SOURCE'] # # During GenFds phase call DSC parser, will go into this branch. # elif "ECP_SOURCE" in GlobalData.gCommandLineDefines.keys(): __IncludeMacros['ECP_SOURCE'] = GlobalData.gCommandLineDefines['ECP_SOURCE'] __IncludeMacros['EFI_SOURCE'] = GlobalData.gGlobalDefines['EFI_SOURCE'] __IncludeMacros['EDK_SOURCE'] = GlobalData.gGlobalDefines['EDK_SOURCE'] # # Allow using MACROs comes from [Defines] section to keep compatible. # __IncludeMacros.update(self._Macros) IncludedFile = NormPath(ReplaceMacro(self._ValueList[1], __IncludeMacros, RaiseError=True)) # # First search the include file under the same directory as DSC file # IncludedFile1 = PathClass(IncludedFile, self.MetaFile.Dir) ErrorCode, ErrorInfo1 = IncludedFile1.Validate() if ErrorCode != 0: # # Also search file under the WORKSPACE directory # IncludedFile1 = PathClass(IncludedFile, GlobalData.gWorkspace) ErrorCode, ErrorInfo2 = IncludedFile1.Validate() if ErrorCode != 0: EdkLogger.error('parser', ErrorCode, File=self._FileWithError, Line=self._LineIndex+1, ExtraData=ErrorInfo1 + "\n"+ ErrorInfo2) self._FileWithError = IncludedFile1 IncludedFileTable = MetaFileStorage(self._Table.Cur, IncludedFile1, MODEL_FILE_DSC, True) Owner = self._Content[self._ContentIndex-1][0] Parser = DscParser(IncludedFile1, self._FileType, IncludedFileTable, Owner=Owner, From=Owner) # set the parser status with current status Parser._SectionName = self._SectionName Parser._SectionType = self._SectionType Parser._Scope = self._Scope Parser._Enabled = self._Enabled # Parse the included file Parser.Start() # update current status with sub-parser's status self._SectionName = Parser._SectionName self._SectionType = Parser._SectionType self._Scope = Parser._Scope self._Enabled = Parser._Enabled # Insert all records in the table for the included file into dsc file table Records = IncludedFileTable.GetAll() if Records: self._Content[self._ContentIndex:self._ContentIndex] = Records self._Content.pop(self._ContentIndex-1) self._ValueList = None self._ContentIndex -= 1 def __ProcessSkuId(self): self._ValueList = [ReplaceMacro(Value, self._Macros, RaiseError=True) for Value in self._ValueList] def __ProcessLibraryInstance(self): self._ValueList = [ReplaceMacro(Value, self._Macros) for Value in self._ValueList] def __ProcessLibraryClass(self): self._ValueList[1] = ReplaceMacro(self._ValueList[1], self._Macros, RaiseError=True) def __ProcessPcd(self): ValueList = GetSplitValueList(self._ValueList[2]) # # PCD value can be an expression # if len(ValueList) > 1 and ValueList[1] == 'VOID*': PcdValue = ValueList[0] try: ValueList[0] = ValueExpression(PcdValue, self._Macros)(True) except WrnExpression, Value: ValueList[0] = Value.result else: PcdValue = ValueList[-1] try: ValueList[-1] = ValueExpression(PcdValue, self._Macros)(True) except WrnExpression, Value: ValueList[-1] = Value.result if ValueList[-1] == 'True': ValueList[-1] = '1' if ValueList[-1] == 'False': ValueList[-1] = '0' self._ValueList[2] = '|'.join(ValueList) def __ProcessComponent(self): self._ValueList[0] = ReplaceMacro(self._ValueList[0], self._Macros) def __ProcessSourceOverridePath(self): self._ValueList[0] = ReplaceMacro(self._ValueList[0], self._Macros) def __ProcessBuildOption(self): self._ValueList = [ReplaceMacro(Value, self._Macros, RaiseError=False) for Value in self._ValueList] _SectionParser = { MODEL_META_DATA_HEADER : _DefineParser, MODEL_EFI_SKU_ID : _SkuIdParser, MODEL_EFI_LIBRARY_INSTANCE : _LibraryInstanceParser, MODEL_EFI_LIBRARY_CLASS : _LibraryClassParser, MODEL_PCD_FIXED_AT_BUILD : _PcdParser, MODEL_PCD_PATCHABLE_IN_MODULE : _PcdParser, MODEL_PCD_FEATURE_FLAG : _PcdParser, MODEL_PCD_DYNAMIC_DEFAULT : _PcdParser, MODEL_PCD_DYNAMIC_HII : _PcdParser, MODEL_PCD_DYNAMIC_VPD : _PcdParser, MODEL_PCD_DYNAMIC_EX_DEFAULT : _PcdParser, MODEL_PCD_DYNAMIC_EX_HII : _PcdParser, MODEL_PCD_DYNAMIC_EX_VPD : _PcdParser, MODEL_META_DATA_COMPONENT : _ComponentParser, MODEL_META_DATA_COMPONENT_SOURCE_OVERRIDE_PATH : _CompponentSourceOverridePathParser, MODEL_META_DATA_BUILD_OPTION : _BuildOptionParser, MODEL_UNKNOWN : MetaFileParser._Skip, MODEL_META_DATA_USER_EXTENSION : MetaFileParser._Skip, MODEL_META_DATA_SECTION_HEADER : MetaFileParser._SectionHeaderParser, MODEL_META_DATA_SUBSECTION_HEADER : _SubsectionHeaderParser, } _Macros = property(_GetMacros) ## DEC file parser class # # @param FilePath The path of platform description file # @param FileType The raw data of DSC file # @param Table Database used to retrieve module/package information # @param Macros Macros used for replacement in file # class DecParser(MetaFileParser): # DEC file supported data types (one type per section) DataType = { TAB_DEC_DEFINES.upper() : MODEL_META_DATA_HEADER, TAB_DSC_DEFINES_DEFINE : MODEL_META_DATA_DEFINE, TAB_INCLUDES.upper() : MODEL_EFI_INCLUDE, TAB_LIBRARY_CLASSES.upper() : MODEL_EFI_LIBRARY_CLASS, TAB_GUIDS.upper() : MODEL_EFI_GUID, TAB_PPIS.upper() : MODEL_EFI_PPI, TAB_PROTOCOLS.upper() : MODEL_EFI_PROTOCOL, TAB_PCDS_FIXED_AT_BUILD_NULL.upper() : MODEL_PCD_FIXED_AT_BUILD, TAB_PCDS_PATCHABLE_IN_MODULE_NULL.upper() : MODEL_PCD_PATCHABLE_IN_MODULE, TAB_PCDS_FEATURE_FLAG_NULL.upper() : MODEL_PCD_FEATURE_FLAG, TAB_PCDS_DYNAMIC_NULL.upper() : MODEL_PCD_DYNAMIC, TAB_PCDS_DYNAMIC_EX_NULL.upper() : MODEL_PCD_DYNAMIC_EX, } ## Constructor of DecParser # # Initialize object of DecParser # # @param FilePath The path of platform description file # @param FileType The raw data of DSC file # @param Table Database used to retrieve module/package information # @param Macros Macros used for replacement in file # def __init__(self, FilePath, FileType, Table): # prevent re-initialization if hasattr(self, "_Table"): return MetaFileParser.__init__(self, FilePath, FileType, Table) self._Comments = [] self._Version = 0x00010005 # Only EDK2 dec file is supported self.TblFile = EccGlobalData.gDb.TblFile self.FileID = -1 ## Parser starter def Start(self): Content = '' try: Content = open(str(self.MetaFile), 'r').readlines() except: EdkLogger.error("Parser", FILE_READ_FAILURE, ExtraData=self.MetaFile) # # Insert a record for file # Filename = NormPath(self.MetaFile) FileID = self.TblFile.GetFileId(Filename) if FileID: self.FileID = FileID else: self.FileID = self.TblFile.InsertFile(Filename, MODEL_FILE_DEC) for Index in range(0, len(Content)): Line, Comment = CleanString2(Content[Index]) self._CurrentLine = Line self._LineIndex = Index # save comment for later use if Comment: self._Comments.append((Comment, self._LineIndex+1)) # skip empty line if Line == '': continue # section header if Line[0] == TAB_SECTION_START and Line[-1] == TAB_SECTION_END: self._SectionHeaderParser() self._Comments = [] continue elif len(self._SectionType) == 0: self._Comments = [] continue # section content self._ValueList = ['','',''] self._SectionParser[self._SectionType[0]](self) if self._ValueList == None or self._ItemType == MODEL_META_DATA_DEFINE: self._ItemType = -1 self._Comments = [] continue # # Model, Value1, Value2, Value3, Arch, BelongsToItem=-1, LineBegin=-1, # ColumnBegin=-1, LineEnd=-1, ColumnEnd=-1, FeatureFlag='', Enabled=-1 # for Arch, ModuleType, Type in self._Scope: self._LastItem = self._Store( Type, self._ValueList[0], self._ValueList[1], self._ValueList[2], Arch, ModuleType, self._Owner[-1], self.FileID, self._LineIndex+1, -1, self._LineIndex+1, -1, 0 ) for Comment, LineNo in self._Comments: self._Store( MODEL_META_DATA_COMMENT, Comment, self._ValueList[0], self._ValueList[1], Arch, ModuleType, self._LastItem, self.FileID, LineNo, -1, LineNo, -1, 0 ) self._Comments = [] self._Done() def _GetApplicableSectionMacro(self): Macros = {} for S1, S2, SectionType in self._Scope: for Scope1, Scope2 in [("COMMON", "COMMON"), ("COMMON", S2), (S1, "COMMON"), (S1, S2)]: if (SectionType, Scope1, Scope2) in self._SectionsMacroDict: Macros.update(self._SectionsMacroDict[(SectionType, Scope1, Scope2)]) return Macros ## Section header parser # # The section header is always in following format: # # [section_name.arch<.platform|module_type>] # def _SectionHeaderParser(self): self._Scope = [] self._SectionName = '' self._SectionType = [] ArchList = set() for Item in GetSplitValueList(self._CurrentLine[1:-1], TAB_COMMA_SPLIT): if Item == '': continue ItemList = GetSplitValueList(Item, TAB_SPLIT) # different types of PCD are permissible in one section self._SectionName = ItemList[0].upper() if self._SectionName in self.DataType: if self.DataType[self._SectionName] not in self._SectionType: self._SectionType.append(self.DataType[self._SectionName]) else: EdkLogger.warn("Parser", "Unrecognized section", File=self.MetaFile, Line=self._LineIndex+1, ExtraData=self._CurrentLine) continue if MODEL_PCD_FEATURE_FLAG in self._SectionType and len(self._SectionType) > 1: EdkLogger.error( 'Parser', FORMAT_INVALID, "%s must not be in the same section of other types of PCD" % TAB_PCDS_FEATURE_FLAG_NULL, File=self.MetaFile, Line=self._LineIndex+1, ExtraData=self._CurrentLine ) # S1 is always Arch if len(ItemList) > 1: S1 = ItemList[1].upper() else: S1 = 'COMMON' ArchList.add(S1) # S2 may be Platform or ModuleType if len(ItemList) > 2: S2 = ItemList[2].upper() else: S2 = 'COMMON' if [S1, S2, self.DataType[self._SectionName]] not in self._Scope: self._Scope.append([S1, S2, self.DataType[self._SectionName]]) # 'COMMON' must not be used with specific ARCHs at the same section if 'COMMON' in ArchList and len(ArchList) > 1: EdkLogger.error('Parser', FORMAT_INVALID, "'common' ARCH must not be used with specific ARCHs", File=self.MetaFile, Line=self._LineIndex+1, ExtraData=self._CurrentLine) ## [guids], [ppis] and [protocols] section parser @ParseMacro def _GuidParser(self): TokenList = GetSplitValueList(self._CurrentLine, TAB_EQUAL_SPLIT, 1) if len(TokenList) < 2: EdkLogger.error('Parser', FORMAT_INVALID, "No GUID name or value specified", ExtraData=self._CurrentLine + " ( = )", File=self.MetaFile, Line=self._LineIndex+1) if TokenList[0] == '': EdkLogger.error('Parser', FORMAT_INVALID, "No GUID name specified", ExtraData=self._CurrentLine + " ( = )", File=self.MetaFile, Line=self._LineIndex+1) if TokenList[1] == '': EdkLogger.error('Parser', FORMAT_INVALID, "No GUID value specified", ExtraData=self._CurrentLine + " ( = )", File=self.MetaFile, Line=self._LineIndex+1) if TokenList[1][0] != '{' or TokenList[1][-1] != '}' or GuidStructureStringToGuidString(TokenList[1]) == '': EdkLogger.error('Parser', FORMAT_INVALID, "Invalid GUID value format", ExtraData=self._CurrentLine + \ " ( = )", File=self.MetaFile, Line=self._LineIndex+1) self._ValueList[0] = TokenList[0] #Parse the Guid value format GuidValueList = TokenList[1].strip(' {}').split(',') Index = 0 HexList = [] if len(GuidValueList) == 11: for GuidValue in GuidValueList: GuidValue = GuidValue.strip() if GuidValue.startswith('0x') or GuidValue.startswith('0X'): HexList.append('0x' + str(GuidValue[2:])) Index += 1 continue else: if GuidValue.startswith('{'): GuidValue = GuidValue.lstrip(' {') HexList.append('0x' + str(GuidValue[2:])) Index += 1 self._ValueList[1] = "{ %s, %s, %s, { %s, %s, %s, %s, %s, %s, %s, %s }}" % (HexList[0], HexList[1], HexList[2],HexList[3],HexList[4],HexList[5],HexList[6],HexList[7],HexList[8],HexList[9],HexList[10]) else: EdkLogger.error('Parser', FORMAT_INVALID, "Invalid GUID value format", ExtraData=self._CurrentLine + \ " ( = )", File=self.MetaFile, Line=self._LineIndex+1) self._ValueList[0] = '' ## PCD sections parser # # [PcdsFixedAtBuild] # [PcdsPatchableInModule] # [PcdsFeatureFlag] # [PcdsDynamicEx # [PcdsDynamic] # @ParseMacro def _PcdParser(self): TokenList = GetSplitValueList(self._CurrentLine, TAB_VALUE_SPLIT, 1) self._ValueList[0:1] = GetSplitValueList(TokenList[0], TAB_SPLIT) # check PCD information if self._ValueList[0] == '' or self._ValueList[1] == '': EdkLogger.error('Parser', FORMAT_INVALID, "No token space GUID or PCD name specified", ExtraData=self._CurrentLine + \ " (.|||)", File=self.MetaFile, Line=self._LineIndex+1) # check PCD datum information if len(TokenList) < 2 or TokenList[1] == '': EdkLogger.error('Parser', FORMAT_INVALID, "No PCD Datum information given", ExtraData=self._CurrentLine + \ " (.|||)", File=self.MetaFile, Line=self._LineIndex+1) ValueRe = re.compile(r'^\s*L?\".*\|.*\"') PtrValue = ValueRe.findall(TokenList[1]) # Has VOID* type string, may contain "|" character in the string. if len(PtrValue) != 0: ptrValueList = re.sub(ValueRe, '', TokenList[1]) ValueList = GetSplitValueList(ptrValueList) ValueList[0] = PtrValue[0] else: ValueList = GetSplitValueList(TokenList[1]) # check if there's enough datum information given if len(ValueList) != 3: EdkLogger.error('Parser', FORMAT_INVALID, "Invalid PCD Datum information given", ExtraData=self._CurrentLine + \ " (.|||)", File=self.MetaFile, Line=self._LineIndex+1) # check default value if ValueList[0] == '': EdkLogger.error('Parser', FORMAT_INVALID, "Missing DefaultValue in PCD Datum information", ExtraData=self._CurrentLine + \ " (.|||)", File=self.MetaFile, Line=self._LineIndex+1) # check datum type if ValueList[1] == '': EdkLogger.error('Parser', FORMAT_INVALID, "Missing DatumType in PCD Datum information", ExtraData=self._CurrentLine + \ " (.|||)", File=self.MetaFile, Line=self._LineIndex+1) # check token of the PCD if ValueList[2] == '': EdkLogger.error('Parser', FORMAT_INVALID, "Missing Token in PCD Datum information", ExtraData=self._CurrentLine + \ " (.|||)", File=self.MetaFile, Line=self._LineIndex+1) # check format of default value against the datum type IsValid, Cause = CheckPcdDatum(ValueList[1], ValueList[0]) if not IsValid: EdkLogger.error('Parser', FORMAT_INVALID, Cause, ExtraData=self._CurrentLine, File=self.MetaFile, Line=self._LineIndex+1) if EccGlobalData.gConfig.UniCheckPCDInfo == '1' or EccGlobalData.gConfig.UniCheckAll == '1' or EccGlobalData.gConfig.CheckAll == '1': # check Description, Prompt information PatternDesc = re.compile('##\s*([\x21-\x7E\s]*)', re.S) PatternPrompt = re.compile('#\s+@Prompt\s+([\x21-\x7E\s]*)', re.S) Description = None Prompt = None # check @ValidRange, @ValidList and @Expression format valid ErrorCodeValid = '0x0 <= %s <= 0xFFFFFFFF' PatternValidRangeIn = '(NOT)?\s*(\d+\s*-\s*\d+|0[xX][a-fA-F0-9]+\s*-\s*0[xX][a-fA-F0-9]+|LT\s*\d+|LT\s*0[xX][a-fA-F0-9]+|GT\s*\d+|GT\s*0[xX][a-fA-F0-9]+|LE\s*\d+|LE\s*0[xX][a-fA-F0-9]+|GE\s*\d+|GE\s*0[xX][a-fA-F0-9]+|XOR\s*\d+|XOR\s*0[xX][a-fA-F0-9]+|EQ\s*\d+|EQ\s*0[xX][a-fA-F0-9]+)' PatternValidRng = re.compile('^' + '(NOT)?\s*' + PatternValidRangeIn + '$') for Comment in self._Comments: Comm = Comment[0].strip() if not Comm: continue if not Description: Description = PatternDesc.findall(Comm) if not Prompt: Prompt = PatternPrompt.findall(Comm) if Comm[0] == '#': ValidFormt = Comm.lstrip('#') ValidFormt = ValidFormt.lstrip() if ValidFormt[0:11] == '@ValidRange': ValidFormt = ValidFormt[11:] ValidFormt = ValidFormt.lstrip() try: ErrorCode, Expression = ValidFormt.split('|', 1) except ValueError: ErrorCode = '0x0' Expression = ValidFormt ErrorCode, Expression = ErrorCode.strip(), Expression.strip() try: if not eval(ErrorCodeValid % ErrorCode): EdkLogger.warn('Parser', '@ValidRange ErrorCode(%s) of PCD %s is not valid UINT32 value.' % (ErrorCode, TokenList[0])) except: EdkLogger.warn('Parser', '@ValidRange ErrorCode(%s) of PCD %s is not valid UINT32 value.' % (ErrorCode, TokenList[0])) if not PatternValidRng.search(Expression): EdkLogger.warn('Parser', '@ValidRange Expression(%s) of PCD %s is incorrect format.' % (Expression, TokenList[0])) if ValidFormt[0:10] == '@ValidList': ValidFormt = ValidFormt[10:] ValidFormt = ValidFormt.lstrip() try: ErrorCode, Expression = ValidFormt.split('|', 1) except ValueError: ErrorCode = '0x0' Expression = ValidFormt ErrorCode, Expression = ErrorCode.strip(), Expression.strip() try: if not eval(ErrorCodeValid % ErrorCode): EdkLogger.warn('Parser', '@ValidList ErrorCode(%s) of PCD %s is not valid UINT32 value.' % (ErrorCode, TokenList[0])) except: EdkLogger.warn('Parser', '@ValidList ErrorCode(%s) of PCD %s is not valid UINT32 value.' % (ErrorCode, TokenList[0])) Values = Expression.split(',') for Value in Values: Value = Value.strip() try: eval(Value) except: EdkLogger.warn('Parser', '@ValidList Expression of PCD %s include a invalid value(%s).' % (TokenList[0], Value)) break if ValidFormt[0:11] == '@Expression': ValidFormt = ValidFormt[11:] ValidFormt = ValidFormt.lstrip() try: ErrorCode, Expression = ValidFormt.split('|', 1) except ValueError: ErrorCode = '0x0' Expression = ValidFormt ErrorCode, Expression = ErrorCode.strip(), Expression.strip() try: if not eval(ErrorCodeValid % ErrorCode): EdkLogger.warn('Parser', '@Expression ErrorCode(%s) of PCD %s is not valid UINT32 value.' % (ErrorCode, TokenList[0])) except: EdkLogger.warn('Parser', '@Expression ErrorCode(%s) of PCD %s is not valid UINT32 value.' % (ErrorCode, TokenList[0])) if not Expression: EdkLogger.warn('Parser', '@Expression Expression of PCD %s is incorrect format.' % TokenList[0]) if not Description: EdkLogger.warn('Parser', 'PCD %s Description information is not provided.' % TokenList[0]) if not Prompt: EdkLogger.warn('Parser', 'PCD %s Prompt information is not provided.' % TokenList[0]) # check Description, Prompt localization information if self._UniObj: self._UniObj.CheckPcdInfo(TokenList[0]) if ValueList[0] in ['True', 'true', 'TRUE']: ValueList[0] = '1' elif ValueList[0] in ['False', 'false', 'FALSE']: ValueList[0] = '0' self._ValueList[2] = ValueList[0].strip() + '|' + ValueList[1].strip() + '|' + ValueList[2].strip() _SectionParser = { MODEL_META_DATA_HEADER : MetaFileParser._DefineParser, MODEL_EFI_INCLUDE : MetaFileParser._PathParser, MODEL_EFI_LIBRARY_CLASS : MetaFileParser._PathParser, MODEL_EFI_GUID : _GuidParser, MODEL_EFI_PPI : _GuidParser, MODEL_EFI_PROTOCOL : _GuidParser, MODEL_PCD_FIXED_AT_BUILD : _PcdParser, MODEL_PCD_PATCHABLE_IN_MODULE : _PcdParser, MODEL_PCD_FEATURE_FLAG : _PcdParser, MODEL_PCD_DYNAMIC : _PcdParser, MODEL_PCD_DYNAMIC_EX : _PcdParser, MODEL_UNKNOWN : MetaFileParser._Skip, MODEL_META_DATA_USER_EXTENSION : MetaFileParser._Skip, } ## FdfObject # # This class defined basic Fdf object which is used by inheriting # # @param object: Inherited from object class # class FdfObject(object): def __init__(self): object.__init__() ## Fdf # # This class defined the structure used in Fdf object # # @param FdfObject: Inherited from FdfObject class # @param Filename: Input value for Ffilename of Fdf file, default is None # @param WorkspaceDir: Input value for current workspace directory, default is None # class Fdf(FdfObject): def __init__(self, Filename = None, IsToDatabase = False, WorkspaceDir = None, Database = None): self.WorkspaceDir = WorkspaceDir self.IsToDatabase = IsToDatabase self.Cur = Database.Cur self.TblFile = Database.TblFile self.TblFdf = Database.TblFdf self.FileID = -1 self.FileList = {} # # Load Fdf file if filename is not None # if Filename != None: try: self.LoadFdfFile(Filename) except Exception: pass # # Insert a FDF file record into database # def InsertFile(self, Filename): FileID = -1 Filename = NormPath(Filename) if Filename not in self.FileList: FileID = self.TblFile.InsertFile(Filename, MODEL_FILE_FDF) self.FileList[Filename] = FileID return self.FileList[Filename] ## Load Fdf file # # Load the file if it exists # # @param Filename: Input value for filename of Fdf file # def LoadFdfFile(self, Filename): FileList = [] # # Parse Fdf file # Filename = NormPath(Filename) Fdf = FdfParser(Filename) Fdf.ParseFile() # # Insert inf file and pcd information # if self.IsToDatabase: (Model, Value1, Value2, Value3, Scope1, Scope2, BelongsToItem, BelongsToFile, StartLine, StartColumn, EndLine, EndColumn, Enabled) = \ (0, '', '', '', 'COMMON', 'COMMON', -1, -1, -1, -1, -1, -1, 0) for Index in range(0, len(Fdf.Profile.PcdDict)): pass for Key in Fdf.Profile.PcdDict.keys(): Model = MODEL_PCD Value1 = Key[1] Value2 = Key[0] FileName = Fdf.Profile.PcdFileLineDict[Key][0] StartLine = Fdf.Profile.PcdFileLineDict[Key][1] BelongsToFile = self.InsertFile(FileName) self.TblFdf.Insert(Model, Value1, Value2, Value3, Scope1, Scope2, BelongsToItem, BelongsToFile, StartLine, StartColumn, EndLine, EndColumn, Enabled) for Index in range(0, len(Fdf.Profile.InfList)): Model = MODEL_META_DATA_COMPONENT Value1 = Fdf.Profile.InfList[Index] Value2 = '' FileName = Fdf.Profile.InfFileLineList[Index][0] StartLine = Fdf.Profile.InfFileLineList[Index][1] BelongsToFile = self.InsertFile(FileName) self.TblFdf.Insert(Model, Value1, Value2, Value3, Scope1, Scope2, BelongsToItem, BelongsToFile, StartLine, StartColumn, EndLine, EndColumn, Enabled) class UniParser(object): # IsExtraUni defined the UNI file is Module UNI or extra Module UNI # IsModuleUni defined the UNI file is Module UNI or Package UNI def __init__(self, FilePath, IsExtraUni=False, IsModuleUni=True): self.FilePath = FilePath self.FileName = os.path.basename(FilePath) self.IsExtraUni = IsExtraUni self.IsModuleUni = IsModuleUni self.FileIn = None self.Missing = [] self.__read() def __read(self): try: self.FileIn = CodecOpenLongFilePath(self.FilePath, Mode='rb', Encoding='utf_8').read() except UnicodeError: self.FileIn = CodecOpenLongFilePath(self.FilePath, Mode='rb', Encoding='utf_16').read() except UnicodeError: self.FileIn = CodecOpenLongFilePath(self.FilePath, Mode='rb', Encoding='utf_16_le').read() except IOError: self.FileIn = "" def Start(self): if self.IsModuleUni: if self.IsExtraUni: ModuleName = self.CheckKeyValid('STR_PROPERTIES_MODULE_NAME') self.PrintLog('STR_PROPERTIES_MODULE_NAME', ModuleName) else: ModuleAbstract = self.CheckKeyValid('STR_MODULE_ABSTRACT') self.PrintLog('STR_MODULE_ABSTRACT', ModuleAbstract) ModuleDescription = self.CheckKeyValid('STR_MODULE_DESCRIPTION') self.PrintLog('STR_MODULE_DESCRIPTION', ModuleDescription) else: if self.IsExtraUni: PackageName = self.CheckKeyValid('STR_PROPERTIES_PACKAGE_NAME') self.PrintLog('STR_PROPERTIES_PACKAGE_NAME', PackageName) else: PackageAbstract = self.CheckKeyValid('STR_PACKAGE_ABSTRACT') self.PrintLog('STR_PACKAGE_ABSTRACT', PackageAbstract) PackageDescription = self.CheckKeyValid('STR_PACKAGE_DESCRIPTION') self.PrintLog('STR_PACKAGE_DESCRIPTION', PackageDescription) def CheckKeyValid(self, Key, Contents=None): if not Contents: Contents = self.FileIn KeyPattern = re.compile('#string\s+%s\s+.*?#language.*?".*?"' % Key, re.S) if KeyPattern.search(Contents): return True return False def CheckPcdInfo(self, PcdCName): PromptKey = 'STR_%s_PROMPT' % PcdCName.replace('.', '_') PcdPrompt = self.CheckKeyValid(PromptKey) self.PrintLog(PromptKey, PcdPrompt) HelpKey = 'STR_%s_HELP' % PcdCName.replace('.', '_') PcdHelp = self.CheckKeyValid(HelpKey) self.PrintLog(HelpKey, PcdHelp) def PrintLog(self, Key, Value): if not Value and Key not in self.Missing: Msg = '%s is missing in the %s file.' % (Key, self.FileName) EdkLogger.warn('Parser', Msg) EccGlobalData.gDb.TblReport.Insert(EccToolError.ERROR_GENERAL_CHECK_UNI_HELP_INFO, OtherMsg=Msg, BelongsToTable='File', BelongsToItem=-2) self.Missing.append(Key) ## # # This acts like the main() function for the script, unless it is 'import'ed into another # script. # if __name__ == '__main__': pass