/gem5/ext/ply/test/ |
H A D | yacc_sr.py | 11 from calclex import tokens
|
H A D | yacc_term1.py | 11 from calclex import tokens
|
H A D | yacc_unused.py | 11 from calclex import tokens
|
H A D | yacc_unused_rule.py | 11 from calclex import tokens
|
H A D | yacc_uprec.py | 11 from calclex import tokens
|
H A D | yacc_uprec2.py | 11 from calclex import tokens
|
/gem5/ext/googletest/googletest/scripts/ |
H A D | pump.py | 383 """A generator that yields the tokens in the given string.""" 475 def ParseElseNode(tokens): 477 return PopToken(tokens, token_type) 479 next = PeekToken(tokens) 485 code_node = ParseCodeNode(tokens) 492 code_node = ParseCodeNode(tokens) 494 inner_else_node = ParseElseNode(tokens) 498 return ParseElseNode(tokens) 503 def ParseAtomicCodeNode(tokens): 505 return PopToken(tokens, token_typ [all...] |
/gem5/ext/ply/example/GardenSnake/ |
H A D | GardenSnake.py | 47 tokens = ( variable 125 # Don't generate newline tokens when inside of parenthesis, eg 155 # WS will only occur before any other tokens on a line. 157 # I have three filters. One tags tokens by adding two attributes. 172 def track_tokens_filter(lexer, tokens): 176 for token in tokens: 224 def indentation_filter(tokens): 230 for token in tokens: 247 # WS tokens are never passed to the parser 303 tokens [all...] |
/gem5/ext/ply/example/yply/ |
H A D | yparse.py | 7 tokens = ylex.tokens variable 23 print "tokens = ", repr(tokenlist)
|
H A D | ylex.py | 11 tokens = ( variable
|
/gem5/ext/ply/example/newclasscalc/ |
H A D | calc.py | 28 tokens = () variable in class:Parser 62 tokens = ( variable in class:Calc
|
/gem5/ext/ply/example/classcalc/ |
H A D | calc.py | 26 tokens = () variable in class:Parser 59 tokens = ( variable in class:Calc
|
/gem5/ext/ply/example/calc/ |
H A D | calc.py | 14 tokens = ( variable
|
/gem5/ext/ply/example/calcdebug/ |
H A D | calc.py | 14 tokens = ( variable
|
/gem5/ext/ply/example/closurecalc/ |
H A D | calc.py | 27 tokens = (
|
/gem5/ext/ply/example/unicalc/ |
H A D | calc.py | 7 # This example uses unicode strings for tokens, docstrings, and input. 13 tokens = ( variable
|
/gem5/ext/ply/ply/ |
H A D | ctokens.py | 10 tokens = [ variable
|
H A D | lex.py | 68 # Token class. This class is used to represent the tokens produced. 134 self.lextokens = None # List of valid tokens 548 self.tokens = [] 573 # Get the tokens map 575 tokens = self.ldict.get("tokens",None) 576 if not tokens: 581 if not isinstance(tokens,(list, tuple)): 582 self.log.error("tokens must be a list or tuple") 586 if not tokens [all...] |
H A D | yacc.py | 1399 # Look for literal tokens 1579 # Find all symbols that were used the grammar, but not defined as tokens or 2759 # start symbol, error function, tokens, precedence list, action functions, 2767 self.tokens = None 2807 if self.tokens: 2808 sig.update(" ".join(self.tokens).encode('latin-1')) 2888 # Get the tokens map 2890 tokens = self.pdict.get("tokens",None) 2891 if not tokens [all...] |
/gem5/ext/ply/example/optcalc/ |
H A D | calc.py | 14 tokens = ( variable
|
/gem5/ext/ply/example/ansic/ |
H A D | clex.py | 20 tokens = reserved + ( variable
|
H A D | cparse.py | 12 tokens = clex.tokens variable
|
/gem5/ext/ply/example/BASIC/ |
H A D | basparse.py | 7 tokens = basiclex.tokens variable
|
/gem5/src/mem/slicc/ |
H A D | parser.py | 130 tokens = [ 'EQ', 'NE', 'LT', 'GT', 'LE', 'GE', variable in class:SLICC 138 tokens += reserved.values()
|
/gem5/src/arch/ |
H A D | micro_asm.py | 185 tokens = reserved + ( variable 292 # Basic regular expressions to pick out simple tokens
|