You can not select more than 25 topics
			Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
		
		
		
		
		
			
		
			
				
					
					
						
							855 lines
						
					
					
						
							23 KiB
						
					
					
				
			
		
		
		
			
			
			
				
					
				
				
					
				
			
		
		
	
	
							855 lines
						
					
					
						
							23 KiB
						
					
					
				| #!/usr/bin/env python | |
| # | |
| # Copyright 2008, Google Inc. | |
| # All rights reserved. | |
| # | |
| # Redistribution and use in source and binary forms, with or without | |
| # modification, are permitted provided that the following conditions are | |
| # met: | |
| # | |
| #     * Redistributions of source code must retain the above copyright | |
| # notice, this list of conditions and the following disclaimer. | |
| #     * Redistributions in binary form must reproduce the above | |
| # copyright notice, this list of conditions and the following disclaimer | |
| # in the documentation and/or other materials provided with the | |
| # distribution. | |
| #     * Neither the name of Google Inc. nor the names of its | |
| # contributors may be used to endorse or promote products derived from | |
| # this software without specific prior written permission. | |
| # | |
| # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS | |
| # "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT | |
| # LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR | |
| # A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT | |
| # OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, | |
| # SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT | |
| # LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, | |
| # DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY | |
| # THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT | |
| # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE | |
| # OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. | |
| 
 | |
| """pump v0.2.0 - Pretty Useful for Meta Programming. | |
|  | |
| A tool for preprocessor meta programming.  Useful for generating | |
| repetitive boilerplate code.  Especially useful for writing C++ | |
| classes, functions, macros, and templates that need to work with | |
| various number of arguments. | |
|  | |
| USAGE: | |
|        pump.py SOURCE_FILE | |
|  | |
| EXAMPLES: | |
|        pump.py foo.cc.pump | |
|          Converts foo.cc.pump to foo.cc. | |
|  | |
| GRAMMAR: | |
|        CODE ::= ATOMIC_CODE* | |
|        ATOMIC_CODE ::= $var ID = EXPRESSION | |
|            | $var ID = [[ CODE ]] | |
|            | $range ID EXPRESSION..EXPRESSION | |
|            | $for ID SEPARATOR [[ CODE ]] | |
|            | $($) | |
|            | $ID | |
|            | $(EXPRESSION) | |
|            | $if EXPRESSION [[ CODE ]] ELSE_BRANCH | |
|            | [[ CODE ]] | |
|            | RAW_CODE | |
|        SEPARATOR ::= RAW_CODE | EMPTY | |
|        ELSE_BRANCH ::= $else [[ CODE ]] | |
|            | $elif EXPRESSION [[ CODE ]] ELSE_BRANCH | |
|            | EMPTY | |
|        EXPRESSION has Python syntax. | |
| """ | |
| 
 | |
| __author__ = 'wan@google.com (Zhanyong Wan)' | |
| 
 | |
| import os | |
| import re | |
| import sys | |
| 
 | |
| 
 | |
| TOKEN_TABLE = [ | |
|     (re.compile(r'\$var\s+'), '$var'), | |
|     (re.compile(r'\$elif\s+'), '$elif'), | |
|     (re.compile(r'\$else\s+'), '$else'), | |
|     (re.compile(r'\$for\s+'), '$for'), | |
|     (re.compile(r'\$if\s+'), '$if'), | |
|     (re.compile(r'\$range\s+'), '$range'), | |
|     (re.compile(r'\$[_A-Za-z]\w*'), '$id'), | |
|     (re.compile(r'\$\(\$\)'), '$($)'), | |
|     (re.compile(r'\$'), '$'), | |
|     (re.compile(r'\[\[\n?'), '[['), | |
|     (re.compile(r'\]\]\n?'), ']]'), | |
|     ] | |
| 
 | |
| 
 | |
| class Cursor: | |
|   """Represents a position (line and column) in a text file.""" | |
| 
 | |
|   def __init__(self, line=-1, column=-1): | |
|     self.line = line | |
|     self.column = column | |
| 
 | |
|   def __eq__(self, rhs): | |
|     return self.line == rhs.line and self.column == rhs.column | |
| 
 | |
|   def __ne__(self, rhs): | |
|     return not self == rhs | |
| 
 | |
|   def __lt__(self, rhs): | |
|     return self.line < rhs.line or ( | |
|         self.line == rhs.line and self.column < rhs.column) | |
| 
 | |
|   def __le__(self, rhs): | |
|     return self < rhs or self == rhs | |
| 
 | |
|   def __gt__(self, rhs): | |
|     return rhs < self | |
| 
 | |
|   def __ge__(self, rhs): | |
|     return rhs <= self | |
| 
 | |
|   def __str__(self): | |
|     if self == Eof(): | |
|       return 'EOF' | |
|     else: | |
|       return '%s(%s)' % (self.line + 1, self.column) | |
| 
 | |
|   def __add__(self, offset): | |
|     return Cursor(self.line, self.column + offset) | |
| 
 | |
|   def __sub__(self, offset): | |
|     return Cursor(self.line, self.column - offset) | |
| 
 | |
|   def Clone(self): | |
|     """Returns a copy of self.""" | |
| 
 | |
|     return Cursor(self.line, self.column) | |
| 
 | |
| 
 | |
| # Special cursor to indicate the end-of-file. | |
| def Eof(): | |
|   """Returns the special cursor to denote the end-of-file.""" | |
|   return Cursor(-1, -1) | |
| 
 | |
| 
 | |
| class Token: | |
|   """Represents a token in a Pump source file.""" | |
| 
 | |
|   def __init__(self, start=None, end=None, value=None, token_type=None): | |
|     if start is None: | |
|       self.start = Eof() | |
|     else: | |
|       self.start = start | |
|     if end is None: | |
|       self.end = Eof() | |
|     else: | |
|       self.end = end | |
|     self.value = value | |
|     self.token_type = token_type | |
| 
 | |
|   def __str__(self): | |
|     return 'Token @%s: \'%s\' type=%s' % ( | |
|         self.start, self.value, self.token_type) | |
| 
 | |
|   def Clone(self): | |
|     """Returns a copy of self.""" | |
| 
 | |
|     return Token(self.start.Clone(), self.end.Clone(), self.value, | |
|                  self.token_type) | |
| 
 | |
| 
 | |
| def StartsWith(lines, pos, string): | |
|   """Returns True iff the given position in lines starts with 'string'.""" | |
| 
 | |
|   return lines[pos.line][pos.column:].startswith(string) | |
| 
 | |
| 
 | |
| def FindFirstInLine(line, token_table): | |
|   best_match_start = -1 | |
|   for (regex, token_type) in token_table: | |
|     m = regex.search(line) | |
|     if m: | |
|       # We found regex in lines | |
|       if best_match_start < 0 or m.start() < best_match_start: | |
|         best_match_start = m.start() | |
|         best_match_length = m.end() - m.start() | |
|         best_match_token_type = token_type | |
| 
 | |
|   if best_match_start < 0: | |
|     return None | |
| 
 | |
|   return (best_match_start, best_match_length, best_match_token_type) | |
| 
 | |
| 
 | |
| def FindFirst(lines, token_table, cursor): | |
|   """Finds the first occurrence of any string in strings in lines.""" | |
| 
 | |
|   start = cursor.Clone() | |
|   cur_line_number = cursor.line | |
|   for line in lines[start.line:]: | |
|     if cur_line_number == start.line: | |
|       line = line[start.column:] | |
|     m = FindFirstInLine(line, token_table) | |
|     if m: | |
|       # We found a regex in line. | |
|       (start_column, length, token_type) = m | |
|       if cur_line_number == start.line: | |
|         start_column += start.column | |
|       found_start = Cursor(cur_line_number, start_column) | |
|       found_end = found_start + length | |
|       return MakeToken(lines, found_start, found_end, token_type) | |
|     cur_line_number += 1 | |
|   # We failed to find str in lines | |
|   return None | |
| 
 | |
| 
 | |
| def SubString(lines, start, end): | |
|   """Returns a substring in lines.""" | |
| 
 | |
|   if end == Eof(): | |
|     end = Cursor(len(lines) - 1, len(lines[-1])) | |
| 
 | |
|   if start >= end: | |
|     return '' | |
| 
 | |
|   if start.line == end.line: | |
|     return lines[start.line][start.column:end.column] | |
| 
 | |
|   result_lines = ([lines[start.line][start.column:]] + | |
|                   lines[start.line + 1:end.line] + | |
|                   [lines[end.line][:end.column]]) | |
|   return ''.join(result_lines) | |
| 
 | |
| 
 | |
| def StripMetaComments(str): | |
|   """Strip meta comments from each line in the given string.""" | |
| 
 | |
|   # First, completely remove lines containing nothing but a meta | |
|   # comment, including the trailing \n. | |
|   str = re.sub(r'^\s*\$\$.*\n', '', str) | |
| 
 | |
|   # Then, remove meta comments from contentful lines. | |
|   return re.sub(r'\s*\$\$.*', '', str) | |
| 
 | |
| 
 | |
| def MakeToken(lines, start, end, token_type): | |
|   """Creates a new instance of Token.""" | |
| 
 | |
|   return Token(start, end, SubString(lines, start, end), token_type) | |
| 
 | |
| 
 | |
| def ParseToken(lines, pos, regex, token_type): | |
|   line = lines[pos.line][pos.column:] | |
|   m = regex.search(line) | |
|   if m and not m.start(): | |
|     return MakeToken(lines, pos, pos + m.end(), token_type) | |
|   else: | |
|     print 'ERROR: %s expected at %s.' % (token_type, pos) | |
|     sys.exit(1) | |
| 
 | |
| 
 | |
| ID_REGEX = re.compile(r'[_A-Za-z]\w*') | |
| EQ_REGEX = re.compile(r'=') | |
| REST_OF_LINE_REGEX = re.compile(r'.*?(?=$|\$\$)') | |
| OPTIONAL_WHITE_SPACES_REGEX = re.compile(r'\s*') | |
| WHITE_SPACE_REGEX = re.compile(r'\s') | |
| DOT_DOT_REGEX = re.compile(r'\.\.') | |
| 
 | |
| 
 | |
| def Skip(lines, pos, regex): | |
|   line = lines[pos.line][pos.column:] | |
|   m = re.search(regex, line) | |
|   if m and not m.start(): | |
|     return pos + m.end() | |
|   else: | |
|     return pos | |
| 
 | |
| 
 | |
| def SkipUntil(lines, pos, regex, token_type): | |
|   line = lines[pos.line][pos.column:] | |
|   m = re.search(regex, line) | |
|   if m: | |
|     return pos + m.start() | |
|   else: | |
|     print ('ERROR: %s expected on line %s after column %s.' % | |
|            (token_type, pos.line + 1, pos.column)) | |
|     sys.exit(1) | |
| 
 | |
| 
 | |
| def ParseExpTokenInParens(lines, pos): | |
|   def ParseInParens(pos): | |
|     pos = Skip(lines, pos, OPTIONAL_WHITE_SPACES_REGEX) | |
|     pos = Skip(lines, pos, r'\(') | |
|     pos = Parse(pos) | |
|     pos = Skip(lines, pos, r'\)') | |
|     return pos | |
| 
 | |
|   def Parse(pos): | |
|     pos = SkipUntil(lines, pos, r'\(|\)', ')') | |
|     if SubString(lines, pos, pos + 1) == '(': | |
|       pos = Parse(pos + 1) | |
|       pos = Skip(lines, pos, r'\)') | |
|       return Parse(pos) | |
|     else: | |
|       return pos | |
| 
 | |
|   start = pos.Clone() | |
|   pos = ParseInParens(pos) | |
|   return MakeToken(lines, start, pos, 'exp') | |
| 
 | |
| 
 | |
| def RStripNewLineFromToken(token): | |
|   if token.value.endswith('\n'): | |
|     return Token(token.start, token.end, token.value[:-1], token.token_type) | |
|   else: | |
|     return token | |
| 
 | |
| 
 | |
| def TokenizeLines(lines, pos): | |
|   while True: | |
|     found = FindFirst(lines, TOKEN_TABLE, pos) | |
|     if not found: | |
|       yield MakeToken(lines, pos, Eof(), 'code') | |
|       return | |
| 
 | |
|     if found.start == pos: | |
|       prev_token = None | |
|       prev_token_rstripped = None | |
|     else: | |
|       prev_token = MakeToken(lines, pos, found.start, 'code') | |
|       prev_token_rstripped = RStripNewLineFromToken(prev_token) | |
| 
 | |
|     if found.token_type == '$var': | |
|       if prev_token_rstripped: | |
|         yield prev_token_rstripped | |
|       yield found | |
|       id_token = ParseToken(lines, found.end, ID_REGEX, 'id') | |
|       yield id_token | |
|       pos = Skip(lines, id_token.end, OPTIONAL_WHITE_SPACES_REGEX) | |
| 
 | |
|       eq_token = ParseToken(lines, pos, EQ_REGEX, '=') | |
|       yield eq_token | |
|       pos = Skip(lines, eq_token.end, r'\s*') | |
| 
 | |
|       if SubString(lines, pos, pos + 2) != '[[': | |
|         exp_token = ParseToken(lines, pos, REST_OF_LINE_REGEX, 'exp') | |
|         yield exp_token | |
|         pos = Cursor(exp_token.end.line + 1, 0) | |
|     elif found.token_type == '$for': | |
|       if prev_token_rstripped: | |
|         yield prev_token_rstripped | |
|       yield found | |
|       id_token = ParseToken(lines, found.end, ID_REGEX, 'id') | |
|       yield id_token | |
|       pos = Skip(lines, id_token.end, WHITE_SPACE_REGEX) | |
|     elif found.token_type == '$range': | |
|       if prev_token_rstripped: | |
|         yield prev_token_rstripped | |
|       yield found | |
|       id_token = ParseToken(lines, found.end, ID_REGEX, 'id') | |
|       yield id_token | |
|       pos = Skip(lines, id_token.end, OPTIONAL_WHITE_SPACES_REGEX) | |
| 
 | |
|       dots_pos = SkipUntil(lines, pos, DOT_DOT_REGEX, '..') | |
|       yield MakeToken(lines, pos, dots_pos, 'exp') | |
|       yield MakeToken(lines, dots_pos, dots_pos + 2, '..') | |
|       pos = dots_pos + 2 | |
|       new_pos = Cursor(pos.line + 1, 0) | |
|       yield MakeToken(lines, pos, new_pos, 'exp') | |
|       pos = new_pos | |
|     elif found.token_type == '$': | |
|       if prev_token: | |
|         yield prev_token | |
|       yield found | |
|       exp_token = ParseExpTokenInParens(lines, found.end) | |
|       yield exp_token | |
|       pos = exp_token.end | |
|     elif (found.token_type == ']]' or found.token_type == '$if' or | |
|           found.token_type == '$elif' or found.token_type == '$else'): | |
|       if prev_token_rstripped: | |
|         yield prev_token_rstripped | |
|       yield found | |
|       pos = found.end | |
|     else: | |
|       if prev_token: | |
|         yield prev_token | |
|       yield found | |
|       pos = found.end | |
| 
 | |
| 
 | |
| def Tokenize(s): | |
|   """A generator that yields the tokens in the given string.""" | |
|   if s != '': | |
|     lines = s.splitlines(True) | |
|     for token in TokenizeLines(lines, Cursor(0, 0)): | |
|       yield token | |
| 
 | |
| 
 | |
| class CodeNode: | |
|   def __init__(self, atomic_code_list=None): | |
|     self.atomic_code = atomic_code_list | |
| 
 | |
| 
 | |
| class VarNode: | |
|   def __init__(self, identifier=None, atomic_code=None): | |
|     self.identifier = identifier | |
|     self.atomic_code = atomic_code | |
| 
 | |
| 
 | |
| class RangeNode: | |
|   def __init__(self, identifier=None, exp1=None, exp2=None): | |
|     self.identifier = identifier | |
|     self.exp1 = exp1 | |
|     self.exp2 = exp2 | |
| 
 | |
| 
 | |
| class ForNode: | |
|   def __init__(self, identifier=None, sep=None, code=None): | |
|     self.identifier = identifier | |
|     self.sep = sep | |
|     self.code = code | |
| 
 | |
| 
 | |
| class ElseNode: | |
|   def __init__(self, else_branch=None): | |
|     self.else_branch = else_branch | |
| 
 | |
| 
 | |
| class IfNode: | |
|   def __init__(self, exp=None, then_branch=None, else_branch=None): | |
|     self.exp = exp | |
|     self.then_branch = then_branch | |
|     self.else_branch = else_branch | |
| 
 | |
| 
 | |
| class RawCodeNode: | |
|   def __init__(self, token=None): | |
|     self.raw_code = token | |
| 
 | |
| 
 | |
| class LiteralDollarNode: | |
|   def __init__(self, token): | |
|     self.token = token | |
| 
 | |
| 
 | |
| class ExpNode: | |
|   def __init__(self, token, python_exp): | |
|     self.token = token | |
|     self.python_exp = python_exp | |
| 
 | |
| 
 | |
| def PopFront(a_list): | |
|   head = a_list[0] | |
|   a_list[:1] = [] | |
|   return head | |
| 
 | |
| 
 | |
| def PushFront(a_list, elem): | |
|   a_list[:0] = [elem] | |
| 
 | |
| 
 | |
| def PopToken(a_list, token_type=None): | |
|   token = PopFront(a_list) | |
|   if token_type is not None and token.token_type != token_type: | |
|     print 'ERROR: %s expected at %s' % (token_type, token.start) | |
|     print 'ERROR: %s found instead' % (token,) | |
|     sys.exit(1) | |
| 
 | |
|   return token | |
| 
 | |
| 
 | |
| def PeekToken(a_list): | |
|   if not a_list: | |
|     return None | |
| 
 | |
|   return a_list[0] | |
| 
 | |
| 
 | |
| def ParseExpNode(token): | |
|   python_exp = re.sub(r'([_A-Za-z]\w*)', r'self.GetValue("\1")', token.value) | |
|   return ExpNode(token, python_exp) | |
| 
 | |
| 
 | |
| def ParseElseNode(tokens): | |
|   def Pop(token_type=None): | |
|     return PopToken(tokens, token_type) | |
| 
 | |
|   next = PeekToken(tokens) | |
|   if not next: | |
|     return None | |
|   if next.token_type == '$else': | |
|     Pop('$else') | |
|     Pop('[[') | |
|     code_node = ParseCodeNode(tokens) | |
|     Pop(']]') | |
|     return code_node | |
|   elif next.token_type == '$elif': | |
|     Pop('$elif') | |
|     exp = Pop('code') | |
|     Pop('[[') | |
|     code_node = ParseCodeNode(tokens) | |
|     Pop(']]') | |
|     inner_else_node = ParseElseNode(tokens) | |
|     return CodeNode([IfNode(ParseExpNode(exp), code_node, inner_else_node)]) | |
|   elif not next.value.strip(): | |
|     Pop('code') | |
|     return ParseElseNode(tokens) | |
|   else: | |
|     return None | |
| 
 | |
| 
 | |
| def ParseAtomicCodeNode(tokens): | |
|   def Pop(token_type=None): | |
|     return PopToken(tokens, token_type) | |
| 
 | |
|   head = PopFront(tokens) | |
|   t = head.token_type | |
|   if t == 'code': | |
|     return RawCodeNode(head) | |
|   elif t == '$var': | |
|     id_token = Pop('id') | |
|     Pop('=') | |
|     next = PeekToken(tokens) | |
|     if next.token_type == 'exp': | |
|       exp_token = Pop() | |
|       return VarNode(id_token, ParseExpNode(exp_token)) | |
|     Pop('[[') | |
|     code_node = ParseCodeNode(tokens) | |
|     Pop(']]') | |
|     return VarNode(id_token, code_node) | |
|   elif t == '$for': | |
|     id_token = Pop('id') | |
|     next_token = PeekToken(tokens) | |
|     if next_token.token_type == 'code': | |
|       sep_token = next_token | |
|       Pop('code') | |
|     else: | |
|       sep_token = None | |
|     Pop('[[') | |
|     code_node = ParseCodeNode(tokens) | |
|     Pop(']]') | |
|     return ForNode(id_token, sep_token, code_node) | |
|   elif t == '$if': | |
|     exp_token = Pop('code') | |
|     Pop('[[') | |
|     code_node = ParseCodeNode(tokens) | |
|     Pop(']]') | |
|     else_node = ParseElseNode(tokens) | |
|     return IfNode(ParseExpNode(exp_token), code_node, else_node) | |
|   elif t == '$range': | |
|     id_token = Pop('id') | |
|     exp1_token = Pop('exp') | |
|     Pop('..') | |
|     exp2_token = Pop('exp') | |
|     return RangeNode(id_token, ParseExpNode(exp1_token), | |
|                      ParseExpNode(exp2_token)) | |
|   elif t == '$id': | |
|     return ParseExpNode(Token(head.start + 1, head.end, head.value[1:], 'id')) | |
|   elif t == '$($)': | |
|     return LiteralDollarNode(head) | |
|   elif t == '$': | |
|     exp_token = Pop('exp') | |
|     return ParseExpNode(exp_token) | |
|   elif t == '[[': | |
|     code_node = ParseCodeNode(tokens) | |
|     Pop(']]') | |
|     return code_node | |
|   else: | |
|     PushFront(tokens, head) | |
|     return None | |
| 
 | |
| 
 | |
| def ParseCodeNode(tokens): | |
|   atomic_code_list = [] | |
|   while True: | |
|     if not tokens: | |
|       break | |
|     atomic_code_node = ParseAtomicCodeNode(tokens) | |
|     if atomic_code_node: | |
|       atomic_code_list.append(atomic_code_node) | |
|     else: | |
|       break | |
|   return CodeNode(atomic_code_list) | |
| 
 | |
| 
 | |
| def ParseToAST(pump_src_text): | |
|   """Convert the given Pump source text into an AST.""" | |
|   tokens = list(Tokenize(pump_src_text)) | |
|   code_node = ParseCodeNode(tokens) | |
|   return code_node | |
| 
 | |
| 
 | |
| class Env: | |
|   def __init__(self): | |
|     self.variables = [] | |
|     self.ranges = [] | |
| 
 | |
|   def Clone(self): | |
|     clone = Env() | |
|     clone.variables = self.variables[:] | |
|     clone.ranges = self.ranges[:] | |
|     return clone | |
| 
 | |
|   def PushVariable(self, var, value): | |
|     # If value looks like an int, store it as an int. | |
|     try: | |
|       int_value = int(value) | |
|       if ('%s' % int_value) == value: | |
|         value = int_value | |
|     except Exception: | |
|       pass | |
|     self.variables[:0] = [(var, value)] | |
| 
 | |
|   def PopVariable(self): | |
|     self.variables[:1] = [] | |
| 
 | |
|   def PushRange(self, var, lower, upper): | |
|     self.ranges[:0] = [(var, lower, upper)] | |
| 
 | |
|   def PopRange(self): | |
|     self.ranges[:1] = [] | |
| 
 | |
|   def GetValue(self, identifier): | |
|     for (var, value) in self.variables: | |
|       if identifier == var: | |
|         return value | |
| 
 | |
|     print 'ERROR: meta variable %s is undefined.' % (identifier,) | |
|     sys.exit(1) | |
| 
 | |
|   def EvalExp(self, exp): | |
|     try: | |
|       result = eval(exp.python_exp) | |
|     except Exception, e: | |
|       print 'ERROR: caught exception %s: %s' % (e.__class__.__name__, e) | |
|       print ('ERROR: failed to evaluate meta expression %s at %s' % | |
|              (exp.python_exp, exp.token.start)) | |
|       sys.exit(1) | |
|     return result | |
| 
 | |
|   def GetRange(self, identifier): | |
|     for (var, lower, upper) in self.ranges: | |
|       if identifier == var: | |
|         return (lower, upper) | |
| 
 | |
|     print 'ERROR: range %s is undefined.' % (identifier,) | |
|     sys.exit(1) | |
| 
 | |
| 
 | |
| class Output: | |
|   def __init__(self): | |
|     self.string = '' | |
| 
 | |
|   def GetLastLine(self): | |
|     index = self.string.rfind('\n') | |
|     if index < 0: | |
|       return '' | |
| 
 | |
|     return self.string[index + 1:] | |
| 
 | |
|   def Append(self, s): | |
|     self.string += s | |
| 
 | |
| 
 | |
| def RunAtomicCode(env, node, output): | |
|   if isinstance(node, VarNode): | |
|     identifier = node.identifier.value.strip() | |
|     result = Output() | |
|     RunAtomicCode(env.Clone(), node.atomic_code, result) | |
|     value = result.string | |
|     env.PushVariable(identifier, value) | |
|   elif isinstance(node, RangeNode): | |
|     identifier = node.identifier.value.strip() | |
|     lower = int(env.EvalExp(node.exp1)) | |
|     upper = int(env.EvalExp(node.exp2)) | |
|     env.PushRange(identifier, lower, upper) | |
|   elif isinstance(node, ForNode): | |
|     identifier = node.identifier.value.strip() | |
|     if node.sep is None: | |
|       sep = '' | |
|     else: | |
|       sep = node.sep.value | |
|     (lower, upper) = env.GetRange(identifier) | |
|     for i in range(lower, upper + 1): | |
|       new_env = env.Clone() | |
|       new_env.PushVariable(identifier, i) | |
|       RunCode(new_env, node.code, output) | |
|       if i != upper: | |
|         output.Append(sep) | |
|   elif isinstance(node, RawCodeNode): | |
|     output.Append(node.raw_code.value) | |
|   elif isinstance(node, IfNode): | |
|     cond = env.EvalExp(node.exp) | |
|     if cond: | |
|       RunCode(env.Clone(), node.then_branch, output) | |
|     elif node.else_branch is not None: | |
|       RunCode(env.Clone(), node.else_branch, output) | |
|   elif isinstance(node, ExpNode): | |
|     value = env.EvalExp(node) | |
|     output.Append('%s' % (value,)) | |
|   elif isinstance(node, LiteralDollarNode): | |
|     output.Append('$') | |
|   elif isinstance(node, CodeNode): | |
|     RunCode(env.Clone(), node, output) | |
|   else: | |
|     print 'BAD' | |
|     print node | |
|     sys.exit(1) | |
| 
 | |
| 
 | |
| def RunCode(env, code_node, output): | |
|   for atomic_code in code_node.atomic_code: | |
|     RunAtomicCode(env, atomic_code, output) | |
| 
 | |
| 
 | |
| def IsSingleLineComment(cur_line): | |
|   return '//' in cur_line | |
| 
 | |
| 
 | |
| def IsInPreprocessorDirective(prev_lines, cur_line): | |
|   if cur_line.lstrip().startswith('#'): | |
|     return True | |
|   return prev_lines and prev_lines[-1].endswith('\\') | |
| 
 | |
| 
 | |
| def WrapComment(line, output): | |
|   loc = line.find('//') | |
|   before_comment = line[:loc].rstrip() | |
|   if before_comment == '': | |
|     indent = loc | |
|   else: | |
|     output.append(before_comment) | |
|     indent = len(before_comment) - len(before_comment.lstrip()) | |
|   prefix = indent*' ' + '// ' | |
|   max_len = 80 - len(prefix) | |
|   comment = line[loc + 2:].strip() | |
|   segs = [seg for seg in re.split(r'(\w+\W*)', comment) if seg != ''] | |
|   cur_line = '' | |
|   for seg in segs: | |
|     if len((cur_line + seg).rstrip()) < max_len: | |
|       cur_line += seg | |
|     else: | |
|       if cur_line.strip() != '': | |
|         output.append(prefix + cur_line.rstrip()) | |
|       cur_line = seg.lstrip() | |
|   if cur_line.strip() != '': | |
|     output.append(prefix + cur_line.strip()) | |
| 
 | |
| 
 | |
| def WrapCode(line, line_concat, output): | |
|   indent = len(line) - len(line.lstrip()) | |
|   prefix = indent*' '  # Prefix of the current line | |
|   max_len = 80 - indent - len(line_concat)  # Maximum length of the current line | |
|   new_prefix = prefix + 4*' '  # Prefix of a continuation line | |
|   new_max_len = max_len - 4  # Maximum length of a continuation line | |
|   # Prefers to wrap a line after a ',' or ';'. | |
|   segs = [seg for seg in re.split(r'([^,;]+[,;]?)', line.strip()) if seg != ''] | |
|   cur_line = ''  # The current line without leading spaces. | |
|   for seg in segs: | |
|     # If the line is still too long, wrap at a space. | |
|     while cur_line == '' and len(seg.strip()) > max_len: | |
|       seg = seg.lstrip() | |
|       split_at = seg.rfind(' ', 0, max_len) | |
|       output.append(prefix + seg[:split_at].strip() + line_concat) | |
|       seg = seg[split_at + 1:] | |
|       prefix = new_prefix | |
|       max_len = new_max_len | |
| 
 | |
|     if len((cur_line + seg).rstrip()) < max_len: | |
|       cur_line = (cur_line + seg).lstrip() | |
|     else: | |
|       output.append(prefix + cur_line.rstrip() + line_concat) | |
|       prefix = new_prefix | |
|       max_len = new_max_len | |
|       cur_line = seg.lstrip() | |
|   if cur_line.strip() != '': | |
|     output.append(prefix + cur_line.strip()) | |
| 
 | |
| 
 | |
| def WrapPreprocessorDirective(line, output): | |
|   WrapCode(line, ' \\', output) | |
| 
 | |
| 
 | |
| def WrapPlainCode(line, output): | |
|   WrapCode(line, '', output) | |
| 
 | |
| 
 | |
| def IsMultiLineIWYUPragma(line): | |
|   return re.search(r'/\* IWYU pragma: ', line) | |
| 
 | |
| 
 | |
| def IsHeaderGuardIncludeOrOneLineIWYUPragma(line): | |
|   return (re.match(r'^#(ifndef|define|endif\s*//)\s*[\w_]+\s*$', line) or | |
|           re.match(r'^#include\s', line) or | |
|           # Don't break IWYU pragmas, either; that causes iwyu.py problems. | |
|           re.search(r'// IWYU pragma: ', line)) | |
| 
 | |
| 
 | |
| def WrapLongLine(line, output): | |
|   line = line.rstrip() | |
|   if len(line) <= 80: | |
|     output.append(line) | |
|   elif IsSingleLineComment(line): | |
|     if IsHeaderGuardIncludeOrOneLineIWYUPragma(line): | |
|       # The style guide made an exception to allow long header guard lines, | |
|       # includes and IWYU pragmas. | |
|       output.append(line) | |
|     else: | |
|       WrapComment(line, output) | |
|   elif IsInPreprocessorDirective(output, line): | |
|     if IsHeaderGuardIncludeOrOneLineIWYUPragma(line): | |
|       # The style guide made an exception to allow long header guard lines, | |
|       # includes and IWYU pragmas. | |
|       output.append(line) | |
|     else: | |
|       WrapPreprocessorDirective(line, output) | |
|   elif IsMultiLineIWYUPragma(line): | |
|     output.append(line) | |
|   else: | |
|     WrapPlainCode(line, output) | |
| 
 | |
| 
 | |
| def BeautifyCode(string): | |
|   lines = string.splitlines() | |
|   output = [] | |
|   for line in lines: | |
|     WrapLongLine(line, output) | |
|   output2 = [line.rstrip() for line in output] | |
|   return '\n'.join(output2) + '\n' | |
| 
 | |
| 
 | |
| def ConvertFromPumpSource(src_text): | |
|   """Return the text generated from the given Pump source text.""" | |
|   ast = ParseToAST(StripMetaComments(src_text)) | |
|   output = Output() | |
|   RunCode(Env(), ast, output) | |
|   return BeautifyCode(output.string) | |
| 
 | |
| 
 | |
| def main(argv): | |
|   if len(argv) == 1: | |
|     print __doc__ | |
|     sys.exit(1) | |
| 
 | |
|   file_path = argv[-1] | |
|   output_str = ConvertFromPumpSource(file(file_path, 'r').read()) | |
|   if file_path.endswith('.pump'): | |
|     output_file_path = file_path[:-5] | |
|   else: | |
|     output_file_path = '-' | |
|   if output_file_path == '-': | |
|     print output_str, | |
|   else: | |
|     output_file = file(output_file_path, 'w') | |
|     output_file.write('// This file was GENERATED by command:\n') | |
|     output_file.write('//     %s %s\n' % | |
|                       (os.path.basename(__file__), os.path.basename(file_path))) | |
|     output_file.write('// DO NOT EDIT BY HAND!!!\n\n') | |
|     output_file.write(output_str) | |
|     output_file.close() | |
| 
 | |
| 
 | |
| if __name__ == '__main__': | |
|   main(sys.argv)
 |