• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1#!/usr/bin/env python3
2
3#
4# Copyright (C) 2018 The Android Open Source Project
5#
6# Licensed under the Apache License, Version 2.0 (the "License");
7# you may not use this file except in compliance with the License.
8# You may obtain a copy of the License at
9#
10#      http://www.apache.org/licenses/LICENSE-2.0
11#
12# Unless required by applicable law or agreed to in writing, software
13# distributed under the License is distributed on an "AS IS" BASIS,
14# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
15# See the License for the specific language governing permissions and
16# limitations under the License.
17#
18
19"""This module implements a Android.bp parser."""
20
21import collections
22import glob
23import itertools
24import os
25import re
26import sys
27
28
29#------------------------------------------------------------------------------
30# Python 2 compatibility
31#------------------------------------------------------------------------------
32
33if sys.version_info >= (3,):
34    py3_chr = chr  # pylint: disable=invalid-name
35else:
36    def py3_chr(codepoint):
37        """Convert an integer character codepoint into a utf-8 string."""
38        return unichr(codepoint).encode('utf-8')
39
40try:
41    from enum import Enum
42except ImportError:
43    class _Enum(object):  # pylint: disable=too-few-public-methods
44        """A name-value pair for each enumeration."""
45
46        __slot__ = ('name', 'value')
47
48
49        def __init__(self, name, value):
50            """Create a name-value pair."""
51            self.name = name
52            self.value = value
53
54
55        def __repr__(self):
56            """Return the name of the enumeration."""
57            return self.name
58
59
60    class _EnumMeta(type):  # pylint: disable=too-few-public-methods
61        """Metaclass for Enum base class."""
62
63        def __new__(mcs, name, bases, attrs):
64            """Collects enumerations from attributes of the derived classes."""
65            enums = []
66            new_attrs = {'_enums': enums}
67
68            for key, value in attrs.iteritems():
69                if key.startswith('_'):
70                    new_attrs[key] = value
71                else:
72                    item = _Enum(key, value)
73                    enums.append(item)
74                    new_attrs[key] = item
75
76            return type.__new__(mcs, name, bases, new_attrs)
77
78
79        def __iter__(cls):
80            """Iterate the list of enumerations."""
81            return iter(cls._enums)
82
83
84    class Enum(object):  # pylint: disable=too-few-public-methods
85        """Enum base class."""
86        __metaclass__ = _EnumMeta
87
88
89#------------------------------------------------------------------------------
90# Lexer
91#------------------------------------------------------------------------------
92
93class Token(Enum):  # pylint: disable=too-few-public-methods
94    """Token enumerations."""
95
96    EOF = 0
97
98    IDENT = 1
99    LPAREN = 2
100    RPAREN = 3
101    LBRACKET = 4
102    RBRACKET = 5
103    LBRACE = 6
104    RBRACE = 7
105    COLON = 8
106    ASSIGN = 9
107    ASSIGNPLUS = 10
108    PLUS = 11
109    COMMA = 12
110    STRING = 13
111    INTEGER = 14
112
113    COMMENT = 15
114    SPACE = 16
115
116
117class LexerError(ValueError):
118    """Lexer error exception class."""
119
120    def __init__(self, buf, pos, message):
121        """Create a lexer error exception object."""
122        super(LexerError, self).__init__(message)
123        self.message = message
124        self.line, self.column = Lexer.compute_line_column(buf, pos)
125
126
127    def __str__(self):
128        """Convert lexer error to string representation."""
129        return 'LexerError: {}:{}: {}'.format(
130            self.line, self.column, self.message)
131
132
133class Lexer(object):
134    """Lexer to tokenize the input string."""
135
136    def __init__(self, buf, offset=0, path=None):
137        """Tokenize the source code in buf starting from offset.
138
139        Args:
140            buf (string) The source code to be tokenized.
141            offset (int) The position to start.
142        """
143
144        self.buf = buf
145
146        self.start = None
147        self.end = offset
148        self.token = None
149        self.literal = None
150        self.path = path
151
152        self._next()
153
154
155    def consume(self, *tokens):
156        """Consume one or more token."""
157
158        for token in tokens:
159            if token == self.token:
160                self._next()
161            else:
162                raise LexerError(self.buf, self.start,
163                                 'unexpected token ' + self.token.name)
164
165
166    def _next(self):
167        """Read next non-comment non-space token."""
168
169        buf_len = len(self.buf)
170        while self.end < buf_len:
171            self.start = self.end
172            self.token, self.end, self.literal = self.lex(self.buf, self.start)
173            if self.token != Token.SPACE and self.token != Token.COMMENT:
174                return
175
176        self.start = self.end
177        self.token = Token.EOF
178        self.literal = None
179
180
181    @staticmethod
182    def compute_line_column(buf, pos):
183        """Compute the line number and the column number of a given position in
184        the buffer."""
185
186        prior = buf[0:pos]
187        newline_pos = prior.rfind('\n')
188        if newline_pos == -1:
189            return (1, pos + 1)
190        return (prior.count('\n') + 1, pos - newline_pos)
191
192
193    UNICODE_CHARS_PATTERN = re.compile('[^\\\\\\n"]+')
194
195
196    ESCAPE_CHAR_TABLE = {
197        'a': '\a', 'b': '\b', 'f': '\f', 'n': '\n', 'r': '\r', 't': '\t',
198        'v': '\v', '\\': '\\', '\'': '\'', '\"': '\"',
199    }
200
201
202    OCT_TABLE = {str(i) for i in range(8)}
203
204
205    @staticmethod
206    def decode_oct(buf, offset, start, end):
207        """Read characters from buf[start:end] and interpret them as an octal
208        integer."""
209
210        if end > len(buf):
211            raise LexerError(buf, offset, 'bad octal escape sequence')
212        try:
213            codepoint = int(buf[start:end], 8)
214        except ValueError:
215            raise LexerError(buf, offset, 'bad octal escape sequence')
216        if codepoint > 0xff:
217            raise LexerError(buf, offset, 'bad octal escape sequence')
218        return codepoint
219
220
221    @staticmethod
222    def decode_hex(buf, offset, start, end):
223        """Read characters from buf[start:end] and interpret them as a
224        hexadecimal integer."""
225
226        if end > len(buf):
227            raise LexerError(buf, offset, 'bad hex escape sequence')
228        try:
229            return int(buf[start:end], 16)
230        except ValueError:
231            raise LexerError(buf, offset, 'bad hex escape sequence')
232
233
234    @classmethod
235    def lex_interpreted_string(cls, buf, offset):
236        """Tokenize a golang interpreted string.
237
238        Args:
239            buf (str)    The source code buffer.
240            offset (int) The position to find a golang interpreted string
241                         literal.
242
243        Returns:
244            A tuple with the end of matched buffer and the interpreted string
245            literal.
246        """
247
248        buf_len = len(buf)
249        pos = offset + 1
250        literal = ''
251        while pos < buf_len:
252            # Match unicode characters
253            match = cls.UNICODE_CHARS_PATTERN.match(buf, pos)
254            if match:
255                literal += match.group(0)
256                pos = match.end()
257            # Read the next character
258            try:
259                char = buf[pos]
260            except IndexError:
261                raise LexerError(buf, pos,
262                                 'unclosed interpreted string literal')
263            if char == '\\':
264                # Escape sequences
265                try:
266                    char = buf[pos + 1]
267                except IndexError:
268                    raise LexerError(buf, pos, 'bad escape sequence')
269                if char in cls.OCT_TABLE:
270                    literal += chr(cls.decode_oct(buf, pos, pos + 1, pos + 4))
271                    pos += 4
272                elif char == 'x':
273                    literal += chr(cls.decode_hex(buf, pos, pos + 2, pos + 4))
274                    pos += 4
275                elif char == 'u':
276                    literal += py3_chr(
277                        cls.decode_hex(buf, pos, pos + 2, pos + 6))
278                    pos += 6
279                elif char == 'U':
280                    literal += py3_chr(
281                        cls.decode_hex(buf, pos, pos + 2, pos + 10))
282                    pos += 10
283                else:
284                    try:
285                        literal += cls.ESCAPE_CHAR_TABLE[char]
286                        pos += 2
287                    except KeyError:
288                        raise LexerError(buf, pos, 'bad escape sequence')
289                continue
290            if char == '"':
291                # End of string literal
292                return (pos + 1, literal)
293            raise LexerError(buf, pos, 'unclosed interpreted string literal')
294
295
296    @classmethod
297    def lex_string(cls, buf, offset):
298        """Tokenize a golang string literal.
299
300        Args:
301            buf (str)    The source code buffer.
302            offset (int) The position to find a golang string literal.
303
304        Returns:
305            A tuple with the end of matched buffer and the interpreted string
306            literal.
307        """
308
309        char = buf[offset]
310        if char == '`':
311            try:
312                end = buf.index('`', offset + 1)
313                return (end + 1, buf[offset + 1 : end])
314            except ValueError:
315                raise LexerError(buf, len(buf), 'unclosed raw string literal')
316        if char == '"':
317            return cls.lex_interpreted_string(buf, offset)
318        raise LexerError(buf, offset, 'no string literal start character')
319
320
321    LEXER_PATTERNS = (
322        (Token.IDENT, '[A-Za-z_][0-9A-Za-z_]*'),
323        (Token.LPAREN, '\\('),
324        (Token.RPAREN, '\\)'),
325        (Token.LBRACKET, '\\['),
326        (Token.RBRACKET, '\\]'),
327        (Token.LBRACE, '\\{'),
328        (Token.RBRACE, '\\}'),
329        (Token.COLON, ':'),
330        (Token.ASSIGN, '='),
331        (Token.ASSIGNPLUS, '\\+='),
332        (Token.PLUS, '\\+'),
333        (Token.COMMA, ','),
334        (Token.STRING, '["`]'),
335        (Token.INTEGER, '-{0,1}[0-9]+'),
336
337        (Token.COMMENT,
338         '/(?:(?:/[^\\n]*)|(?:\\*(?:(?:[^*]*)|(?:\\*+[^/*]))*\\*+/))'),
339        (Token.SPACE, '\\s+'),
340    )
341
342
343    LEXER_MATCHER = re.compile('|'.join(
344        '(' + pattern + ')' for _, pattern in LEXER_PATTERNS))
345
346
347    @classmethod
348    def lex(cls, buf, offset):
349        """Tokenize a token from buf[offset].
350
351        Args:
352            buf (string) The source code buffer.
353            offset (int) The position to find and tokenize a token.
354
355        Return:
356            A tuple with three elements.  The first element is the token id.
357            The second element is the end of the token.  The third element is
358            the value for strings or identifiers.
359        """
360
361        match = cls.LEXER_MATCHER.match(buf, offset)
362        if not match:
363            raise LexerError(buf, offset, 'unknown token')
364        token = cls.LEXER_PATTERNS[match.lastindex - 1][0]
365
366        if token == Token.STRING:
367            end, literal = cls.lex_string(buf, offset)
368        else:
369            end = match.end()
370            if token in {Token.IDENT, Token.INTEGER}:
371                literal = buf[offset:end]
372            else:
373                literal = None
374
375        return (token, end, literal)
376
377
378#------------------------------------------------------------------------------
379# AST
380#------------------------------------------------------------------------------
381
382class Expr(object):  # pylint: disable=too-few-public-methods
383    """Base class for all expressions."""
384
385    def eval(self, env):
386        """Evaluate the expression under an environment."""
387        raise NotImplementedError()
388
389
390class String(Expr, str):
391    """String constant literal."""
392
393    def eval(self, env):
394        """Evaluate the string expression under an environment."""
395        return self
396
397
398class Bool(Expr):  # pylint: disable=too-few-public-methods
399    """Boolean constant literal."""
400
401    __slots__ = ('value',)
402
403
404    def __init__(self, value):
405        """Create a boolean constant literal."""
406        self.value = value
407
408
409    def __repr__(self):
410        """Convert a boolean constant literal to string representation."""
411        return repr(self.value)
412
413
414    def __bool__(self):
415        """Convert boolean constant literal to Python bool type."""
416        return self.value
417
418    __nonzero__ = __bool__
419
420
421    def __eq__(self, rhs):
422        """Compare whether two instances are equal."""
423        return self.value == rhs.value
424
425
426    def __hash__(self):
427        """Compute the hashed value."""
428        return hash(self.value)
429
430
431    def eval(self, env):
432        """Evaluate the boolean expression under an environment."""
433        return self
434
435
436class Integer(Expr):  # pylint: disable=too-few-public-methods
437    """Integer constant literal."""
438
439    __slots__ = ('value',)
440
441
442    def __init__(self, value):
443        """Create an integer constant literal."""
444        self.value = value
445
446
447    def __repr__(self):
448        """Convert an integer constant literal to string representation."""
449        return repr(self.value)
450
451
452    def __bool__(self):
453        """Convert an integer constant literal to Python bool type."""
454        return bool(self.value)
455
456    __nonzero__ = __bool__
457
458
459    def __int__(self):
460        """Convert an integer constant literal to Python int type."""
461        return self.value
462
463
464    def __eq__(self, rhs):
465        """Compare whether two instances are equal."""
466        return self.value == rhs.value
467
468
469    def __hash__(self):
470        """Compute the hashed value."""
471        return hash(self.value)
472
473
474    def eval(self, env):
475        """Evaluate the integer expression under an environment."""
476        return self
477
478
479class VarRef(Expr):  # pylint: disable=too-few-public-methods
480    """A reference to a variable."""
481
482    def __init__(self, name, value):
483        """Create a variable reference with a name and the value under static
484        scoping."""
485        self.name = name
486        self.value = value
487
488
489    def __repr__(self):
490        """Convert a variable reference to string representation."""
491        return self.name
492
493
494    def eval(self, env):
495        """Evaluate the identifier under an environment."""
496        if self.value is None:
497            return env[self.name].eval(env)
498        return self.value.eval(env)
499
500
501class List(Expr, list):
502    """List expression."""
503
504    def eval(self, env):
505        """Evaluate list elements under an environment."""
506        return List(item.eval(env) for item in self)
507
508
509class Dict(Expr, collections.OrderedDict):
510    """Dictionary expression."""
511
512    def __repr__(self):
513        attrs = ', '.join(key + ': ' + repr(value)
514                          for key, value in self.items())
515        return '{' + attrs + '}'
516
517
518    def eval(self, env):
519        """Evaluate dictionary values under an environment."""
520        return Dict((key, value.eval(env)) for key, value in self.items())
521
522
523class Concat(Expr):  # pylint: disable=too-few-public-methods
524    """List/string/integer plus operator."""
525
526    __slots__ = ('lhs', 'rhs')
527
528
529    def __init__(self, lhs, rhs):
530        """Create a list/string/integer plus expression."""
531        self.lhs = lhs
532        self.rhs = rhs
533
534
535    def __repr__(self):
536        return '(' + repr(self.lhs) + ' + ' + repr(self.rhs) + ')'
537
538
539    def eval(self, env):
540        """Evaluate list/string/integer plus operator under an environment."""
541        lhs = self.lhs.eval(env)
542        rhs = self.rhs.eval(env)
543        if isinstance(lhs, List) and isinstance(rhs, List):
544            return List(itertools.chain(lhs, rhs))
545        if isinstance(lhs, String) and isinstance(rhs, String):
546            return String(lhs + rhs)
547        if isinstance(lhs, Integer) and isinstance(rhs, Integer):
548            return Integer(int(lhs) + int(rhs))
549        raise TypeError('bad plus operands')
550
551
552#------------------------------------------------------------------------------
553# Parser
554#------------------------------------------------------------------------------
555
556class ParseError(ValueError):
557    """Parser error exception class."""
558
559    def __init__(self, lexer, message):
560        """Create a parser error exception object."""
561        super(ParseError, self).__init__(message)
562        self.message = message
563        self.line, self.column = \
564            Lexer.compute_line_column(lexer.buf, lexer.start)
565
566
567    def __str__(self):
568        """Convert parser error to string representation."""
569        return 'ParseError: {}:{}: {}'.format(
570            self.line, self.column, self.message)
571
572
573class Parser(object):
574    """Parser to parse Android.bp files."""
575
576    def __init__(self, lexer, inherited_env=None):
577        """Initialize the parser with the lexer."""
578        self.lexer = lexer
579
580        self.var_defs = []
581        self.vars = {} if inherited_env is None else dict(inherited_env)
582        self.modules = []
583
584
585    def parse(self):
586        """Parse AST from tokens."""
587        lexer = self.lexer
588        while lexer.token != Token.EOF:
589            if lexer.token == Token.IDENT:
590                ident = self.parse_ident_lvalue()
591                if lexer.token in {Token.ASSIGN, Token.ASSIGNPLUS}:
592                    self.parse_assign(ident, lexer.token)
593                elif lexer.token in {Token.LBRACE, Token.LPAREN}:
594                    self.parse_module_definition(ident)
595                else:
596                    raise ParseError(lexer,
597                                     'unexpected token ' + lexer.token.name)
598            else:
599                raise ParseError(lexer, 'unexpected token ' + lexer.token.name)
600        lexer.consume(Token.EOF)
601
602
603    def create_var_ref(self, name):
604        """Create a variable reference."""
605        return VarRef(name, self.vars.get(name))
606
607
608    def define_var(self, name, value):
609        """Define a variable."""
610        self.var_defs.append((name, value))
611        self.vars[name] = value
612
613
614    def parse_assign(self, ident, assign_token):
615        """Parse an assignment statement."""
616        lexer = self.lexer
617        lexer.consume(assign_token)
618        value = self.parse_expression()
619        if assign_token == Token.ASSIGNPLUS:
620            value = Concat(self.create_var_ref(ident), value)
621        self.define_var(ident, value)
622
623
624    def parse_module_definition(self, module_ident):
625        """Parse a module definition."""
626        properties = self.parse_dict()
627        properties['_path'] = String(self.lexer.path)
628        self.modules.append((module_ident, properties))
629
630
631    def parse_ident_lvalue(self):
632        """Parse an identifier as an l-value."""
633        ident = self.lexer.literal
634        self.lexer.consume(Token.IDENT)
635        return ident
636
637
638    def parse_ident_rvalue(self):
639        """Parse an identifier as a r-value.
640
641        Returns:
642            Returns VarRef if the literal is not 'true' nor 'false'.
643
644            Returns Bool(true/false) if the literal is either 'true' or 'false'.
645        """
646        lexer = self.lexer
647        if lexer.literal in {'true', 'false'}:
648            result = Bool(lexer.literal == 'true')
649        else:
650            result = self.create_var_ref(lexer.literal)
651        lexer.consume(Token.IDENT)
652        return result
653
654
655    def parse_string(self):
656        """Parse a string."""
657        lexer = self.lexer
658        string = String(lexer.literal)
659        lexer.consume(Token.STRING)
660        return string
661
662
663    def parse_integer(self):
664        """Parse an integer."""
665        lexer = self.lexer
666        integer = Integer(int(lexer.literal))
667        lexer.consume(Token.INTEGER)
668        return integer
669
670
671    def parse_operand(self):
672        """Parse an operand."""
673        lexer = self.lexer
674        token = lexer.token
675        if token == Token.STRING:
676            return self.parse_string()
677        if token == Token.IDENT:
678            return self.parse_ident_rvalue()
679        if token == Token.INTEGER:
680            return self.parse_integer()
681        if token == Token.LBRACKET:
682            return self.parse_list()
683        if token == Token.LBRACE:
684            return self.parse_dict()
685        if token == Token.LPAREN:
686            lexer.consume(Token.LPAREN)
687            operand = self.parse_expression()
688            lexer.consume(Token.RPAREN)
689            return operand
690        raise ParseError(lexer, 'unexpected token ' + token.name)
691
692
693    def parse_expression(self):
694        """Parse an expression."""
695        lexer = self.lexer
696        expr = self.parse_operand()
697        while lexer.token == Token.PLUS:
698            lexer.consume(Token.PLUS)
699            expr = Concat(expr, self.parse_operand())
700        return expr
701
702
703    def parse_list(self):
704        """Parse a list."""
705        result = List()
706        lexer = self.lexer
707        lexer.consume(Token.LBRACKET)
708        while lexer.token != Token.RBRACKET:
709            result.append(self.parse_expression())
710            if lexer.token == Token.COMMA:
711                lexer.consume(Token.COMMA)
712        lexer.consume(Token.RBRACKET)
713        return result
714
715
716    def parse_dict(self):
717        """Parse a dict."""
718        result = Dict()
719        lexer = self.lexer
720
721        is_func_syntax = lexer.token == Token.LPAREN
722        if is_func_syntax:
723            lexer.consume(Token.LPAREN)
724        else:
725            lexer.consume(Token.LBRACE)
726
727        while lexer.token != Token.RBRACE and lexer.token != Token.RPAREN:
728            if lexer.token != Token.IDENT:
729                raise ParseError(lexer, 'unexpected token ' + lexer.token.name)
730            key = self.parse_ident_lvalue()
731
732            if lexer.token == Token.ASSIGN:
733                lexer.consume(Token.ASSIGN)
734            else:
735                lexer.consume(Token.COLON)
736
737            value = self.parse_expression()
738            result[key] = value
739
740            if lexer.token == Token.COMMA:
741                lexer.consume(Token.COMMA)
742
743        if is_func_syntax:
744            lexer.consume(Token.RPAREN)
745        else:
746            lexer.consume(Token.RBRACE)
747
748        return result
749
750
751class RecursiveParser(object):
752    """This is a recursive parser which will parse blueprint files
753    recursively."""
754
755
756    # Default Blueprint file name
757    _DEFAULT_SUB_NAME = 'Android.bp'
758
759
760    def __init__(self):
761        """Initialize a recursive parser."""
762        self.visited = set()
763        self.modules = []
764
765
766    @staticmethod
767    def glob_sub_files(pattern, sub_file_name):
768        """List the sub file paths that match with the pattern with
769        wildcards."""
770
771        for path in glob.glob(pattern):
772            if os.path.isfile(path):
773                if os.path.basename(path) == sub_file_name:
774                    yield path
775            else:
776                sub_file_path = os.path.join(path, sub_file_name)
777                if os.path.isfile(sub_file_path):
778                    yield sub_file_path
779
780
781    @classmethod
782    def find_sub_files_from_env(cls, rootdir, env, use_subdirs,
783                                default_sub_name=_DEFAULT_SUB_NAME):
784        """Find the sub files from the names specified in build, subdirs, and
785        optional_subdirs."""
786
787        subs = []
788
789        if 'build' in env:
790            subs.extend(os.path.join(rootdir, filename)
791                        for filename in env['build'].eval(env))
792        if use_subdirs:
793            sub_name = env['subname'] if 'subname' in env else default_sub_name
794
795            if 'subdirs' in env:
796                for path in env['subdirs'].eval(env):
797                    subs.extend(cls.glob_sub_files(os.path.join(rootdir, path),
798                                                   sub_name))
799            if 'optional_subdirs' in env:
800                for path in env['optional_subdirs'].eval(env):
801                    subs.extend(cls.glob_sub_files(os.path.join(rootdir, path),
802                                                   sub_name))
803        return subs
804
805
806    @staticmethod
807    def _read_file(path, env):
808        """Read a blueprint file and return modules and the environment."""
809        with open(path, 'r') as bp_file:
810            content = bp_file.read()
811        parser = Parser(Lexer(content, path=path), env)
812        parser.parse()
813        return (parser.modules, parser.vars)
814
815
816    def _parse_file(self, path, env, evaluate):
817        """Parse a blueprint file and append to self.modules."""
818        modules, sub_env = self._read_file(path, env)
819        if evaluate:
820            modules = [(ident, attrs.eval(env)) for ident, attrs in modules]
821        self.modules += modules
822        return sub_env
823
824
825    def _parse_file_recursive(self, path, env, evaluate, use_subdirs):
826        """Parse a blueprint file and recursively."""
827
828        self.visited.add(path)
829
830        sub_env = self._parse_file(path, env, evaluate)
831
832        rootdir = os.path.dirname(path)
833
834        sub_file_paths = self.find_sub_files_from_env(rootdir, sub_env,
835                                                      use_subdirs)
836
837        sub_env.pop('build', None)
838        sub_env.pop('subdirs', None)
839        sub_env.pop('optional_subdirs', None)
840
841        for sub_file_path in sub_file_paths:
842            if sub_file_path not in self.visited:
843                self._parse_file_recursive(sub_file_path, sub_env, evaluate,
844                                           use_subdirs)
845        return sub_env
846
847
848    def _scan_and_parse_all_file_recursive(self, filename, path, env, evaluate):
849        """Scan all files with the specified name and parse them."""
850
851        rootdir = os.path.dirname(path)
852        assert rootdir, 'rootdir is empty but must be non-empty'
853
854        envs = [(rootdir, env)]
855        assert env is not None
856
857        # Scan directories for all blueprint files
858        for basedir, dirnames, filenames in os.walk(rootdir):
859            # Drop irrelevant environments
860            while not basedir.startswith(envs[-1][0]):
861                envs.pop()
862
863            # Filter sub directories
864            new_dirnames = []
865            for name in dirnames:
866                if name in {'.git', '.repo'}:
867                    continue
868                if basedir == rootdir and name == 'out':
869                    continue
870                new_dirnames.append(name)
871            dirnames[:] = new_dirnames
872
873            # Parse blueprint files
874            if filename in filenames:
875                try:
876                    path = os.path.join(basedir, filename)
877                    sys.stdout.flush()
878                    sub_env = self._parse_file_recursive(path, envs[-1][1],
879                                                         evaluate, False)
880                    assert sub_env is not None
881                    envs.append((basedir, sub_env))
882                except IOError:
883                    pass
884
885
886    def parse_file(self, path, env=None, evaluate=True,
887                   default_sub_name=_DEFAULT_SUB_NAME):
888        """Parse blueprint files recursively."""
889
890        if env is None:
891            env = {}
892
893        path = os.path.abspath(path)
894
895        sub_env = self._read_file(path, env)[1]
896
897        if 'subdirs' in sub_env or 'optional_subdirs' in sub_env:
898            self._parse_file_recursive(path, env, evaluate, True)
899        else:
900            self._scan_and_parse_all_file_recursive(
901                default_sub_name, path, env, evaluate)
902
903
904#------------------------------------------------------------------------------
905# Transformation
906#------------------------------------------------------------------------------
907
908def _build_named_modules_dict(modules):
909    """Build a name-to-module dict."""
910    named_modules = {}
911    for i, (ident, attrs) in enumerate(modules):
912        name = attrs.get('name')
913        if name is not None:
914            named_modules[name] = [ident, attrs, i]
915    return named_modules
916
917
918def _po_sorted_modules(modules, named_modules):
919    """Sort modules in post order."""
920    modules = [(ident, attrs, i) for i, (ident, attrs) in enumerate(modules)]
921
922    # Build module dependency graph.
923    edges = {}
924    for ident, attrs, module_id in modules:
925        defaults = attrs.get('defaults')
926        if defaults:
927            edges[module_id] = set(
928                named_modules[default][2] for default in defaults)
929
930    # Traverse module graph in post order.
931    post_order = []
932    visited = set()
933
934    def _traverse(module_id):
935        visited.add(module_id)
936        for next_module_id in edges.get(module_id, []):
937            if next_module_id not in visited:
938                _traverse(next_module_id)
939        post_order.append(modules[module_id])
940
941    for module_id in range(len(modules)):
942        if module_id not in visited:
943            _traverse(module_id)
944
945    return post_order
946
947
948def evaluate_default(attrs, default_attrs):
949    """Add default attributes if the keys do not exist."""
950    for key, value in default_attrs.items():
951        if key not in attrs:
952            attrs[key] = value
953        else:
954            attrs_value = attrs[key]
955            if isinstance(value, Dict) and isinstance(attrs_value, Dict):
956                attrs[key] = evaluate_default(attrs_value, value)
957    return attrs
958
959
960def evaluate_defaults(modules):
961    """Add default attributes to all modules if the keys do not exist."""
962    named_modules = _build_named_modules_dict(modules)
963    for ident, attrs, i in _po_sorted_modules(modules, named_modules):
964        for default in attrs.get('defaults', []):
965            attrs = evaluate_default(attrs, named_modules[default][1])
966        modules[i] = (ident, attrs)
967    return modules
968
969
970def fill_module_namespaces(root_bp, modules):
971    """Collect soong_namespace definition and set a `_namespace` property to
972    each module definitions."""
973
974    # Collect all namespaces
975    rootdir = os.path.dirname(os.path.abspath(root_bp))
976    namespaces = {rootdir}
977    for ident, attrs in modules:
978        if ident == 'soong_namespace':
979            namespaces.add(os.path.dirname(attrs['_path']))
980
981    # Build a path matcher for module namespaces
982    namespaces = sorted(namespaces, reverse=True)
983    path_matcher = re.compile(
984        '|'.join('(' + re.escape(x) + '/.*)' for x in namespaces))
985
986    # Trim the root directory prefix
987    rootdir_prefix_len = len(rootdir) + 1
988    namespaces = [path[rootdir_prefix_len:] for path in namespaces]
989
990    # Fill in module namespaces
991    for ident, attrs in modules:
992        match = path_matcher.match(attrs['_path'])
993        attrs['_namespace'] = namespaces[match.lastindex - 1]
994
995    return modules
996