mirror of
https://git.yoctoproject.org/git/poky
synced 2026-01-01 13:58:04 +00:00
One of the current frustrations with the sstate checksums is that
code like base_contains('X', 'y',...) adds a full dependency on X
and varies depend even on whitespace changes in X.
This patch adds special handling of the contains functions to expand
the first parameter and check for the flag specified by the second
parameter (assuming its a string).
The result is then appended to the value of the variable with a "Set"
or "Unset" status. If the flag is added/removed, the stored variable
value changes and hence the checksum changes. No dependency on X
is added so it is free to change with regard to other flags or
whitespace.
This code is far from ideal, ideally we'd have properly typed variables
however it fixes a major annoyance of the current checksums and
is of enough value its worth adding in a stopgap solution. It shouldn't
significantly restrict any propely typed variable implementation in
future.
(Bitbake rev: ed2d0a22a80299de0cfd377999950cf4b26c512e)
Signed-off-by: Richard Purdie <richard.purdie@linuxfoundation.org>
Signed-off-by: Christopher Larson <chris_larson@mentor.com>
Signed-off-by: Richard Purdie <richard.purdie@linuxfoundation.org>
326 lines
12 KiB
Python
326 lines
12 KiB
Python
import ast
|
|
import codegen
|
|
import logging
|
|
import collections
|
|
import os.path
|
|
import bb.utils, bb.data
|
|
from itertools import chain
|
|
from pysh import pyshyacc, pyshlex, sherrors
|
|
from bb.cache import MultiProcessCache
|
|
|
|
|
|
logger = logging.getLogger('BitBake.CodeParser')
|
|
|
|
try:
|
|
import cPickle as pickle
|
|
except ImportError:
|
|
import pickle
|
|
logger.info('Importing cPickle failed. Falling back to a very slow implementation.')
|
|
|
|
|
|
def check_indent(codestr):
|
|
"""If the code is indented, add a top level piece of code to 'remove' the indentation"""
|
|
|
|
i = 0
|
|
while codestr[i] in ["\n", "\t", " "]:
|
|
i = i + 1
|
|
|
|
if i == 0:
|
|
return codestr
|
|
|
|
if codestr[i-1] == "\t" or codestr[i-1] == " ":
|
|
return "if 1:\n" + codestr
|
|
|
|
return codestr
|
|
|
|
|
|
class CodeParserCache(MultiProcessCache):
|
|
cache_file_name = "bb_codeparser.dat"
|
|
CACHE_VERSION = 4
|
|
|
|
def __init__(self):
|
|
MultiProcessCache.__init__(self)
|
|
self.pythoncache = self.cachedata[0]
|
|
self.shellcache = self.cachedata[1]
|
|
self.pythoncacheextras = self.cachedata_extras[0]
|
|
self.shellcacheextras = self.cachedata_extras[1]
|
|
|
|
def init_cache(self, d):
|
|
MultiProcessCache.init_cache(self, d)
|
|
|
|
# cachedata gets re-assigned in the parent
|
|
self.pythoncache = self.cachedata[0]
|
|
self.shellcache = self.cachedata[1]
|
|
|
|
def compress_keys(self, data):
|
|
# When the dicts are originally created, python calls intern() on the set keys
|
|
# which significantly improves memory usage. Sadly the pickle/unpickle process
|
|
# doesn't call intern() on the keys and results in the same strings being duplicated
|
|
# in memory. This also means pickle will save the same string multiple times in
|
|
# the cache file. By interning the data here, the cache file shrinks dramatically
|
|
# meaning faster load times and the reloaded cache files also consume much less
|
|
# memory. This is worth any performance hit from this loops and the use of the
|
|
# intern() data storage.
|
|
# Python 3.x may behave better in this area
|
|
for h in data[0]:
|
|
data[0][h]["refs"] = self.internSet(data[0][h]["refs"])
|
|
data[0][h]["execs"] = self.internSet(data[0][h]["execs"])
|
|
for h in data[1]:
|
|
data[1][h]["execs"] = self.internSet(data[1][h]["execs"])
|
|
return
|
|
|
|
def create_cachedata(self):
|
|
data = [{}, {}]
|
|
return data
|
|
|
|
codeparsercache = CodeParserCache()
|
|
|
|
def parser_cache_init(d):
|
|
codeparsercache.init_cache(d)
|
|
|
|
def parser_cache_save(d):
|
|
codeparsercache.save_extras(d)
|
|
|
|
def parser_cache_savemerge(d):
|
|
codeparsercache.save_merge(d)
|
|
|
|
Logger = logging.getLoggerClass()
|
|
class BufferedLogger(Logger):
|
|
def __init__(self, name, level=0, target=None):
|
|
Logger.__init__(self, name)
|
|
self.setLevel(level)
|
|
self.buffer = []
|
|
self.target = target
|
|
|
|
def handle(self, record):
|
|
self.buffer.append(record)
|
|
|
|
def flush(self):
|
|
for record in self.buffer:
|
|
self.target.handle(record)
|
|
self.buffer = []
|
|
|
|
class PythonParser():
|
|
getvars = ("d.getVar", "bb.data.getVar", "data.getVar", "d.appendVar", "d.prependVar")
|
|
containsfuncs = ("bb.utils.contains", "base_contains", "oe.utils.contains")
|
|
execfuncs = ("bb.build.exec_func", "bb.build.exec_task")
|
|
|
|
def warn(self, func, arg):
|
|
"""Warn about calls of bitbake APIs which pass a non-literal
|
|
argument for the variable name, as we're not able to track such
|
|
a reference.
|
|
"""
|
|
|
|
try:
|
|
funcstr = codegen.to_source(func)
|
|
argstr = codegen.to_source(arg)
|
|
except TypeError:
|
|
self.log.debug(2, 'Failed to convert function and argument to source form')
|
|
else:
|
|
self.log.debug(1, self.unhandled_message % (funcstr, argstr))
|
|
|
|
def visit_Call(self, node):
|
|
name = self.called_node_name(node.func)
|
|
if name in self.getvars or name in self.containsfuncs:
|
|
if isinstance(node.args[0], ast.Str):
|
|
varname = node.args[0].s
|
|
if name in self.containsfuncs and isinstance(node.args[1], ast.Str):
|
|
self.contains[varname].add(node.args[1].s)
|
|
else:
|
|
self.references.add(node.args[0].s)
|
|
else:
|
|
self.warn(node.func, node.args[0])
|
|
elif name in self.execfuncs:
|
|
if isinstance(node.args[0], ast.Str):
|
|
self.var_execs.add(node.args[0].s)
|
|
else:
|
|
self.warn(node.func, node.args[0])
|
|
elif name and isinstance(node.func, (ast.Name, ast.Attribute)):
|
|
self.execs.add(name)
|
|
|
|
def called_node_name(self, node):
|
|
"""Given a called node, return its original string form"""
|
|
components = []
|
|
while node:
|
|
if isinstance(node, ast.Attribute):
|
|
components.append(node.attr)
|
|
node = node.value
|
|
elif isinstance(node, ast.Name):
|
|
components.append(node.id)
|
|
return '.'.join(reversed(components))
|
|
else:
|
|
break
|
|
|
|
def __init__(self, name, log):
|
|
self.var_execs = set()
|
|
self.contains = collections.defaultdict(set)
|
|
self.execs = set()
|
|
self.references = set()
|
|
self.log = BufferedLogger('BitBake.Data.%s' % name, logging.DEBUG, log)
|
|
|
|
self.unhandled_message = "in call of %s, argument '%s' is not a string literal"
|
|
self.unhandled_message = "while parsing %s, %s" % (name, self.unhandled_message)
|
|
|
|
def parse_python(self, node):
|
|
h = hash(str(node))
|
|
|
|
if h in codeparsercache.pythoncache:
|
|
self.references = codeparsercache.pythoncache[h]["refs"]
|
|
self.execs = codeparsercache.pythoncache[h]["execs"]
|
|
self.contains = codeparsercache.pythoncache[h]["contains"]
|
|
return
|
|
|
|
if h in codeparsercache.pythoncacheextras:
|
|
self.references = codeparsercache.pythoncacheextras[h]["refs"]
|
|
self.execs = codeparsercache.pythoncacheextras[h]["execs"]
|
|
self.contains = codeparsercache.pythoncacheextras[h]["contains"]
|
|
return
|
|
|
|
code = compile(check_indent(str(node)), "<string>", "exec",
|
|
ast.PyCF_ONLY_AST)
|
|
|
|
for n in ast.walk(code):
|
|
if n.__class__.__name__ == "Call":
|
|
self.visit_Call(n)
|
|
|
|
self.references.update(self.var_execs)
|
|
|
|
codeparsercache.pythoncacheextras[h] = {}
|
|
codeparsercache.pythoncacheextras[h]["refs"] = self.references
|
|
codeparsercache.pythoncacheextras[h]["execs"] = self.execs
|
|
codeparsercache.pythoncacheextras[h]["contains"] = self.contains
|
|
|
|
class ShellParser():
|
|
def __init__(self, name, log):
|
|
self.funcdefs = set()
|
|
self.allexecs = set()
|
|
self.execs = set()
|
|
self.log = BufferedLogger('BitBake.Data.%s' % name, logging.DEBUG, log)
|
|
self.unhandled_template = "unable to handle non-literal command '%s'"
|
|
self.unhandled_template = "while parsing %s, %s" % (name, self.unhandled_template)
|
|
|
|
def parse_shell(self, value):
|
|
"""Parse the supplied shell code in a string, returning the external
|
|
commands it executes.
|
|
"""
|
|
|
|
h = hash(str(value))
|
|
|
|
if h in codeparsercache.shellcache:
|
|
self.execs = codeparsercache.shellcache[h]["execs"]
|
|
return self.execs
|
|
|
|
if h in codeparsercache.shellcacheextras:
|
|
self.execs = codeparsercache.shellcacheextras[h]["execs"]
|
|
return self.execs
|
|
|
|
try:
|
|
tokens, _ = pyshyacc.parse(value, eof=True, debug=False)
|
|
except pyshlex.NeedMore:
|
|
raise sherrors.ShellSyntaxError("Unexpected EOF")
|
|
|
|
for token in tokens:
|
|
self.process_tokens(token)
|
|
self.execs = set(cmd for cmd in self.allexecs if cmd not in self.funcdefs)
|
|
|
|
codeparsercache.shellcacheextras[h] = {}
|
|
codeparsercache.shellcacheextras[h]["execs"] = self.execs
|
|
|
|
return self.execs
|
|
|
|
def process_tokens(self, tokens):
|
|
"""Process a supplied portion of the syntax tree as returned by
|
|
pyshyacc.parse.
|
|
"""
|
|
|
|
def function_definition(value):
|
|
self.funcdefs.add(value.name)
|
|
return [value.body], None
|
|
|
|
def case_clause(value):
|
|
# Element 0 of each item in the case is the list of patterns, and
|
|
# Element 1 of each item in the case is the list of commands to be
|
|
# executed when that pattern matches.
|
|
words = chain(*[item[0] for item in value.items])
|
|
cmds = chain(*[item[1] for item in value.items])
|
|
return cmds, words
|
|
|
|
def if_clause(value):
|
|
main = chain(value.cond, value.if_cmds)
|
|
rest = value.else_cmds
|
|
if isinstance(rest, tuple) and rest[0] == "elif":
|
|
return chain(main, if_clause(rest[1]))
|
|
else:
|
|
return chain(main, rest)
|
|
|
|
def simple_command(value):
|
|
return None, chain(value.words, (assign[1] for assign in value.assigns))
|
|
|
|
token_handlers = {
|
|
"and_or": lambda x: ((x.left, x.right), None),
|
|
"async": lambda x: ([x], None),
|
|
"brace_group": lambda x: (x.cmds, None),
|
|
"for_clause": lambda x: (x.cmds, x.items),
|
|
"function_definition": function_definition,
|
|
"if_clause": lambda x: (if_clause(x), None),
|
|
"pipeline": lambda x: (x.commands, None),
|
|
"redirect_list": lambda x: ([x.cmd], None),
|
|
"subshell": lambda x: (x.cmds, None),
|
|
"while_clause": lambda x: (chain(x.condition, x.cmds), None),
|
|
"until_clause": lambda x: (chain(x.condition, x.cmds), None),
|
|
"simple_command": simple_command,
|
|
"case_clause": case_clause,
|
|
}
|
|
|
|
for token in tokens:
|
|
name, value = token
|
|
try:
|
|
more_tokens, words = token_handlers[name](value)
|
|
except KeyError:
|
|
raise NotImplementedError("Unsupported token type " + name)
|
|
|
|
if more_tokens:
|
|
self.process_tokens(more_tokens)
|
|
|
|
if words:
|
|
self.process_words(words)
|
|
|
|
def process_words(self, words):
|
|
"""Process a set of 'words' in pyshyacc parlance, which includes
|
|
extraction of executed commands from $() blocks, as well as grabbing
|
|
the command name argument.
|
|
"""
|
|
|
|
words = list(words)
|
|
for word in list(words):
|
|
wtree = pyshlex.make_wordtree(word[1])
|
|
for part in wtree:
|
|
if not isinstance(part, list):
|
|
continue
|
|
|
|
if part[0] in ('`', '$('):
|
|
command = pyshlex.wordtree_as_string(part[1:-1])
|
|
self.parse_shell(command)
|
|
|
|
if word[0] in ("cmd_name", "cmd_word"):
|
|
if word in words:
|
|
words.remove(word)
|
|
|
|
usetoken = False
|
|
for word in words:
|
|
if word[0] in ("cmd_name", "cmd_word") or \
|
|
(usetoken and word[0] == "TOKEN"):
|
|
if "=" in word[1]:
|
|
usetoken = True
|
|
continue
|
|
|
|
cmd = word[1]
|
|
if cmd.startswith("$"):
|
|
self.log.debug(1, self.unhandled_template % cmd)
|
|
elif cmd == "eval":
|
|
command = " ".join(word for _, word in words[1:])
|
|
self.parse_shell(command)
|
|
else:
|
|
self.allexecs.add(cmd)
|
|
break
|