You've already forked python-uncompyle6
mirror of
https://github.com/rocky/python-uncompyle6.git
synced 2025-08-03 00:45:53 +08:00
1338 lines
57 KiB
Python
1338 lines
57 KiB
Python
# Copyright (c) 2016-2017, 2019 Rocky Bernstein
|
|
"""
|
|
Python 3.7 base code. We keep non-custom-generated grammar rules out of this file.
|
|
"""
|
|
from uncompyle6.scanners.tok import Token
|
|
from uncompyle6.parser import PythonParser, PythonParserSingle, nop_func
|
|
from uncompyle6.parsers.treenode import SyntaxTree
|
|
from spark_parser import DEFAULT_DEBUG as PARSER_DEFAULT_DEBUG
|
|
|
|
|
|
class Python37BaseParser(PythonParser):
|
|
def __init__(self, debug_parser=PARSER_DEFAULT_DEBUG):
|
|
self.added_rules = set()
|
|
super(Python37BaseParser, self).__init__(
|
|
SyntaxTree, "stmts", debug=debug_parser
|
|
)
|
|
self.new_rules = set()
|
|
|
|
@staticmethod
|
|
def call_fn_name(token):
|
|
"""Customize CALL_FUNCTION to add the number of positional arguments"""
|
|
if token.attr is not None:
|
|
return "%s_%i" % (token.kind, token.attr)
|
|
else:
|
|
return "%s_0" % (token.kind)
|
|
|
|
def add_make_function_rule(self, rule, opname, attr, customize):
|
|
"""Python 3.3 added a an addtional LOAD_STR before MAKE_FUNCTION and
|
|
this has an effect on many rules.
|
|
"""
|
|
new_rule = rule % "LOAD_STR "
|
|
self.add_unique_rule(new_rule, opname, attr, customize)
|
|
|
|
def custom_build_class_rule(self, opname, i, token, tokens, customize):
|
|
"""
|
|
# Should the first rule be somehow folded into the 2nd one?
|
|
build_class ::= LOAD_BUILD_CLASS mkfunc
|
|
LOAD_CLASSNAME {expr}^n-1 CALL_FUNCTION_n
|
|
LOAD_CONST CALL_FUNCTION_n
|
|
build_class ::= LOAD_BUILD_CLASS mkfunc
|
|
expr
|
|
call
|
|
CALL_FUNCTION_3
|
|
"""
|
|
# FIXME: I bet this can be simplified
|
|
# look for next MAKE_FUNCTION
|
|
for i in range(i + 1, len(tokens)):
|
|
if tokens[i].kind.startswith("MAKE_FUNCTION"):
|
|
break
|
|
elif tokens[i].kind.startswith("MAKE_CLOSURE"):
|
|
break
|
|
pass
|
|
assert i < len(
|
|
tokens
|
|
), "build_class needs to find MAKE_FUNCTION or MAKE_CLOSURE"
|
|
assert (
|
|
tokens[i + 1].kind == "LOAD_STR"
|
|
), "build_class expecting CONST after MAKE_FUNCTION/MAKE_CLOSURE"
|
|
call_fn_tok = None
|
|
for i in range(i, len(tokens)):
|
|
if tokens[i].kind.startswith("CALL_FUNCTION"):
|
|
call_fn_tok = tokens[i]
|
|
break
|
|
if not call_fn_tok:
|
|
raise RuntimeError(
|
|
"build_class custom rule for %s needs to find CALL_FUNCTION" % opname
|
|
)
|
|
|
|
# customize build_class rule
|
|
# FIXME: What's the deal with the two rules? Different Python versions?
|
|
# Different situations? Note that the above rule is based on the CALL_FUNCTION
|
|
# token found, while this one doesn't.
|
|
# 3.6+ handling
|
|
call_function = call_fn_tok.kind
|
|
if call_function.startswith("CALL_FUNCTION_KW"):
|
|
self.addRule("classdef ::= build_class_kw store", nop_func)
|
|
rule = "build_class_kw ::= LOAD_BUILD_CLASS mkfunc %sLOAD_CONST %s" % (
|
|
"expr " * (call_fn_tok.attr - 1),
|
|
call_function,
|
|
)
|
|
else:
|
|
call_function = self.call_fn_name(call_fn_tok)
|
|
rule = "build_class ::= LOAD_BUILD_CLASS mkfunc %s%s" % (
|
|
"expr " * (call_fn_tok.attr - 1),
|
|
call_function,
|
|
)
|
|
self.addRule(rule, nop_func)
|
|
return
|
|
|
|
# FIXME FIXME FIXME: The below is an utter mess. Come up with a better
|
|
# organization for this. For example, arrange organize by opcode base?
|
|
|
|
def customize_grammar_rules(self, tokens, customize):
|
|
|
|
is_pypy = False
|
|
|
|
# For a rough break out on the first word. This may
|
|
# include instructions that don't need customization,
|
|
# but we'll do a finer check after the rough breakout.
|
|
customize_instruction_basenames = frozenset(
|
|
(
|
|
"BEFORE",
|
|
"BUILD",
|
|
"CALL",
|
|
"CONTINUE",
|
|
"DELETE",
|
|
"FORMAT",
|
|
"GET",
|
|
"JUMP",
|
|
"LOAD",
|
|
"LOOKUP",
|
|
"MAKE",
|
|
"RETURN",
|
|
"RAISE",
|
|
"SETUP",
|
|
"UNPACK",
|
|
)
|
|
)
|
|
|
|
# Opcode names in the custom_ops_processed set have rules that get added
|
|
# unconditionally and the rules are constant. So they need to be done
|
|
# only once and if we see the opcode a second we don't have to consider
|
|
# adding more rules.
|
|
#
|
|
# Note: BUILD_TUPLE_UNPACK_WITH_CALL gets considered by
|
|
# default because it starts with BUILD. So we'll set to ignore it from
|
|
# the start.
|
|
custom_ops_processed = set(("BUILD_TUPLE_UNPACK_WITH_CALL",))
|
|
|
|
# A set of instruction operation names that exist in the token stream.
|
|
# We use this customize the grammar that we create.
|
|
# 2.6-compatible set comprehensions
|
|
self.seen_ops = frozenset([t.kind for t in tokens])
|
|
self.seen_op_basenames = frozenset(
|
|
[opname[: opname.rfind("_")] for opname in self.seen_ops]
|
|
)
|
|
|
|
# Loop over instructions adding custom grammar rules based on
|
|
# a specific instruction seen.
|
|
|
|
if "PyPy" in customize:
|
|
is_pypy = True
|
|
self.addRule(
|
|
"""
|
|
stmt ::= assign3_pypy
|
|
stmt ::= assign2_pypy
|
|
assign3_pypy ::= expr expr expr store store store
|
|
assign2_pypy ::= expr expr store store
|
|
stmt ::= if_expr_lambda
|
|
stmt ::= conditional_not_lambda
|
|
if_expr_lambda ::= expr jmp_false expr return_if_lambda
|
|
return_lambda LAMBDA_MARKER
|
|
conditional_not_lambda
|
|
::= expr jmp_true expr return_if_lambda
|
|
return_lambda LAMBDA_MARKER
|
|
""",
|
|
nop_func,
|
|
)
|
|
|
|
n = len(tokens)
|
|
|
|
# Determine if we have an iteration CALL_FUNCTION_1.
|
|
has_get_iter_call_function1 = False
|
|
for i, token in enumerate(tokens):
|
|
if (
|
|
token == "GET_ITER"
|
|
and i < n - 2
|
|
and self.call_fn_name(tokens[i + 1]) == "CALL_FUNCTION_1"
|
|
):
|
|
has_get_iter_call_function1 = True
|
|
|
|
for i, token in enumerate(tokens):
|
|
opname = token.kind
|
|
|
|
# Do a quick breakout before testing potentially
|
|
# each of the dozen or so instruction in if elif.
|
|
if (
|
|
opname[: opname.find("_")] not in customize_instruction_basenames
|
|
or opname in custom_ops_processed
|
|
):
|
|
continue
|
|
|
|
opname_base = opname[: opname.rfind("_")]
|
|
|
|
# The order of opname listed is roughly sorted below
|
|
|
|
if opname == "LOAD_ASSERT" and "PyPy" in customize:
|
|
rules_str = """
|
|
stmt ::= JUMP_IF_NOT_DEBUG stmts COME_FROM
|
|
"""
|
|
self.add_unique_doc_rules(rules_str, customize)
|
|
|
|
elif opname == "BEFORE_ASYNC_WITH":
|
|
rules_str = """
|
|
stmt ::= async_with_stmt
|
|
stmt ::= async_with_as_stmt
|
|
"""
|
|
|
|
if self.version < 3.8:
|
|
rules_str += """
|
|
async_with_stmt ::= expr
|
|
BEFORE_ASYNC_WITH GET_AWAITABLE LOAD_CONST YIELD_FROM
|
|
SETUP_ASYNC_WITH POP_TOP suite_stmts_opt
|
|
POP_BLOCK LOAD_CONST COME_FROM_ASYNC_WITH
|
|
WITH_CLEANUP_START
|
|
GET_AWAITABLE LOAD_CONST YIELD_FROM
|
|
WITH_CLEANUP_FINISH END_FINALLY
|
|
async_with_as_stmt ::= expr
|
|
BEFORE_ASYNC_WITH GET_AWAITABLE LOAD_CONST YIELD_FROM
|
|
SETUP_ASYNC_WITH store suite_stmts_opt
|
|
POP_BLOCK LOAD_CONST COME_FROM_ASYNC_WITH
|
|
WITH_CLEANUP_START
|
|
GET_AWAITABLE LOAD_CONST YIELD_FROM
|
|
WITH_CLEANUP_FINISH END_FINALLY
|
|
"""
|
|
else:
|
|
rules_str += """
|
|
async_with_stmt ::= expr
|
|
BEFORE_ASYNC_WITH GET_AWAITABLE LOAD_CONST YIELD_FROM
|
|
SETUP_ASYNC_WITH POP_TOP suite_stmts
|
|
POP_TOP POP_BLOCK BEGIN_FINALLY COME_FROM_ASYNC_WITH
|
|
WITH_CLEANUP_START
|
|
GET_AWAITABLE LOAD_CONST YIELD_FROM
|
|
WITH_CLEANUP_FINISH END_FINALLY
|
|
async_with_as_stmt ::= expr
|
|
BEFORE_ASYNC_WITH GET_AWAITABLE LOAD_CONST YIELD_FROM
|
|
SETUP_ASYNC_WITH store suite_stmts
|
|
POP_TOP POP_BLOCK BEGIN_FINALLY COME_FROM_ASYNC_WITH
|
|
WITH_CLEANUP_START
|
|
GET_AWAITABLE LOAD_CONST YIELD_FROM
|
|
WITH_CLEANUP_FINISH END_FINALLY
|
|
"""
|
|
self.addRule(rules_str, nop_func)
|
|
|
|
elif opname_base == "BUILD_CONST_KEY_MAP":
|
|
kvlist_n = "expr " * (token.attr)
|
|
rule = "dict ::= %sLOAD_CONST %s" % (kvlist_n, opname)
|
|
self.addRule(rule, nop_func)
|
|
|
|
elif opname.startswith("BUILD_LIST_UNPACK"):
|
|
v = token.attr
|
|
rule = "build_list_unpack ::= %s%s" % ("expr " * v, opname)
|
|
self.addRule(rule, nop_func)
|
|
rule = "expr ::= build_list_unpack"
|
|
self.addRule(rule, nop_func)
|
|
|
|
elif opname_base in ("BUILD_MAP", "BUILD_MAP_UNPACK"):
|
|
|
|
if opname == "BUILD_MAP_UNPACK":
|
|
self.addRule(
|
|
"""
|
|
expr ::= unmap_dict
|
|
unmap_dict ::= dict_comp BUILD_MAP_UNPACK
|
|
""",
|
|
nop_func,
|
|
)
|
|
pass
|
|
elif opname.startswith("BUILD_MAP_UNPACK_WITH_CALL"):
|
|
v = token.attr
|
|
rule = "build_map_unpack_with_call ::= %s%s" % ("expr " * v, opname)
|
|
self.addRule(rule, nop_func)
|
|
|
|
kvlist_n = "kvlist_%s" % token.attr
|
|
if opname == "BUILD_MAP_n":
|
|
# PyPy sometimes has no count. Sigh.
|
|
rule = (
|
|
"dict_comp_func ::= BUILD_MAP_n LOAD_FAST for_iter store "
|
|
"comp_iter JUMP_BACK RETURN_VALUE RETURN_LAST"
|
|
)
|
|
self.add_unique_rule(rule, "dict_comp_func", 1, customize)
|
|
|
|
kvlist_n = "kvlist_n"
|
|
rule = "kvlist_n ::= kvlist_n kv3"
|
|
self.add_unique_rule(rule, "kvlist_n", 0, customize)
|
|
rule = "kvlist_n ::="
|
|
self.add_unique_rule(rule, "kvlist_n", 1, customize)
|
|
rule = "dict ::= BUILD_MAP_n kvlist_n"
|
|
|
|
if not opname.startswith("BUILD_MAP_WITH_CALL"):
|
|
# FIXME: Use the attr
|
|
# so this doesn't run into exponential parsing time.
|
|
if opname.startswith("BUILD_MAP_UNPACK"):
|
|
# FIXME: start here. The LHS should be unmap_dict, not dict.
|
|
# FIXME: really we need a combination of dict_entry-like things.
|
|
# It just so happens the most common case is not to mix
|
|
# dictionary comphensions with dictionary, elements
|
|
if "LOAD_DICTCOMP" in self.seen_ops:
|
|
rule = "dict ::= %s%s" % ("dict_comp " * token.attr, opname)
|
|
self.addRule(rule, nop_func)
|
|
rule = """
|
|
expr ::= unmap_dict
|
|
unmap_dict ::= %s%s
|
|
""" % (
|
|
"expr " * token.attr,
|
|
opname,
|
|
)
|
|
else:
|
|
rule = "%s ::= %s %s" % (
|
|
kvlist_n,
|
|
"expr " * (token.attr * 2),
|
|
opname,
|
|
)
|
|
self.add_unique_rule(rule, opname, token.attr, customize)
|
|
rule = "dict ::= %s" % kvlist_n
|
|
self.add_unique_rule(rule, opname, token.attr, customize)
|
|
|
|
elif opname.startswith("BUILD_MAP_UNPACK_WITH_CALL"):
|
|
v = token.attr
|
|
rule = "build_map_unpack_with_call ::= %s%s" % ("expr " * v, opname)
|
|
self.addRule(rule, nop_func)
|
|
|
|
elif opname.startswith("BUILD_TUPLE_UNPACK_WITH_CALL"):
|
|
v = token.attr
|
|
rule = (
|
|
"build_tuple_unpack_with_call ::= "
|
|
+ "expr1024 " * int(v // 1024)
|
|
+ "expr32 " * int((v // 32) % 32)
|
|
+ "expr " * (v % 32)
|
|
+ opname
|
|
)
|
|
self.addRule(rule, nop_func)
|
|
rule = "starred ::= %s %s" % ("expr " * v, opname)
|
|
self.addRule(rule, nop_func)
|
|
|
|
elif opname_base in (
|
|
"BUILD_LIST",
|
|
"BUILD_SET",
|
|
"BUILD_TUPLE",
|
|
"BUILD_TUPLE_UNPACK",
|
|
):
|
|
v = token.attr
|
|
|
|
is_LOAD_CLOSURE = False
|
|
if opname_base == "BUILD_TUPLE":
|
|
# If is part of a "load_closure", then it is not part of a
|
|
# "list".
|
|
is_LOAD_CLOSURE = True
|
|
for j in range(v):
|
|
if tokens[i - j - 1].kind != "LOAD_CLOSURE":
|
|
is_LOAD_CLOSURE = False
|
|
break
|
|
if is_LOAD_CLOSURE:
|
|
rule = "load_closure ::= %s%s" % (("LOAD_CLOSURE " * v), opname)
|
|
self.add_unique_rule(rule, opname, token.attr, customize)
|
|
if not is_LOAD_CLOSURE or v == 0:
|
|
# We do this complicated test to speed up parsing of
|
|
# pathelogically long literals, especially those over 1024.
|
|
build_count = token.attr
|
|
thousands = build_count // 1024
|
|
thirty32s = (build_count // 32) % 32
|
|
if thirty32s > 0:
|
|
rule = "expr32 ::=%s" % (" expr" * 32)
|
|
self.add_unique_rule(rule, opname_base, build_count, customize)
|
|
pass
|
|
if thousands > 0:
|
|
self.add_unique_rule(
|
|
"expr1024 ::=%s" % (" expr32" * 32),
|
|
opname_base,
|
|
build_count,
|
|
customize,
|
|
)
|
|
pass
|
|
collection = opname_base[opname_base.find("_") + 1 :].lower()
|
|
rule = (
|
|
("%s ::= " % collection)
|
|
+ "expr1024 " * thousands
|
|
+ "expr32 " * thirty32s
|
|
+ "expr " * (build_count % 32)
|
|
+ opname
|
|
)
|
|
self.add_unique_rules(["expr ::= %s" % collection, rule], customize)
|
|
continue
|
|
continue
|
|
elif opname_base == "BUILD_SLICE":
|
|
if token.attr == 2:
|
|
self.add_unique_rules(
|
|
[
|
|
"expr ::= build_slice2",
|
|
"build_slice2 ::= expr expr BUILD_SLICE_2",
|
|
],
|
|
customize,
|
|
)
|
|
else:
|
|
assert token.attr == 3, (
|
|
"BUILD_SLICE value must be 2 or 3; is %s" % v
|
|
)
|
|
self.add_unique_rules(
|
|
[
|
|
"expr ::= build_slice3",
|
|
"build_slice3 ::= expr expr expr BUILD_SLICE_3",
|
|
],
|
|
customize,
|
|
)
|
|
|
|
elif opname.startswith("BUILD_STRING"):
|
|
v = token.attr
|
|
rules_str = """
|
|
expr ::= joined_str
|
|
joined_str ::= %sBUILD_STRING_%d
|
|
""" % (
|
|
"expr " * v,
|
|
v,
|
|
)
|
|
self.add_unique_doc_rules(rules_str, customize)
|
|
if "FORMAT_VALUE_ATTR" in self.seen_ops:
|
|
rules_str = """
|
|
formatted_value_attr ::= expr expr FORMAT_VALUE_ATTR expr BUILD_STRING
|
|
expr ::= formatted_value_attr
|
|
"""
|
|
self.add_unique_doc_rules(rules_str, customize)
|
|
|
|
elif opname in frozenset(
|
|
(
|
|
"CALL_FUNCTION",
|
|
"CALL_FUNCTION_EX",
|
|
"CALL_FUNCTION_EX_KW",
|
|
"CALL_FUNCTION_VAR",
|
|
"CALL_FUNCTION_VAR_KW",
|
|
)
|
|
) or opname.startswith("CALL_FUNCTION_KW"):
|
|
|
|
if opname == "CALL_FUNCTION" and token.attr == 1:
|
|
rule = """
|
|
dict_comp ::= LOAD_DICTCOMP LOAD_STR MAKE_FUNCTION_0 expr
|
|
GET_ITER CALL_FUNCTION_1
|
|
classdefdeco1 ::= expr classdefdeco2 CALL_FUNCTION_1
|
|
"""
|
|
self.addRule(rule, nop_func)
|
|
|
|
self.custom_classfunc_rule(opname, token, customize, tokens[i + 1])
|
|
# Note: don't add to custom_ops_processed.
|
|
|
|
elif opname_base == "CALL_METHOD":
|
|
# PyPy and Python 3.7+ only - DRY with parse2
|
|
|
|
args_pos, args_kw = self.get_pos_kw(token)
|
|
|
|
# number of apply equiv arguments:
|
|
nak = (len(opname_base) - len("CALL_METHOD")) // 3
|
|
rule = (
|
|
"call ::= expr "
|
|
+ ("pos_arg " * args_pos)
|
|
+ ("kwarg " * args_kw)
|
|
+ "expr " * nak
|
|
+ opname
|
|
)
|
|
self.add_unique_rule(rule, opname, token.attr, customize)
|
|
|
|
elif opname == "CONTINUE":
|
|
self.addRule("continue ::= CONTINUE", nop_func)
|
|
custom_ops_processed.add(opname)
|
|
elif opname == "CONTINUE_LOOP":
|
|
self.addRule("continue ::= CONTINUE_LOOP", nop_func)
|
|
custom_ops_processed.add(opname)
|
|
elif opname == "DELETE_ATTR":
|
|
self.addRule("del_stmt ::= expr DELETE_ATTR", nop_func)
|
|
custom_ops_processed.add(opname)
|
|
elif opname == "DELETE_DEREF":
|
|
self.addRule(
|
|
"""
|
|
stmt ::= del_deref_stmt
|
|
del_deref_stmt ::= DELETE_DEREF
|
|
""",
|
|
nop_func,
|
|
)
|
|
custom_ops_processed.add(opname)
|
|
elif opname == "DELETE_SUBSCR":
|
|
self.addRule(
|
|
"""
|
|
del_stmt ::= delete_subscript
|
|
delete_subscript ::= expr expr DELETE_SUBSCR
|
|
""",
|
|
nop_func,
|
|
)
|
|
custom_ops_processed.add(opname)
|
|
|
|
elif opname == "FORMAT_VALUE":
|
|
rules_str = """
|
|
expr ::= formatted_value1
|
|
formatted_value1 ::= expr FORMAT_VALUE
|
|
"""
|
|
self.add_unique_doc_rules(rules_str, customize)
|
|
|
|
elif opname == "FORMAT_VALUE_ATTR":
|
|
rules_str = """
|
|
expr ::= formatted_value2
|
|
formatted_value2 ::= expr expr FORMAT_VALUE_ATTR
|
|
"""
|
|
self.add_unique_doc_rules(rules_str, customize)
|
|
|
|
elif opname == "GET_ITER":
|
|
self.addRule(
|
|
"""
|
|
expr ::= get_iter
|
|
attribute ::= expr GET_ITER
|
|
""",
|
|
nop_func,
|
|
)
|
|
custom_ops_processed.add(opname)
|
|
elif opname == "JUMP_IF_NOT_DEBUG":
|
|
v = token.attr
|
|
self.addRule(
|
|
"""
|
|
stmt ::= assert_pypy
|
|
stmt ::= assert2_pypy", nop_func)
|
|
assert_pypy ::= JUMP_IF_NOT_DEBUG assert_expr jmp_true
|
|
LOAD_ASSERT RAISE_VARARGS_1 COME_FROM
|
|
assert2_pypy ::= JUMP_IF_NOT_DEBUG assert_expr jmp_true
|
|
LOAD_ASSERT expr CALL_FUNCTION_1
|
|
RAISE_VARARGS_1 COME_FROM
|
|
assert2_pypy ::= JUMP_IF_NOT_DEBUG assert_expr jmp_true
|
|
LOAD_ASSERT expr CALL_FUNCTION_1
|
|
RAISE_VARARGS_1 COME_FROM,
|
|
""",
|
|
nop_func,
|
|
)
|
|
custom_ops_processed.add(opname)
|
|
elif opname == "LOAD_BUILD_CLASS":
|
|
self.custom_build_class_rule(opname, i, token, tokens, customize)
|
|
# Note: don't add to custom_ops_processed.
|
|
elif opname == "LOAD_CLASSDEREF":
|
|
# Python 3.4+
|
|
self.addRule("expr ::= LOAD_CLASSDEREF", nop_func)
|
|
custom_ops_processed.add(opname)
|
|
elif opname == "LOAD_CLASSNAME":
|
|
self.addRule("expr ::= LOAD_CLASSNAME", nop_func)
|
|
custom_ops_processed.add(opname)
|
|
elif opname == "LOAD_DICTCOMP":
|
|
if has_get_iter_call_function1:
|
|
rule_pat = (
|
|
"dict_comp ::= LOAD_DICTCOMP %sMAKE_FUNCTION_0 expr "
|
|
"GET_ITER CALL_FUNCTION_1"
|
|
)
|
|
self.add_make_function_rule(rule_pat, opname, token.attr, customize)
|
|
pass
|
|
custom_ops_processed.add(opname)
|
|
elif opname == "LOAD_ATTR":
|
|
self.addRule(
|
|
"""
|
|
expr ::= attribute
|
|
attribute ::= expr LOAD_ATTR
|
|
""",
|
|
nop_func,
|
|
)
|
|
custom_ops_processed.add(opname)
|
|
elif opname == "LOAD_LISTCOMP":
|
|
self.add_unique_rule("expr ::= listcomp", opname, token.attr, customize)
|
|
custom_ops_processed.add(opname)
|
|
elif opname == "LOAD_SETCOMP":
|
|
# Should this be generalized and put under MAKE_FUNCTION?
|
|
if has_get_iter_call_function1:
|
|
self.addRule("expr ::= set_comp", nop_func)
|
|
rule_pat = (
|
|
"set_comp ::= LOAD_SETCOMP %sMAKE_FUNCTION_0 expr "
|
|
"GET_ITER CALL_FUNCTION_1"
|
|
)
|
|
self.add_make_function_rule(rule_pat, opname, token.attr, customize)
|
|
pass
|
|
custom_ops_processed.add(opname)
|
|
elif opname == "LOOKUP_METHOD":
|
|
# A PyPy speciality - DRY with parse3
|
|
self.addRule(
|
|
"""
|
|
expr ::= attribute
|
|
attribute ::= expr LOOKUP_METHOD
|
|
""",
|
|
nop_func,
|
|
)
|
|
custom_ops_processed.add(opname)
|
|
elif opname.startswith("MAKE_CLOSURE"):
|
|
# DRY with MAKE_FUNCTION
|
|
# Note: this probably doesn't handle kwargs proprerly
|
|
|
|
if opname == "MAKE_CLOSURE_0" and "LOAD_DICTCOMP" in self.seen_ops:
|
|
# Is there something general going on here?
|
|
# Note that 3.6+ doesn't do this, but we'll remove
|
|
# this rule in parse36.py
|
|
rule = """
|
|
dict_comp ::= load_closure LOAD_DICTCOMP LOAD_STR
|
|
MAKE_CLOSURE_0 expr
|
|
GET_ITER CALL_FUNCTION_1
|
|
"""
|
|
self.addRule(rule, nop_func)
|
|
|
|
args_pos, args_kw, annotate_args = token.attr
|
|
|
|
# FIXME: Fold test into add_make_function_rule
|
|
j = 2
|
|
if is_pypy or (i >= j and tokens[i - j] == "LOAD_LAMBDA"):
|
|
rule_pat = "mklambda ::= %sload_closure LOAD_LAMBDA %%s%s" % (
|
|
"pos_arg " * args_pos,
|
|
opname,
|
|
)
|
|
self.add_make_function_rule(rule_pat, opname, token.attr, customize)
|
|
|
|
if has_get_iter_call_function1:
|
|
rule_pat = (
|
|
"generator_exp ::= %sload_closure load_genexpr %%s%s expr "
|
|
"GET_ITER CALL_FUNCTION_1" % ("pos_arg " * args_pos, opname)
|
|
)
|
|
self.add_make_function_rule(rule_pat, opname, token.attr, customize)
|
|
|
|
if has_get_iter_call_function1:
|
|
if is_pypy or (i >= j and tokens[i - j] == "LOAD_LISTCOMP"):
|
|
# In the tokens we saw:
|
|
# LOAD_LISTCOMP LOAD_CONST MAKE_FUNCTION (>= 3.3) or
|
|
# LOAD_LISTCOMP MAKE_FUNCTION (< 3.3) or
|
|
# and have GET_ITER CALL_FUNCTION_1
|
|
# Todo: For Pypy we need to modify this slightly
|
|
rule_pat = (
|
|
"listcomp ::= %sload_closure LOAD_LISTCOMP %%s%s expr "
|
|
"GET_ITER CALL_FUNCTION_1"
|
|
% ("pos_arg " * args_pos, opname)
|
|
)
|
|
self.add_make_function_rule(
|
|
rule_pat, opname, token.attr, customize
|
|
)
|
|
if is_pypy or (i >= j and tokens[i - j] == "LOAD_SETCOMP"):
|
|
rule_pat = (
|
|
"set_comp ::= %sload_closure LOAD_SETCOMP %%s%s expr "
|
|
"GET_ITER CALL_FUNCTION_1"
|
|
% ("pos_arg " * args_pos, opname)
|
|
)
|
|
self.add_make_function_rule(
|
|
rule_pat, opname, token.attr, customize
|
|
)
|
|
if is_pypy or (i >= j and tokens[i - j] == "LOAD_DICTCOMP"):
|
|
self.add_unique_rule(
|
|
"dict_comp ::= %sload_closure LOAD_DICTCOMP %s "
|
|
"expr GET_ITER CALL_FUNCTION_1"
|
|
% ("pos_arg " * args_pos, opname),
|
|
opname,
|
|
token.attr,
|
|
customize,
|
|
)
|
|
|
|
if args_kw > 0:
|
|
kwargs_str = "kwargs "
|
|
else:
|
|
kwargs_str = ""
|
|
|
|
rule = "mkfunc ::= %s%s%s load_closure LOAD_CODE LOAD_STR %s" % (
|
|
"expr " * args_pos,
|
|
kwargs_str,
|
|
"expr " * annotate_args,
|
|
opname,
|
|
)
|
|
|
|
self.add_unique_rule(rule, opname, token.attr, customize)
|
|
|
|
if args_kw == 0:
|
|
rule = "mkfunc ::= %sload_closure load_genexpr %s" % (
|
|
"pos_arg " * args_pos,
|
|
opname,
|
|
)
|
|
self.add_unique_rule(rule, opname, token.attr, customize)
|
|
|
|
pass
|
|
elif opname_base.startswith("MAKE_FUNCTION"):
|
|
args_pos, args_kw, annotate_args, closure = token.attr
|
|
stack_count = args_pos + args_kw + annotate_args
|
|
if closure:
|
|
if args_pos:
|
|
rule = "mklambda ::= %s%s%s%s" % (
|
|
"expr " * stack_count,
|
|
"load_closure " * closure,
|
|
"BUILD_TUPLE_1 LOAD_LAMBDA LOAD_STR ",
|
|
opname,
|
|
)
|
|
else:
|
|
rule = "mklambda ::= %s%s%s" % (
|
|
"load_closure " * closure,
|
|
"LOAD_LAMBDA LOAD_STR ",
|
|
opname,
|
|
)
|
|
self.add_unique_rule(rule, opname, token.attr, customize)
|
|
|
|
else:
|
|
rule = "mklambda ::= %sLOAD_LAMBDA LOAD_STR %s" % (
|
|
("expr " * stack_count),
|
|
opname,
|
|
)
|
|
self.add_unique_rule(rule, opname, token.attr, customize)
|
|
|
|
rule = "mkfunc ::= %s%s%s%s" % (
|
|
"expr " * stack_count,
|
|
"load_closure " * closure,
|
|
"LOAD_CODE LOAD_STR ",
|
|
opname,
|
|
)
|
|
self.add_unique_rule(rule, opname, token.attr, customize)
|
|
|
|
if has_get_iter_call_function1:
|
|
rule_pat = (
|
|
"generator_exp ::= %sload_genexpr %%s%s expr "
|
|
"GET_ITER CALL_FUNCTION_1" % ("pos_arg " * args_pos, opname)
|
|
)
|
|
self.add_make_function_rule(rule_pat, opname, token.attr, customize)
|
|
rule_pat = (
|
|
"generator_exp ::= %sload_closure load_genexpr %%s%s expr "
|
|
"GET_ITER CALL_FUNCTION_1" % ("pos_arg " * args_pos, opname)
|
|
)
|
|
self.add_make_function_rule(rule_pat, opname, token.attr, customize)
|
|
if is_pypy or (i >= 2 and tokens[i - 2] == "LOAD_LISTCOMP"):
|
|
# 3.6+ sometimes bundles all of the
|
|
# 'exprs' in the rule above into a
|
|
# tuple.
|
|
rule_pat = (
|
|
"listcomp ::= load_closure LOAD_LISTCOMP %%s%s "
|
|
"expr GET_ITER CALL_FUNCTION_1" % (opname,)
|
|
)
|
|
self.add_make_function_rule(
|
|
rule_pat, opname, token.attr, customize
|
|
)
|
|
rule_pat = (
|
|
"listcomp ::= %sLOAD_LISTCOMP %%s%s expr "
|
|
"GET_ITER CALL_FUNCTION_1" % ("expr " * args_pos, opname)
|
|
)
|
|
self.add_make_function_rule(
|
|
rule_pat, opname, token.attr, customize
|
|
)
|
|
|
|
if is_pypy or (i >= 2 and tokens[i - 2] == "LOAD_LAMBDA"):
|
|
rule_pat = "mklambda ::= %s%sLOAD_LAMBDA %%s%s" % (
|
|
("pos_arg " * args_pos),
|
|
("kwarg " * args_kw),
|
|
opname,
|
|
)
|
|
self.add_make_function_rule(rule_pat, opname, token.attr, customize)
|
|
continue
|
|
|
|
args_pos, args_kw, annotate_args, closure = token.attr
|
|
|
|
j = 2
|
|
|
|
if has_get_iter_call_function1:
|
|
rule_pat = (
|
|
"generator_exp ::= %sload_genexpr %%s%s expr "
|
|
"GET_ITER CALL_FUNCTION_1" % ("pos_arg " * args_pos, opname)
|
|
)
|
|
self.add_make_function_rule(rule_pat, opname, token.attr, customize)
|
|
|
|
if is_pypy or (i >= j and tokens[i - j] == "LOAD_LISTCOMP"):
|
|
# In the tokens we saw:
|
|
# LOAD_LISTCOMP LOAD_CONST MAKE_FUNCTION (>= 3.3) or
|
|
# LOAD_LISTCOMP MAKE_FUNCTION (< 3.3) or
|
|
# and have GET_ITER CALL_FUNCTION_1
|
|
# Todo: For Pypy we need to modify this slightly
|
|
rule_pat = (
|
|
"listcomp ::= %sLOAD_LISTCOMP %%s%s expr "
|
|
"GET_ITER CALL_FUNCTION_1" % ("expr " * args_pos, opname)
|
|
)
|
|
self.add_make_function_rule(
|
|
rule_pat, opname, token.attr, customize
|
|
)
|
|
|
|
# FIXME: Fold test into add_make_function_rule
|
|
if is_pypy or (i >= j and tokens[i - j] == "LOAD_LAMBDA"):
|
|
rule_pat = "mklambda ::= %s%sLOAD_LAMBDA %%s%s" % (
|
|
("pos_arg " * args_pos),
|
|
("kwarg " * args_kw),
|
|
opname,
|
|
)
|
|
self.add_make_function_rule(rule_pat, opname, token.attr, customize)
|
|
|
|
if args_kw == 0:
|
|
kwargs = "no_kwargs"
|
|
self.add_unique_rule("no_kwargs ::=", opname, token.attr, customize)
|
|
else:
|
|
kwargs = "kwargs"
|
|
|
|
# positional args before keyword args
|
|
rule = "mkfunc ::= %s%s %s%s" % (
|
|
"pos_arg " * args_pos,
|
|
kwargs,
|
|
"LOAD_CODE LOAD_STR ",
|
|
opname,
|
|
)
|
|
self.add_unique_rule(rule, opname, token.attr, customize)
|
|
|
|
elif opname == "MAKE_FUNCTION_8":
|
|
if "LOAD_DICTCOMP" in self.seen_ops:
|
|
# Is there something general going on here?
|
|
rule = """
|
|
dict_comp ::= load_closure LOAD_DICTCOMP LOAD_STR
|
|
MAKE_FUNCTION_8 expr
|
|
GET_ITER CALL_FUNCTION_1
|
|
"""
|
|
self.addRule(rule, nop_func)
|
|
elif "LOAD_SETCOMP" in self.seen_ops:
|
|
rule = """
|
|
set_comp ::= load_closure LOAD_SETCOMP LOAD_STR
|
|
MAKE_FUNCTION_8 expr
|
|
GET_ITER CALL_FUNCTION_1
|
|
"""
|
|
self.addRule(rule, nop_func)
|
|
|
|
elif opname == "RETURN_VALUE_LAMBDA":
|
|
self.addRule(
|
|
"""
|
|
return_lambda ::= ret_expr RETURN_VALUE_LAMBDA
|
|
""",
|
|
nop_func,
|
|
)
|
|
custom_ops_processed.add(opname)
|
|
elif opname == "RAISE_VARARGS_0":
|
|
self.addRule(
|
|
"""
|
|
stmt ::= raise_stmt0
|
|
raise_stmt0 ::= RAISE_VARARGS_0
|
|
""",
|
|
nop_func,
|
|
)
|
|
custom_ops_processed.add(opname)
|
|
elif opname == "RAISE_VARARGS_1":
|
|
self.addRule(
|
|
"""
|
|
stmt ::= raise_stmt1
|
|
raise_stmt1 ::= expr RAISE_VARARGS_1
|
|
""",
|
|
nop_func,
|
|
)
|
|
custom_ops_processed.add(opname)
|
|
elif opname == "RAISE_VARARGS_2":
|
|
self.addRule(
|
|
"""
|
|
stmt ::= raise_stmt2
|
|
raise_stmt2 ::= expr expr RAISE_VARARGS_2
|
|
""",
|
|
nop_func,
|
|
)
|
|
custom_ops_processed.add(opname)
|
|
|
|
elif opname == "SETUP_EXCEPT":
|
|
self.addRule(
|
|
"""
|
|
try_except ::= SETUP_EXCEPT suite_stmts_opt POP_BLOCK
|
|
except_handler opt_come_from_except
|
|
|
|
tryelsestmt ::= SETUP_EXCEPT suite_stmts_opt POP_BLOCK
|
|
except_handler else_suite come_from_except_clauses
|
|
|
|
tryelsestmt ::= SETUP_EXCEPT suite_stmts_opt POP_BLOCK
|
|
except_handler else_suite come_froms
|
|
|
|
tryelsestmtl ::= SETUP_EXCEPT suite_stmts_opt POP_BLOCK
|
|
except_handler else_suitel come_from_except_clauses
|
|
|
|
stmt ::= tryelsestmtl3
|
|
tryelsestmtl3 ::= SETUP_EXCEPT suite_stmts_opt POP_BLOCK
|
|
except_handler COME_FROM else_suitel
|
|
opt_come_from_except
|
|
""",
|
|
nop_func,
|
|
)
|
|
custom_ops_processed.add(opname)
|
|
|
|
elif opname == "SETUP_WITH":
|
|
rules_str = """
|
|
stmt ::= withstmt
|
|
stmt ::= withasstmt
|
|
|
|
withstmt ::= expr SETUP_WITH POP_TOP suite_stmts_opt COME_FROM_WITH
|
|
WITH_CLEANUP_START WITH_CLEANUP_FINISH END_FINALLY
|
|
withasstmt ::= expr SETUP_WITH store suite_stmts_opt COME_FROM_WITH
|
|
WITH_CLEANUP_START WITH_CLEANUP_FINISH END_FINALLY
|
|
|
|
withstmt ::= expr
|
|
SETUP_WITH POP_TOP suite_stmts_opt
|
|
POP_BLOCK LOAD_CONST COME_FROM_WITH
|
|
WITH_CLEANUP_START WITH_CLEANUP_FINISH END_FINALLY
|
|
withasstmt ::= expr
|
|
SETUP_WITH store suite_stmts_opt
|
|
POP_BLOCK LOAD_CONST COME_FROM_WITH
|
|
WITH_CLEANUP_START WITH_CLEANUP_FINISH END_FINALLY
|
|
|
|
withstmt ::= expr
|
|
SETUP_WITH POP_TOP suite_stmts_opt
|
|
POP_BLOCK LOAD_CONST COME_FROM_WITH
|
|
WITH_CLEANUP_START WITH_CLEANUP_FINISH END_FINALLY
|
|
withasstmt ::= expr
|
|
SETUP_WITH store suite_stmts_opt
|
|
POP_BLOCK LOAD_CONST COME_FROM_WITH
|
|
WITH_CLEANUP_START WITH_CLEANUP_FINISH END_FINALLY
|
|
"""
|
|
if self.version < 3.8:
|
|
rules_str += """
|
|
withstmt ::= expr SETUP_WITH POP_TOP suite_stmts_opt POP_BLOCK
|
|
LOAD_CONST
|
|
WITH_CLEANUP_START WITH_CLEANUP_FINISH END_FINALLY
|
|
"""
|
|
else:
|
|
rules_str += """
|
|
withstmt ::= expr
|
|
SETUP_WITH POP_TOP suite_stmts_opt
|
|
POP_BLOCK LOAD_CONST COME_FROM_WITH
|
|
WITH_CLEANUP_START WITH_CLEANUP_FINISH END_FINALLY
|
|
|
|
withasstmt ::= expr
|
|
SETUP_WITH store suite_stmts_opt
|
|
POP_BLOCK LOAD_CONST COME_FROM_WITH
|
|
|
|
withstmt ::= expr SETUP_WITH POP_TOP suite_stmts_opt POP_BLOCK
|
|
BEGIN_FINALLY COME_FROM_WITH
|
|
WITH_CLEANUP_START WITH_CLEANUP_FINISH
|
|
END_FINALLY
|
|
"""
|
|
self.addRule(rules_str, nop_func)
|
|
|
|
elif opname_base in ("UNPACK_EX",):
|
|
before_count, after_count = token.attr
|
|
rule = (
|
|
"unpack ::= " + opname + " store" * (before_count + after_count + 1)
|
|
)
|
|
self.addRule(rule, nop_func)
|
|
|
|
elif opname_base in ("UNPACK_TUPLE", "UNPACK_SEQUENCE"):
|
|
rule = "unpack ::= " + opname + " store" * token.attr
|
|
self.addRule(rule, nop_func)
|
|
|
|
elif opname_base == "UNPACK_LIST":
|
|
rule = "unpack_list ::= " + opname + " store" * token.attr
|
|
self.addRule(rule, nop_func)
|
|
custom_ops_processed.add(opname)
|
|
pass
|
|
|
|
pass
|
|
|
|
self.check_reduce["and"] = "AST"
|
|
self.check_reduce["aug_assign1"] = "AST"
|
|
self.check_reduce["aug_assign2"] = "AST"
|
|
self.check_reduce["while1stmt"] = "noAST"
|
|
self.check_reduce["while1elsestmt"] = "noAST"
|
|
self.check_reduce["_ifstmts_jump"] = "AST"
|
|
self.check_reduce["ifelsestmt"] = "AST"
|
|
self.check_reduce["iflaststmt"] = "AST"
|
|
self.check_reduce["iflaststmtl"] = "AST"
|
|
self.check_reduce["ifstmt"] = "AST"
|
|
self.check_reduce["ifstmtl"] = "AST"
|
|
self.check_reduce["annotate_tuple"] = "noAST"
|
|
|
|
# FIXME: remove parser errors caused by the below
|
|
# self.check_reduce['while1elsestmt'] = 'noAST'
|
|
|
|
return
|
|
|
|
def custom_classfunc_rule(self, opname, token, customize, next_token):
|
|
"""
|
|
call ::= expr {expr}^n CALL_FUNCTION_n
|
|
call ::= expr {expr}^n CALL_FUNCTION_VAR_n
|
|
call ::= expr {expr}^n CALL_FUNCTION_VAR_KW_n
|
|
call ::= expr {expr}^n CALL_FUNCTION_KW_n
|
|
|
|
classdefdeco2 ::= LOAD_BUILD_CLASS mkfunc {expr}^n-1 CALL_FUNCTION_n
|
|
"""
|
|
args_pos, args_kw = self.get_pos_kw(token)
|
|
|
|
# Additional exprs for * and ** args:
|
|
# 0 if neither
|
|
# 1 for CALL_FUNCTION_VAR or CALL_FUNCTION_KW
|
|
# 2 for * and ** args (CALL_FUNCTION_VAR_KW).
|
|
# Yes, this computation based on instruction name is a little bit hoaky.
|
|
nak = (len(opname) - len("CALL_FUNCTION")) // 3
|
|
uniq_param = args_kw + args_pos
|
|
|
|
if frozenset(("GET_AWAITABLE", "YIELD_FROM")).issubset(self.seen_ops):
|
|
rule = (
|
|
"async_call ::= expr "
|
|
+ ("pos_arg " * args_pos)
|
|
+ ("kwarg " * args_kw)
|
|
+ "expr " * nak
|
|
+ token.kind
|
|
+ " GET_AWAITABLE LOAD_CONST YIELD_FROM"
|
|
)
|
|
self.add_unique_rule(rule, token.kind, uniq_param, customize)
|
|
self.add_unique_rule(
|
|
"expr ::= async_call", token.kind, uniq_param, customize
|
|
)
|
|
|
|
if opname.startswith("CALL_FUNCTION_VAR"):
|
|
token.kind = self.call_fn_name(token)
|
|
if opname.endswith("KW"):
|
|
kw = "expr "
|
|
else:
|
|
kw = ""
|
|
rule = (
|
|
"call ::= expr expr "
|
|
+ ("pos_arg " * args_pos)
|
|
+ ("kwarg " * args_kw)
|
|
+ kw
|
|
+ token.kind
|
|
)
|
|
|
|
# Note: semantic actions make use of the fact of wheter "args_pos"
|
|
# zero or not in creating a template rule.
|
|
self.add_unique_rule(rule, token.kind, args_pos, customize)
|
|
else:
|
|
token.kind = self.call_fn_name(token)
|
|
uniq_param = args_kw + args_pos
|
|
|
|
# Note: 3.5+ have subclassed this method; so we don't handle
|
|
# 'CALL_FUNCTION_VAR' or 'CALL_FUNCTION_EX' here.
|
|
rule = (
|
|
"call ::= expr "
|
|
+ ("pos_arg " * args_pos)
|
|
+ ("kwarg " * args_kw)
|
|
+ "expr " * nak
|
|
+ token.kind
|
|
)
|
|
|
|
self.add_unique_rule(rule, token.kind, uniq_param, customize)
|
|
|
|
if "LOAD_BUILD_CLASS" in self.seen_ops:
|
|
if (
|
|
next_token == "CALL_FUNCTION"
|
|
and next_token.attr == 1
|
|
and args_pos > 1
|
|
):
|
|
rule = "classdefdeco2 ::= LOAD_BUILD_CLASS mkfunc %s%s_%d" % (
|
|
("expr " * (args_pos - 1)),
|
|
opname,
|
|
args_pos,
|
|
)
|
|
self.add_unique_rule(rule, token.kind, uniq_param, customize)
|
|
|
|
def reduce_is_invalid(self, rule, ast, tokens, first, last):
|
|
lhs = rule[0]
|
|
n = len(tokens)
|
|
|
|
if lhs == "and" and ast:
|
|
# FIXME: put in a routine somewhere
|
|
jmp = ast[1]
|
|
if jmp.kind.startswith("jmp_"):
|
|
if last == n:
|
|
return True
|
|
jmp_target = jmp[0].attr
|
|
jmp_offset = jmp[0].offset
|
|
|
|
if tokens[first].off2int() <= jmp_target < tokens[last].off2int():
|
|
return True
|
|
if rule == ("and", ("expr", "jmp_false", "expr", "jmp_false")):
|
|
jmp2_target = ast[3][0].attr
|
|
return jmp_target != jmp2_target
|
|
elif rule == ("and", ("expr", "jmp_false", "expr")):
|
|
if tokens[last] == "POP_JUMP_IF_FALSE":
|
|
# Ok if jump it doesn't jump to last instruction
|
|
return jmp_target != tokens[last].attr
|
|
elif tokens[last] == "JUMP_IF_TRUE_OR_POP":
|
|
# Ok if jump it does jump right after last instruction
|
|
return jmp_target + 2 != tokens[last].attr
|
|
elif rule == ("and", ("expr", "jmp_false", "expr", "COME_FROM")):
|
|
return ast[-1].attr != jmp_offset
|
|
# elif rule == ("and", ("expr", "jmp_false", "expr", "COME_FROM")):
|
|
# return jmp_offset != tokens[first+3].attr
|
|
|
|
return jmp_target != tokens[last].off2int()
|
|
return False
|
|
|
|
elif lhs in ("aug_assign1", "aug_assign2") and ast[0][0] == "and":
|
|
return True
|
|
elif lhs == "annotate_tuple":
|
|
return not isinstance(tokens[first].attr, tuple)
|
|
elif lhs == "while1elsestmt":
|
|
|
|
if last == n:
|
|
# Adjust for fuzziness in parsing
|
|
last -= 1
|
|
|
|
if tokens[last] == "COME_FROM_LOOP":
|
|
last -= 1
|
|
elif tokens[last - 1] == "COME_FROM_LOOP":
|
|
last -= 2
|
|
if tokens[last] in ("JUMP_BACK", "CONTINUE"):
|
|
# These indicate inside a loop, but token[last]
|
|
# should not be in a loop.
|
|
# FIXME: Not quite right: refine by using target
|
|
return True
|
|
|
|
# if SETUP_LOOP target spans the else part, then this is
|
|
# not while1else. Also do for whileTrue?
|
|
last += 1
|
|
# 3.8+ Doesn't have SETUP_LOOP
|
|
return self.version < 3.8 and tokens[first].attr > tokens[last].off2int()
|
|
|
|
elif lhs == "while1stmt":
|
|
|
|
# If there is a fall through to the COME_FROM_LOOP, then this is
|
|
# not a while 1. So the instruction before should either be a
|
|
# JUMP_BACK or the instruction before should not be the target of a
|
|
# jump. (Well that last clause i not quite right; that target could be
|
|
# from dead code. Ugh. We need a more uniform control flow analysis.)
|
|
if last == n or tokens[last - 1] == "COME_FROM_LOOP":
|
|
cfl = last - 1
|
|
else:
|
|
cfl = last
|
|
assert tokens[cfl] == "COME_FROM_LOOP"
|
|
|
|
for i in range(cfl - 1, first, -1):
|
|
if tokens[i] != "POP_BLOCK":
|
|
break
|
|
if tokens[i].kind not in ("JUMP_BACK", "RETURN_VALUE"):
|
|
if not tokens[i].kind.startswith("COME_FROM"):
|
|
return True
|
|
|
|
# Check that the SETUP_LOOP jumps to the offset after the
|
|
# COME_FROM_LOOP
|
|
if 0 <= last < n and tokens[last] in ("COME_FROM_LOOP", "JUMP_BACK"):
|
|
# jump_back should be right before COME_FROM_LOOP?
|
|
last += 1
|
|
if last == n:
|
|
last -= 1
|
|
offset = tokens[last].off2int()
|
|
assert tokens[first] == "SETUP_LOOP"
|
|
if offset != tokens[first].attr:
|
|
return True
|
|
return False
|
|
elif lhs == "_ifstmts_jump" and len(rule[1]) > 1 and ast:
|
|
come_froms = ast[-1]
|
|
# Make sure all of the "come froms" offset at the
|
|
# end of the "if" come from somewhere inside the "if".
|
|
# Since the come_froms are ordered so that lowest
|
|
# offset COME_FROM is last, it is sufficient to test
|
|
# just the last one.
|
|
|
|
# This is complicated, but note that the JUMP_IF instruction comes immediately
|
|
# *before* _ifstmts_jump so that's what we have to test
|
|
# the COME_FROM against. This can be complicated by intervening
|
|
# POP_TOP, and pseudo COME_FROM, ELSE instructions
|
|
#
|
|
pop_jump_index = first - 1
|
|
while pop_jump_index > 0 and tokens[pop_jump_index] in (
|
|
"ELSE",
|
|
"POP_TOP",
|
|
"JUMP_FORWARD",
|
|
"COME_FROM",
|
|
):
|
|
pop_jump_index -= 1
|
|
come_froms = ast[-1]
|
|
|
|
# FIXME: something is fishy when and EXTENDED ARG is needed before the
|
|
# pop_jump_index instruction to get the argment. In this case, the
|
|
# _ifsmtst_jump can jump to a spot beyond the come_froms.
|
|
# That is going on in the non-EXTENDED_ARG case is that the POP_JUMP_IF
|
|
# jumps to a JUMP_(FORWARD) which is changed into an EXTENDED_ARG POP_JUMP_IF
|
|
# to the jumped forwareded address
|
|
if tokens[pop_jump_index].attr > 256:
|
|
return False
|
|
|
|
if isinstance(come_froms, Token):
|
|
return (
|
|
come_froms.attr is not None
|
|
and tokens[pop_jump_index].offset > come_froms.attr
|
|
)
|
|
|
|
elif len(come_froms) == 0:
|
|
return False
|
|
else:
|
|
return tokens[pop_jump_index].offset > come_froms[-1].attr
|
|
|
|
elif lhs in ("ifstmt", "ifstmtl"):
|
|
# FIXME: put in a routine somewhere
|
|
|
|
n = len(tokens)
|
|
if lhs == "ifstmtl":
|
|
if last == n:
|
|
last -= 1
|
|
pass
|
|
if (tokens[last].attr and isinstance(tokens[last].attr, int)):
|
|
return tokens[first].offset < tokens[last].attr
|
|
pass
|
|
|
|
# Make sure jumps don't extend beyond the end of the if statement.
|
|
l = last
|
|
if l == n:
|
|
l -= 1
|
|
if isinstance(tokens[l].offset, str):
|
|
last_offset = int(tokens[l].offset.split("_")[0], 10)
|
|
else:
|
|
last_offset = tokens[l].offset
|
|
for i in range(first, l):
|
|
t = tokens[i]
|
|
if t.kind == "POP_JUMP_IF_FALSE":
|
|
if t.attr > last_offset:
|
|
return True
|
|
pass
|
|
pass
|
|
pass
|
|
|
|
if ast:
|
|
testexpr = ast[0]
|
|
|
|
if (last + 1) < n and tokens[last + 1] == "COME_FROM_LOOP":
|
|
# iflastsmtl jumped outside of loop. No good.
|
|
return True
|
|
|
|
if testexpr[0] in ("testtrue", "testfalse"):
|
|
test = testexpr[0]
|
|
if len(test) > 1 and test[1].kind.startswith("jmp_"):
|
|
if last == n:
|
|
last -= 1
|
|
jmp_target = test[1][0].attr
|
|
if tokens[first].off2int() <= jmp_target < tokens[last].off2int():
|
|
return True
|
|
# jmp_target less than tokens[first] is okay - is to a loop
|
|
# jmp_target equal tokens[last] is also okay: normal non-optimized non-loop jump
|
|
if jmp_target > tokens[last].off2int():
|
|
# One more weird case to look out for
|
|
# if c1:
|
|
# if c2: # Jumps around the *outer* "else"
|
|
# ...
|
|
# else:
|
|
if jmp_target == tokens[last - 1].attr:
|
|
return False
|
|
if last < n and tokens[last].kind.startswith("JUMP"):
|
|
return False
|
|
return True
|
|
|
|
pass
|
|
pass
|
|
return False
|
|
elif lhs in ("iflaststmt", "iflaststmtl") and ast:
|
|
# FIXME: put in a routine somewhere
|
|
testexpr = ast[0]
|
|
|
|
if testexpr[0] in ("testtrue", "testfalse"):
|
|
|
|
test = testexpr[0]
|
|
if len(test) > 1 and test[1].kind.startswith("jmp_"):
|
|
if last == n:
|
|
last -= 1
|
|
jmp_target = test[1][0].attr
|
|
if tokens[first].off2int() <= jmp_target < tokens[last].off2int():
|
|
return True
|
|
# jmp_target less than tokens[first] is okay - is to a loop
|
|
# jmp_target equal tokens[last] is also okay: normal non-optimized non-loop jump
|
|
|
|
if (last + 1) < n and tokens[last - 1] != "JUMP_BACK" and tokens[last + 1] == "COME_FROM_LOOP":
|
|
# iflastsmtl is not at the end of a loop, but jumped outside of loop. No good.
|
|
# FIXME: check that tokens[last] == "POP_BLOCK"? Or allow for it not to appear?
|
|
return True
|
|
|
|
# If the instruction before "first" is a "POP_JUMP_IF_FALSE" which goes
|
|
# to the same target as jmp_target, then this not nested "if .. if .."
|
|
# but rather "if ... and ..."
|
|
if first > 0 and tokens[first - 1] == "POP_JUMP_IF_FALSE":
|
|
return tokens[first - 1].attr == jmp_target
|
|
|
|
if jmp_target > tokens[last].off2int():
|
|
# One more weird case to look out for
|
|
# if c1:
|
|
# if c2: # Jumps around the *outer* "else"
|
|
# ...
|
|
# else:
|
|
if jmp_target == tokens[last - 1].attr:
|
|
return False
|
|
if last < n and tokens[last].kind.startswith("JUMP"):
|
|
return False
|
|
return True
|
|
|
|
pass
|
|
return False
|
|
|
|
# FIXME: put in a routine somewhere
|
|
elif lhs == "ifelsestmt":
|
|
|
|
if (last + 1) < n and tokens[last + 1] == "COME_FROM_LOOP":
|
|
# ifelsestmt jumped outside of loop. No good.
|
|
return True
|
|
|
|
if rule not in (
|
|
(
|
|
"ifelsestmt",
|
|
(
|
|
"testexpr",
|
|
"c_stmts_opt",
|
|
"jump_forward_else",
|
|
"else_suite",
|
|
"_come_froms",
|
|
),
|
|
),
|
|
(
|
|
"ifelsestmt",
|
|
(
|
|
"testexpr",
|
|
"c_stmts_opt",
|
|
"jf_cfs",
|
|
"else_suite",
|
|
"opt_come_from_except",
|
|
),
|
|
),
|
|
):
|
|
return False
|
|
|
|
# Make sure all of the "come froms" offset at the
|
|
# end of the "if" come from somewhere inside the "if".
|
|
# Since the come_froms are ordered so that lowest
|
|
# offset COME_FROM is last, it is sufficient to test
|
|
# just the last one.
|
|
come_froms = ast[-1]
|
|
if come_froms == "opt_come_from_except" and len(come_froms) > 0:
|
|
come_froms = come_froms[0]
|
|
if not isinstance(come_froms, Token):
|
|
return tokens[first].offset > come_froms[-1].attr
|
|
elif tokens[first].offset > come_froms.attr:
|
|
return True
|
|
|
|
# For mysterious reasons a COME_FROM in tokens[last+1] might be part of the grammar rule
|
|
# even though it is not found in come_froms.
|
|
# Work around this.
|
|
if (
|
|
last < n
|
|
and tokens[last] == "COME_FROM"
|
|
and tokens[first].offset > tokens[last].attr
|
|
):
|
|
return True
|
|
|
|
testexpr = ast[0]
|
|
|
|
# Check that the condition portion of the "if"
|
|
# jumps to the "else" part.
|
|
# Compare with parse30.py of uncompyle6
|
|
if testexpr[0] in ("testtrue", "testfalse"):
|
|
test = testexpr[0]
|
|
if len(test) > 1 and test[1].kind.startswith("jmp_"):
|
|
if last == n:
|
|
last -= 1
|
|
jmp = test[1]
|
|
jmp_target = jmp[0].attr
|
|
if tokens[first].off2int() > jmp_target:
|
|
return True
|
|
return (jmp_target > tokens[last].off2int()) and tokens[
|
|
last
|
|
] != "JUMP_FORWARD"
|
|
|
|
return False
|
|
|
|
return False
|