import types as pytypes # avoid confusion with numba.types import copy import ctypes import numba.core.analysis from numba.core import utils, types, typing, errors, ir, rewrites, config, ir_utils from numba import prange from numba.parfors.parfor import internal_prange from numba.core.ir_utils import ( mk_unique_var, next_label, add_offset_to_labels, replace_vars, remove_dels, rename_labels, find_topo_order, merge_adjacent_blocks, GuardException, require, guard, get_definition, find_callname, find_build_sequence, get_np_ufunc_typ, get_ir_of_code, simplify_CFG, canonicalize_array_math, dead_code_elimination, ) from numba.core.analysis import ( compute_cfg_from_blocks, compute_use_defs, compute_live_variables) from numba.core import postproc from numba.np.unsafe.ndarray import empty_inferred as unsafe_empty_inferred import numpy as np import operator import numba.misc.special """ Variable enable_inline_arraycall is only used for testing purpose. """ enable_inline_arraycall = True def callee_ir_validator(func_ir): """Checks the IR of a callee is supported for inlining """ for blk in func_ir.blocks.values(): for stmt in blk.find_insts(ir.Assign): if isinstance(stmt.value, ir.Yield): msg = "The use of yield in a closure is unsupported." raise errors.UnsupportedError(msg, loc=stmt.loc) def _created_inlined_var_name(function_name, var_name): """Creates a name for an inlined variable based on the function name and the variable name. It does this "safely" to avoid the use of characters that are illegal in python variable names as there are occasions when function generation needs valid python name tokens.""" inlined_name = f'{function_name}.{var_name}' # Replace angle brackets, e.g. "" is replaced with "_locals_" new_name = inlined_name.replace('<', '_').replace('>', '_') # The version "version" of the closure function e.g. foo$2 (id 2) is # rewritten as "foo_v2". Further "." is also replaced with "_". new_name = new_name.replace('.', '_').replace('$', '_v') return new_name class InlineClosureCallPass(object): """InlineClosureCallPass class looks for direct calls to locally defined closures, and inlines the body of the closure function to the call site. """ def __init__(self, func_ir, parallel_options, swapped={}, typed=False): self.func_ir = func_ir self.parallel_options = parallel_options self.swapped = swapped self._processed_stencils = [] self.typed = typed def run(self): """Run inline closure call pass. """ # Analysis relies on ir.Del presence, strip out later pp = postproc.PostProcessor(self.func_ir) pp.run(True) modified = False work_list = list(self.func_ir.blocks.items()) debug_print = _make_debug_print("InlineClosureCallPass") debug_print("START") while work_list: label, block = work_list.pop() for i, instr in enumerate(block.body): if isinstance(instr, ir.Assign): lhs = instr.target expr = instr.value if isinstance(expr, ir.Expr) and expr.op == 'call': call_name = guard(find_callname, self.func_ir, expr) func_def = guard(get_definition, self.func_ir, expr.func) if guard(self._inline_reduction, work_list, block, i, expr, call_name): modified = True break # because block structure changed if guard(self._inline_closure, work_list, block, i, func_def): modified = True break # because block structure changed if guard(self._inline_stencil, instr, call_name, func_def): modified = True if enable_inline_arraycall: # Identify loop structure if modified: # Need to do some cleanups if closure inlining kicked in merge_adjacent_blocks(self.func_ir.blocks) cfg = compute_cfg_from_blocks(self.func_ir.blocks) debug_print("start inline arraycall") _debug_dump(cfg) loops = cfg.loops() sized_loops = [(k, len(loops[k].body)) for k in loops.keys()] visited = [] # We go over all loops, bigger loops first (outer first) for k, s in sorted(sized_loops, key=lambda tup: tup[1], reverse=True): visited.append(k) if guard(_inline_arraycall, self.func_ir, cfg, visited, loops[k], self.swapped, self.parallel_options.comprehension, self.typed): modified = True if modified: _fix_nested_array(self.func_ir) if modified: # clean up now dead/unreachable blocks, e.g. unconditionally raising # an exception in an inlined function would render some parts of the # inliner unreachable cfg = compute_cfg_from_blocks(self.func_ir.blocks) for dead in cfg.dead_nodes(): del self.func_ir.blocks[dead] # run dead code elimination dead_code_elimination(self.func_ir) # do label renaming self.func_ir.blocks = rename_labels(self.func_ir.blocks) # inlining done, strip dels remove_dels(self.func_ir.blocks) debug_print("END") def _inline_reduction(self, work_list, block, i, expr, call_name): # only inline reduction in sequential execution, parallel handling # is done in ParforPass. require(not self.parallel_options.reduction) require(call_name == ('reduce', 'builtins') or call_name == ('reduce', '_functools')) if len(expr.args) not in (2, 3): raise TypeError("invalid reduce call, " "two arguments are required (optional initial " "value can also be specified)") check_reduce_func(self.func_ir, expr.args[0]) def reduce_func(f, A, v=None): it = iter(A) if v is not None: s = v else: s = next(it) for a in it: s = f(s, a) return s inline_closure_call(self.func_ir, self.func_ir.func_id.func.__globals__, block, i, reduce_func, work_list=work_list, callee_validator=callee_ir_validator) return True def _inline_stencil(self, instr, call_name, func_def): from numba.stencils.stencil import StencilFunc lhs = instr.target expr = instr.value # We keep the escaping variables of the stencil kernel # alive by adding them to the actual kernel call as extra # keyword arguments, which is ignored anyway. if (isinstance(func_def, ir.Global) and func_def.name == 'stencil' and isinstance(func_def.value, StencilFunc)): if expr.kws: expr.kws += func_def.value.kws else: expr.kws = func_def.value.kws return True # Otherwise we proceed to check if it is a call to numba.stencil require(call_name == ('stencil', 'numba.stencils.stencil') or call_name == ('stencil', 'numba')) require(expr not in self._processed_stencils) self._processed_stencils.append(expr) if not len(expr.args) == 1: raise ValueError("As a minimum Stencil requires" " a kernel as an argument") stencil_def = guard(get_definition, self.func_ir, expr.args[0]) require(isinstance(stencil_def, ir.Expr) and stencil_def.op == "make_function") kernel_ir = get_ir_of_code(self.func_ir.func_id.func.__globals__, stencil_def.code) options = dict(expr.kws) if 'neighborhood' in options: fixed = guard(self._fix_stencil_neighborhood, options) if not fixed: raise ValueError("stencil neighborhood option should be a tuple" " with constant structure such as ((-w, w),)") if 'index_offsets' in options: fixed = guard(self._fix_stencil_index_offsets, options) if not fixed: raise ValueError("stencil index_offsets option should be a tuple" " with constant structure such as (offset, )") sf = StencilFunc(kernel_ir, 'constant', options) sf.kws = expr.kws # hack to keep variables live sf_global = ir.Global('stencil', sf, expr.loc) self.func_ir._definitions[lhs.name] = [sf_global] instr.value = sf_global return True def _fix_stencil_neighborhood(self, options): """ Extract the two-level tuple representing the stencil neighborhood from the program IR to provide a tuple to StencilFunc. """ # build_tuple node with neighborhood for each dimension dims_build_tuple = get_definition(self.func_ir, options['neighborhood']) require(hasattr(dims_build_tuple, 'items')) res = [] for window_var in dims_build_tuple.items: win_build_tuple = get_definition(self.func_ir, window_var) require(hasattr(win_build_tuple, 'items')) res.append(tuple(win_build_tuple.items)) options['neighborhood'] = tuple(res) return True def _fix_stencil_index_offsets(self, options): """ Extract the tuple representing the stencil index offsets from the program IR to provide to StencilFunc. """ offset_tuple = get_definition(self.func_ir, options['index_offsets']) require(hasattr(offset_tuple, 'items')) options['index_offsets'] = tuple(offset_tuple.items) return True def _inline_closure(self, work_list, block, i, func_def): require(isinstance(func_def, ir.Expr) and func_def.op == "make_function") inline_closure_call(self.func_ir, self.func_ir.func_id.func.__globals__, block, i, func_def, work_list=work_list, callee_validator=callee_ir_validator) return True def check_reduce_func(func_ir, func_var): """Checks the function at func_var in func_ir to make sure it's amenable for inlining. Returns the function itself""" reduce_func = guard(get_definition, func_ir, func_var) if reduce_func is None: raise ValueError("Reduce function cannot be found for njit \ analysis") if isinstance(reduce_func, (ir.FreeVar, ir.Global)): if not isinstance(reduce_func.value, numba.core.registry.CPUDispatcher): raise ValueError("Invalid reduction function") # pull out the python function for inlining reduce_func = reduce_func.value.py_func elif not (hasattr(reduce_func, 'code') or hasattr(reduce_func, '__code__')): raise ValueError("Invalid reduction function") f_code = (reduce_func.code if hasattr(reduce_func, 'code') else reduce_func.__code__) if not f_code.co_argcount == 2: raise TypeError("Reduction function should take 2 arguments") return reduce_func class InlineWorker(object): """ A worker class for inlining, this is a more advanced version of `inline_closure_call` in that it permits inlining from function type, Numba IR and code object. It also, runs the entire untyped compiler pipeline on the inlinee to ensure that it is transformed as though it were compiled directly. """ def __init__(self, typingctx=None, targetctx=None, locals=None, pipeline=None, flags=None, validator=callee_ir_validator, typemap=None, calltypes=None): """ Instantiate a new InlineWorker, all arguments are optional though some must be supplied together for certain use cases. The methods will refuse to run if the object isn't configured in the manner needed. Args are the same as those in a numba.core.Compiler.state, except the validator which is a function taking Numba IR and validating it for use when inlining (this is optional and really to just provide better error messages about things which the inliner cannot handle like yield in closure). """ def check(arg, name): if arg is None: raise TypeError("{} must not be None".format(name)) from numba.core.compiler import DefaultPassBuilder # check the stuff needed to run the more advanced compilation pipeline # is valid if any of it is provided compiler_args = (targetctx, locals, pipeline, flags) compiler_group = [x is not None for x in compiler_args] if any(compiler_group) and not all(compiler_group): check(targetctx, 'targetctx') check(locals, 'locals') check(pipeline, 'pipeline') check(flags, 'flags') elif all(compiler_group): check(typingctx, 'typingctx') self._compiler_pipeline = DefaultPassBuilder.define_untyped_pipeline self.typingctx = typingctx self.targetctx = targetctx self.locals = locals self.pipeline = pipeline self.flags = flags self.validator = validator self.debug_print = _make_debug_print("InlineWorker") # check whether this inliner can also support typemap and calltypes # update and if what's provided is valid pair = (typemap, calltypes) pair_is_none = [x is None for x in pair] if any(pair_is_none) and not all(pair_is_none): msg = ("typemap and calltypes must both be either None or have a " "value, got: %s, %s") raise TypeError(msg % pair) self._permit_update_type_and_call_maps = not all(pair_is_none) self.typemap = typemap self.calltypes = calltypes def inline_ir(self, caller_ir, block, i, callee_ir, callee_freevars, arg_typs=None): """ Inlines the callee_ir in the caller_ir at statement index i of block `block`, callee_freevars are the free variables for the callee_ir. If the callee_ir is derived from a function `func` then this is `func.__code__.co_freevars`. If `arg_typs` is given and the InlineWorker instance was initialized with a typemap and calltypes then they will be appropriately updated based on the arg_typs. """ # Always copy the callee IR, it gets mutated def copy_ir(the_ir): kernel_copy = the_ir.copy() kernel_copy.blocks = {} for block_label, block in the_ir.blocks.items(): new_block = copy.deepcopy(the_ir.blocks[block_label]) new_block.body = [] for stmt in the_ir.blocks[block_label].body: scopy = copy.deepcopy(stmt) new_block.body.append(scopy) kernel_copy.blocks[block_label] = new_block return kernel_copy callee_ir = copy_ir(callee_ir) # check that the contents of the callee IR is something that can be # inlined if a validator is present if self.validator is not None: self.validator(callee_ir) # save an unmutated copy of the callee_ir to return callee_ir_original = callee_ir.copy() scope = block.scope instr = block.body[i] call_expr = instr.value callee_blocks = callee_ir.blocks # 1. relabel callee_ir by adding an offset max_label = max(ir_utils._the_max_label.next(), max(caller_ir.blocks.keys())) callee_blocks = add_offset_to_labels(callee_blocks, max_label + 1) callee_blocks = simplify_CFG(callee_blocks) callee_ir.blocks = callee_blocks min_label = min(callee_blocks.keys()) max_label = max(callee_blocks.keys()) # reset globals in ir_utils before we use it ir_utils._the_max_label.update(max_label) self.debug_print("After relabel") _debug_dump(callee_ir) # 2. rename all local variables in callee_ir with new locals created in # caller_ir callee_scopes = _get_all_scopes(callee_blocks) self.debug_print("callee_scopes = ", callee_scopes) # one function should only have one local scope assert(len(callee_scopes) == 1) callee_scope = callee_scopes[0] var_dict = {} for var in callee_scope.localvars._con.values(): if not (var.name in callee_freevars): inlined_name = _created_inlined_var_name( callee_ir.func_id.unique_name, var.name) new_var = scope.redefine(inlined_name, loc=var.loc) var_dict[var.name] = new_var self.debug_print("var_dict = ", var_dict) replace_vars(callee_blocks, var_dict) self.debug_print("After local var rename") _debug_dump(callee_ir) # 3. replace formal parameters with actual arguments callee_func = callee_ir.func_id.func args = _get_callee_args(call_expr, callee_func, block.body[i].loc, caller_ir) # 4. Update typemap if self._permit_update_type_and_call_maps: if arg_typs is None: raise TypeError('arg_typs should have a value not None') self.update_type_and_call_maps(callee_ir, arg_typs) # update_type_and_call_maps replaces blocks callee_blocks = callee_ir.blocks self.debug_print("After arguments rename: ") _debug_dump(callee_ir) _replace_args_with(callee_blocks, args) # 5. split caller blocks into two new_blocks = [] new_block = ir.Block(scope, block.loc) new_block.body = block.body[i + 1:] new_label = next_label() caller_ir.blocks[new_label] = new_block new_blocks.append((new_label, new_block)) block.body = block.body[:i] block.body.append(ir.Jump(min_label, instr.loc)) # 6. replace Return with assignment to LHS topo_order = find_topo_order(callee_blocks) _replace_returns(callee_blocks, instr.target, new_label) # remove the old definition of instr.target too if (instr.target.name in caller_ir._definitions and call_expr in caller_ir._definitions[instr.target.name]): # NOTE: target can have multiple definitions due to control flow caller_ir._definitions[instr.target.name].remove(call_expr) # 7. insert all new blocks, and add back definitions for label in topo_order: # block scope must point to parent's block = callee_blocks[label] block.scope = scope _add_definitions(caller_ir, block) caller_ir.blocks[label] = block new_blocks.append((label, block)) self.debug_print("After merge in") _debug_dump(caller_ir) return callee_ir_original, callee_blocks, var_dict, new_blocks def inline_function(self, caller_ir, block, i, function, arg_typs=None): """ Inlines the function in the caller_ir at statement index i of block `block`. If `arg_typs` is given and the InlineWorker instance was initialized with a typemap and calltypes then they will be appropriately updated based on the arg_typs. """ callee_ir = self.run_untyped_passes(function) freevars = function.__code__.co_freevars return self.inline_ir(caller_ir, block, i, callee_ir, freevars, arg_typs=arg_typs) def run_untyped_passes(self, func, enable_ssa=False): """ Run the compiler frontend's untyped passes over the given Python function, and return the function's canonical Numba IR. Disable SSA transformation by default, since the call site won't be in SSA form and self.inline_ir depends on this being the case. """ from numba.core.compiler import StateDict, _CompileStatus from numba.core.untyped_passes import ExtractByteCode, WithLifting from numba.core import bytecode from numba.parfors.parfor import ParforDiagnostics state = StateDict() state.func_ir = None state.typingctx = self.typingctx state.targetctx = self.targetctx state.locals = self.locals state.pipeline = self.pipeline state.flags = self.flags state.flags.enable_ssa = enable_ssa state.func_id = bytecode.FunctionIdentity.from_function(func) state.typemap = None state.calltypes = None state.type_annotation = None state.status = _CompileStatus(False) state.return_type = None state.parfor_diagnostics = ParforDiagnostics() state.metadata = {} ExtractByteCode().run_pass(state) # This is a lie, just need *some* args for the case where an obj mode # with lift is needed state.args = len(state.bc.func_id.pysig.parameters) * (types.pyobject,) pm = self._compiler_pipeline(state) pm.finalize() pm.run(state) return state.func_ir def update_type_and_call_maps(self, callee_ir, arg_typs): """ Updates the type and call maps based on calling callee_ir with arguments from arg_typs""" from numba.core.ssa import reconstruct_ssa from numba.core.typed_passes import PreLowerStripPhis if not self._permit_update_type_and_call_maps: msg = ("InlineWorker instance not configured correctly, typemap or " "calltypes missing in initialization.") raise ValueError(msg) from numba.core import typed_passes # call branch pruning to simplify IR and avoid inference errors callee_ir._definitions = ir_utils.build_definitions(callee_ir.blocks) numba.core.analysis.dead_branch_prune(callee_ir, arg_typs) # callee's typing may require SSA callee_ir = reconstruct_ssa(callee_ir) callee_ir._definitions = ir_utils.build_definitions(callee_ir.blocks) f_typemap, f_return_type, f_calltypes, _ = typed_passes.type_inference_stage( self.typingctx, self.targetctx, callee_ir, arg_typs, None) callee_ir = PreLowerStripPhis()._strip_phi_nodes(callee_ir) callee_ir._definitions = ir_utils.build_definitions(callee_ir.blocks) canonicalize_array_math(callee_ir, f_typemap, f_calltypes, self.typingctx) # remove argument entries like arg.a from typemap arg_names = [vname for vname in f_typemap if vname.startswith("arg.")] for a in arg_names: f_typemap.pop(a) self.typemap.update(f_typemap) self.calltypes.update(f_calltypes) def inline_closure_call(func_ir, glbls, block, i, callee, typingctx=None, targetctx=None, arg_typs=None, typemap=None, calltypes=None, work_list=None, callee_validator=None, replace_freevars=True): """Inline the body of `callee` at its callsite (`i`-th instruction of `block`) `func_ir` is the func_ir object of the caller function and `glbls` is its global variable environment (func_ir.func_id.func.__globals__). `block` is the IR block of the callsite and `i` is the index of the callsite's node. `callee` is either the called function or a make_function node. `typingctx`, `typemap` and `calltypes` are typing data structures of the caller, available if we are in a typed pass. `arg_typs` includes the types of the arguments at the callsite. `callee_validator` is an optional callable which can be used to validate the IR of the callee to ensure that it contains IR supported for inlining, it takes one argument, the func_ir of the callee Returns IR blocks of the callee and the variable renaming dictionary used for them to facilitate further processing of new blocks. """ scope = block.scope instr = block.body[i] call_expr = instr.value debug_print = _make_debug_print("inline_closure_call") debug_print("Found closure call: ", instr, " with callee = ", callee) # support both function object and make_function Expr callee_code = callee.code if hasattr(callee, 'code') else callee.__code__ callee_closure = callee.closure if hasattr(callee, 'closure') else callee.__closure__ # first, get the IR of the callee if isinstance(callee, pytypes.FunctionType): from numba.core import compiler callee_ir = compiler.run_frontend(callee, inline_closures=True) else: callee_ir = get_ir_of_code(glbls, callee_code) # check that the contents of the callee IR is something that can be inlined # if a validator is supplied if callee_validator is not None: callee_validator(callee_ir) callee_blocks = callee_ir.blocks # 1. relabel callee_ir by adding an offset max_label = max(ir_utils._the_max_label.next(), max(func_ir.blocks.keys())) callee_blocks = add_offset_to_labels(callee_blocks, max_label + 1) callee_blocks = simplify_CFG(callee_blocks) callee_ir.blocks = callee_blocks min_label = min(callee_blocks.keys()) max_label = max(callee_blocks.keys()) # reset globals in ir_utils before we use it ir_utils._the_max_label.update(max_label) debug_print("After relabel") _debug_dump(callee_ir) # 2. rename all local variables in callee_ir with new locals created in func_ir callee_scopes = _get_all_scopes(callee_blocks) debug_print("callee_scopes = ", callee_scopes) # one function should only have one local scope assert(len(callee_scopes) == 1) callee_scope = callee_scopes[0] var_dict = {} for var in callee_scope.localvars._con.values(): if not (var.name in callee_code.co_freevars): inlined_name = _created_inlined_var_name( callee_ir.func_id.unique_name, var.name) new_var = scope.redefine(inlined_name, loc=var.loc) var_dict[var.name] = new_var debug_print("var_dict = ", var_dict) replace_vars(callee_blocks, var_dict) debug_print("After local var rename") _debug_dump(callee_ir) # 3. replace formal parameters with actual arguments args = _get_callee_args(call_expr, callee, block.body[i].loc, func_ir) debug_print("After arguments rename: ") _debug_dump(callee_ir) # 4. replace freevar with actual closure var if callee_closure and replace_freevars: closure = func_ir.get_definition(callee_closure) debug_print("callee's closure = ", closure) if isinstance(closure, tuple): cellget = ctypes.pythonapi.PyCell_Get cellget.restype = ctypes.py_object cellget.argtypes = (ctypes.py_object,) items = tuple(cellget(x) for x in closure) else: assert(isinstance(closure, ir.Expr) and closure.op == 'build_tuple') items = closure.items assert(len(callee_code.co_freevars) == len(items)) _replace_freevars(callee_blocks, items) debug_print("After closure rename") _debug_dump(callee_ir) if typingctx: from numba.core import typed_passes # call branch pruning to simplify IR and avoid inference errors callee_ir._definitions = ir_utils.build_definitions(callee_ir.blocks) numba.core.analysis.dead_branch_prune(callee_ir, arg_typs) try: f_typemap, f_return_type, f_calltypes, _ = typed_passes.type_inference_stage( typingctx, targetctx, callee_ir, arg_typs, None) except Exception as e: f_typemap, f_return_type, f_calltypes, _ = typed_passes.type_inference_stage( typingctx, targetctx, callee_ir, arg_typs, None) pass canonicalize_array_math(callee_ir, f_typemap, f_calltypes, typingctx) # remove argument entries like arg.a from typemap arg_names = [vname for vname in f_typemap if vname.startswith("arg.")] for a in arg_names: f_typemap.pop(a) typemap.update(f_typemap) calltypes.update(f_calltypes) _replace_args_with(callee_blocks, args) # 5. split caller blocks into two new_blocks = [] new_block = ir.Block(scope, block.loc) new_block.body = block.body[i + 1:] new_label = next_label() func_ir.blocks[new_label] = new_block new_blocks.append((new_label, new_block)) block.body = block.body[:i] block.body.append(ir.Jump(min_label, instr.loc)) # 6. replace Return with assignment to LHS topo_order = find_topo_order(callee_blocks) _replace_returns(callee_blocks, instr.target, new_label) # remove the old definition of instr.target too if (instr.target.name in func_ir._definitions and call_expr in func_ir._definitions[instr.target.name]): # NOTE: target can have multiple definitions due to control flow func_ir._definitions[instr.target.name].remove(call_expr) # 7. insert all new blocks, and add back definitions for label in topo_order: # block scope must point to parent's block = callee_blocks[label] block.scope = scope _add_definitions(func_ir, block) func_ir.blocks[label] = block new_blocks.append((label, block)) debug_print("After merge in") _debug_dump(func_ir) if work_list is not None: for block in new_blocks: work_list.append(block) return callee_blocks, var_dict def _get_callee_args(call_expr, callee, loc, func_ir): """Get arguments for calling 'callee', including the default arguments. keyword arguments are currently only handled when 'callee' is a function. """ if call_expr.op == 'call': args = list(call_expr.args) if call_expr.vararg: msg = "Calling a closure with *args is unsupported." raise errors.UnsupportedError(msg, call_expr.loc) elif call_expr.op == 'getattr': args = [call_expr.value] elif ir_utils.is_operator_or_getitem(call_expr): args = call_expr.list_vars() else: raise TypeError("Unsupported ir.Expr.{}".format(call_expr.op)) debug_print = _make_debug_print("inline_closure_call default handling") # handle defaults and kw arguments using pysignature if callee is function if isinstance(callee, pytypes.FunctionType): pysig = numba.core.utils.pysignature(callee) normal_handler = lambda index, param, default: default default_handler = lambda index, param, default: ir.Const(default, loc) # Throw error for stararg # TODO: handle stararg def stararg_handler(index, param, default): raise NotImplementedError( "Stararg not supported in inliner for arg {} {}".format( index, param)) if call_expr.op == 'call': kws = dict(call_expr.kws) else: kws = {} return numba.core.typing.fold_arguments( pysig, args, kws, normal_handler, default_handler, stararg_handler) else: # TODO: handle arguments for make_function case similar to function # case above callee_defaults = (callee.defaults if hasattr(callee, 'defaults') else callee.__defaults__) if callee_defaults: debug_print("defaults = ", callee_defaults) if isinstance(callee_defaults, tuple): # Python 3.5 defaults_list = [] for x in callee_defaults: if isinstance(x, ir.Var): defaults_list.append(x) else: # this branch is predominantly for kwargs from # inlinable functions defaults_list.append(ir.Const(value=x, loc=loc)) args = args + defaults_list elif (isinstance(callee_defaults, ir.Var) or isinstance(callee_defaults, str)): default_tuple = func_ir.get_definition(callee_defaults) assert(isinstance(default_tuple, ir.Expr)) assert(default_tuple.op == "build_tuple") const_vals = [func_ir.get_definition(x) for x in default_tuple.items] args = args + const_vals else: raise NotImplementedError( "Unsupported defaults to make_function: {}".format( defaults)) return args def _make_debug_print(prefix): def debug_print(*args): if config.DEBUG_INLINE_CLOSURE: print(prefix + ": " + "".join(str(x) for x in args)) return debug_print def _debug_dump(func_ir): if config.DEBUG_INLINE_CLOSURE: func_ir.dump() def _get_all_scopes(blocks): """Get all block-local scopes from an IR. """ all_scopes = [] for label, block in blocks.items(): if not (block.scope in all_scopes): all_scopes.append(block.scope) return all_scopes def _replace_args_with(blocks, args): """ Replace ir.Arg(...) with real arguments from call site """ for label, block in blocks.items(): assigns = block.find_insts(ir.Assign) for stmt in assigns: if isinstance(stmt.value, ir.Arg): idx = stmt.value.index assert(idx < len(args)) stmt.value = args[idx] def _replace_freevars(blocks, args): """ Replace ir.FreeVar(...) with real variables from parent function """ for label, block in blocks.items(): assigns = block.find_insts(ir.Assign) for stmt in assigns: if isinstance(stmt.value, ir.FreeVar): idx = stmt.value.index assert(idx < len(args)) if isinstance(args[idx], ir.Var): stmt.value = args[idx] else: stmt.value = ir.Const(args[idx], stmt.loc) def _replace_returns(blocks, target, return_label): """ Return return statement by assigning directly to target, and a jump. """ for label, block in blocks.items(): casts = [] for i in range(len(block.body)): stmt = block.body[i] if isinstance(stmt, ir.Return): assert(i + 1 == len(block.body)) block.body[i] = ir.Assign(stmt.value, target, stmt.loc) block.body.append(ir.Jump(return_label, stmt.loc)) # remove cast of the returned value for cast in casts: if cast.target.name == stmt.value.name: cast.value = cast.value.value elif isinstance(stmt, ir.Assign) and isinstance(stmt.value, ir.Expr) and stmt.value.op == 'cast': casts.append(stmt) def _add_definitions(func_ir, block): """ Add variable definitions found in a block to parent func_ir. """ definitions = func_ir._definitions assigns = block.find_insts(ir.Assign) for stmt in assigns: definitions[stmt.target.name].append(stmt.value) def _find_arraycall(func_ir, block): """Look for statement like "x = numpy.array(y)" or "x[..] = y" immediately after the closure call that creates list y (the i-th statement in block). Return the statement index if found, or raise GuardException. """ array_var = None array_call_index = None list_var_dead_after_array_call = False list_var = None i = 0 while i < len(block.body): instr = block.body[i] if isinstance(instr, ir.Del): # Stop the process if list_var becomes dead if list_var and array_var and instr.value == list_var.name: list_var_dead_after_array_call = True break pass elif isinstance(instr, ir.Assign): # Found array_var = array(list_var) lhs = instr.target expr = instr.value if (guard(find_callname, func_ir, expr) == ('array', 'numpy') and isinstance(expr.args[0], ir.Var)): list_var = expr.args[0] array_var = lhs array_stmt_index = i array_kws = dict(expr.kws) elif (isinstance(instr, ir.SetItem) and isinstance(instr.value, ir.Var) and not list_var): list_var = instr.value # Found array_var[..] = list_var, the case for nested array array_var = instr.target array_def = get_definition(func_ir, array_var) require(guard(_find_unsafe_empty_inferred, func_ir, array_def)) array_stmt_index = i array_kws = {} else: # Bail out otherwise break i = i + 1 # require array_var is found, and list_var is dead after array_call. require(array_var and list_var_dead_after_array_call) _make_debug_print("find_array_call")(block.body[array_stmt_index]) return list_var, array_stmt_index, array_kws def _find_iter_range(func_ir, range_iter_var, swapped): """Find the iterator's actual range if it is either range(n), or range(m, n), otherwise return raise GuardException. """ debug_print = _make_debug_print("find_iter_range") range_iter_def = get_definition(func_ir, range_iter_var) debug_print("range_iter_var = ", range_iter_var, " def = ", range_iter_def) require(isinstance(range_iter_def, ir.Expr) and range_iter_def.op == 'getiter') range_var = range_iter_def.value range_def = get_definition(func_ir, range_var) debug_print("range_var = ", range_var, " range_def = ", range_def) require(isinstance(range_def, ir.Expr) and range_def.op == 'call') func_var = range_def.func func_def = get_definition(func_ir, func_var) debug_print("func_var = ", func_var, " func_def = ", func_def) require(isinstance(func_def, ir.Global) and (func_def.value == range or func_def.value == numba.misc.special.prange)) nargs = len(range_def.args) swapping = [('"array comprehension"', 'closure of'), range_def.func.loc] if nargs == 1: swapped[range_def.func.name] = swapping stop = get_definition(func_ir, range_def.args[0], lhs_only=True) return (0, range_def.args[0], func_def) elif nargs == 2: swapped[range_def.func.name] = swapping start = get_definition(func_ir, range_def.args[0], lhs_only=True) stop = get_definition(func_ir, range_def.args[1], lhs_only=True) return (start, stop, func_def) else: raise GuardException def _inline_arraycall(func_ir, cfg, visited, loop, swapped, enable_prange=False, typed=False): """Look for array(list) call in the exit block of a given loop, and turn list operations into array operations in the loop if the following conditions are met: 1. The exit block contains an array call on the list; 2. The list variable is no longer live after array call; 3. The list is created in the loop entry block; 4. The loop is created from an range iterator whose length is known prior to the loop; 5. There is only one list_append operation on the list variable in the loop body; 6. The block that contains list_append dominates the loop head, which ensures list length is the same as loop length; If any condition check fails, no modification will be made to the incoming IR. """ debug_print = _make_debug_print("inline_arraycall") # There should only be one loop exit require(len(loop.exits) == 1) exit_block = next(iter(loop.exits)) list_var, array_call_index, array_kws = _find_arraycall(func_ir, func_ir.blocks[exit_block]) # check if dtype is present in array call dtype_def = None dtype_mod_def = None if 'dtype' in array_kws: require(isinstance(array_kws['dtype'], ir.Var)) # We require that dtype argument to be a constant of getattr Expr, and we'll # remember its definition for later use. dtype_def = get_definition(func_ir, array_kws['dtype']) require(isinstance(dtype_def, ir.Expr) and dtype_def.op == 'getattr') dtype_mod_def = get_definition(func_ir, dtype_def.value) list_var_def = get_definition(func_ir, list_var) debug_print("list_var = ", list_var, " def = ", list_var_def) if isinstance(list_var_def, ir.Expr) and list_var_def.op == 'cast': list_var_def = get_definition(func_ir, list_var_def.value) # Check if the definition is a build_list require(isinstance(list_var_def, ir.Expr) and list_var_def.op == 'build_list') # The build_list must be empty require(len(list_var_def.items) == 0) # Look for list_append in "last" block in loop body, which should be a block that is # a post-dominator of the loop header. list_append_stmts = [] for label in loop.body: # We have to consider blocks of this loop, but not sub-loops. # To achieve this, we require the set of "in_loops" of "label" to be visited loops. in_visited_loops = [l.header in visited for l in cfg.in_loops(label)] if not all(in_visited_loops): continue block = func_ir.blocks[label] debug_print("check loop body block ", label) for stmt in block.find_insts(ir.Assign): lhs = stmt.target expr = stmt.value if isinstance(expr, ir.Expr) and expr.op == 'call': func_def = get_definition(func_ir, expr.func) if isinstance(func_def, ir.Expr) and func_def.op == 'getattr' \ and func_def.attr == 'append': list_def = get_definition(func_ir, func_def.value) debug_print("list_def = ", list_def, list_def is list_var_def) if list_def is list_var_def: # found matching append call list_append_stmts.append((label, block, stmt)) # Require only one list_append, otherwise we won't know the indices require(len(list_append_stmts) == 1) append_block_label, append_block, append_stmt = list_append_stmts[0] # Check if append_block (besides loop entry) dominates loop header. # Since CFG doesn't give us this info without loop entry, we approximate # by checking if the predecessor set of the header block is the same # as loop_entries plus append_block, which is certainly more restrictive # than necessary, and can be relaxed if needed. preds = set(l for l, b in cfg.predecessors(loop.header)) debug_print("preds = ", preds, (loop.entries | set([append_block_label]))) require(preds == (loop.entries | set([append_block_label]))) # Find iterator in loop header iter_vars = [] iter_first_vars = [] loop_header = func_ir.blocks[loop.header] for stmt in loop_header.find_insts(ir.Assign): expr = stmt.value if isinstance(expr, ir.Expr): if expr.op == 'iternext': iter_def = get_definition(func_ir, expr.value) debug_print("iter_def = ", iter_def) iter_vars.append(expr.value) elif expr.op == 'pair_first': iter_first_vars.append(stmt.target) # Require only one iterator in loop header require(len(iter_vars) == 1 and len(iter_first_vars) == 1) iter_var = iter_vars[0] # variable that holds the iterator object iter_first_var = iter_first_vars[0] # variable that holds the value out of iterator # Final requirement: only one loop entry, and we're going to modify it by: # 1. replacing the list definition with an array definition; # 2. adding a counter for the array iteration. require(len(loop.entries) == 1) loop_entry = func_ir.blocks[next(iter(loop.entries))] terminator = loop_entry.terminator scope = loop_entry.scope loc = loop_entry.loc stmts = [] removed = [] def is_removed(val, removed): if isinstance(val, ir.Var): for x in removed: if x.name == val.name: return True return False # Skip list construction and skip terminator, add the rest to stmts for i in range(len(loop_entry.body) - 1): stmt = loop_entry.body[i] if isinstance(stmt, ir.Assign) and (stmt.value is list_def or is_removed(stmt.value, removed)): removed.append(stmt.target) else: stmts.append(stmt) debug_print("removed variables: ", removed) # Define an index_var to index the array. # If the range happens to be single step ranges like range(n), or range(m, n), # then the index_var correlates to iterator index; otherwise we'll have to # define a new counter. range_def = guard(_find_iter_range, func_ir, iter_var, swapped) index_var = ir.Var(scope, mk_unique_var("index"), loc) if range_def and range_def[0] == 0: # iterator starts with 0, index_var can just be iter_first_var index_var = iter_first_var else: # index_var = -1 # starting the index with -1 since it will incremented in loop header stmts.append(_new_definition(func_ir, index_var, ir.Const(value=-1, loc=loc), loc)) # Insert statement to get the size of the loop iterator size_var = ir.Var(scope, mk_unique_var("size"), loc) if range_def: start, stop, range_func_def = range_def if start == 0: size_val = stop else: size_val = ir.Expr.binop(fn=operator.sub, lhs=stop, rhs=start, loc=loc) # we can parallelize this loop if enable_prange = True, by changing # range function from range, to prange. if enable_prange and isinstance(range_func_def, ir.Global): range_func_def.name = 'internal_prange' range_func_def.value = internal_prange else: # this doesn't work in objmode as it's effectively untyped if typed: len_func_var = ir.Var(scope, mk_unique_var("len_func"), loc) from numba.cpython.rangeobj import length_of_iterator stmts.append(_new_definition(func_ir, len_func_var, ir.Global('length_of_iterator', length_of_iterator, loc=loc), loc)) size_val = ir.Expr.call(len_func_var, (iter_var,), (), loc=loc) else: raise GuardException stmts.append(_new_definition(func_ir, size_var, size_val, loc)) size_tuple_var = ir.Var(scope, mk_unique_var("size_tuple"), loc) stmts.append(_new_definition(func_ir, size_tuple_var, ir.Expr.build_tuple(items=[size_var], loc=loc), loc)) # Insert array allocation array_var = ir.Var(scope, mk_unique_var("array"), loc) empty_func = ir.Var(scope, mk_unique_var("empty_func"), loc) if dtype_def and dtype_mod_def: # when dtype is present, we'll call empty with dtype dtype_mod_var = ir.Var(scope, mk_unique_var("dtype_mod"), loc) dtype_var = ir.Var(scope, mk_unique_var("dtype"), loc) stmts.append(_new_definition(func_ir, dtype_mod_var, dtype_mod_def, loc)) stmts.append(_new_definition(func_ir, dtype_var, ir.Expr.getattr(dtype_mod_var, dtype_def.attr, loc), loc)) stmts.append(_new_definition(func_ir, empty_func, ir.Global('empty', np.empty, loc=loc), loc)) array_kws = [('dtype', dtype_var)] else: # this doesn't work in objmode as it's effectively untyped if typed: # otherwise we'll call unsafe_empty_inferred stmts.append(_new_definition(func_ir, empty_func, ir.Global('unsafe_empty_inferred', unsafe_empty_inferred, loc=loc), loc)) array_kws = [] else: raise GuardException # array_var = empty_func(size_tuple_var) stmts.append(_new_definition(func_ir, array_var, ir.Expr.call(empty_func, (size_tuple_var,), list(array_kws), loc=loc), loc)) # Add back removed just in case they are used by something else for var in removed: stmts.append(_new_definition(func_ir, var, array_var, loc)) # Add back terminator stmts.append(terminator) # Modify loop_entry loop_entry.body = stmts if range_def: if range_def[0] != 0: # when range doesn't start from 0, index_var becomes loop index # (iter_first_var) minus an offset (range_def[0]) terminator = loop_header.terminator assert(isinstance(terminator, ir.Branch)) # find the block in the loop body that header jumps to block_id = terminator.truebr blk = func_ir.blocks[block_id] loc = blk.loc blk.body.insert(0, _new_definition(func_ir, index_var, ir.Expr.binop(fn=operator.sub, lhs=iter_first_var, rhs=range_def[0], loc=loc), loc)) else: # Insert index_var increment to the end of loop header loc = loop_header.loc terminator = loop_header.terminator stmts = loop_header.body[0:-1] next_index_var = ir.Var(scope, mk_unique_var("next_index"), loc) one = ir.Var(scope, mk_unique_var("one"), loc) # one = 1 stmts.append(_new_definition(func_ir, one, ir.Const(value=1,loc=loc), loc)) # next_index_var = index_var + 1 stmts.append(_new_definition(func_ir, next_index_var, ir.Expr.binop(fn=operator.add, lhs=index_var, rhs=one, loc=loc), loc)) # index_var = next_index_var stmts.append(_new_definition(func_ir, index_var, next_index_var, loc)) stmts.append(terminator) loop_header.body = stmts # In append_block, change list_append into array assign for i in range(len(append_block.body)): if append_block.body[i] is append_stmt: debug_print("Replace append with SetItem") append_block.body[i] = ir.SetItem(target=array_var, index=index_var, value=append_stmt.value.args[0], loc=append_stmt.loc) # replace array call, by changing "a = array(b)" to "a = b" stmt = func_ir.blocks[exit_block].body[array_call_index] # stmt can be either array call or SetItem, we only replace array call if isinstance(stmt, ir.Assign) and isinstance(stmt.value, ir.Expr): stmt.value = array_var func_ir._definitions[stmt.target.name] = [stmt.value] return True def _find_unsafe_empty_inferred(func_ir, expr): unsafe_empty_inferred require(isinstance(expr, ir.Expr) and expr.op == 'call') callee = expr.func callee_def = get_definition(func_ir, callee) require(isinstance(callee_def, ir.Global)) _make_debug_print("_find_unsafe_empty_inferred")(callee_def.value) return callee_def.value == unsafe_empty_inferred def _fix_nested_array(func_ir): """Look for assignment like: a[..] = b, where both a and b are numpy arrays, and try to eliminate array b by expanding a with an extra dimension. """ blocks = func_ir.blocks cfg = compute_cfg_from_blocks(blocks) usedefs = compute_use_defs(blocks) empty_deadmap = dict([(label, set()) for label in blocks.keys()]) livemap = compute_live_variables(cfg, blocks, usedefs.defmap, empty_deadmap) def find_array_def(arr): """Find numpy array definition such as arr = numba.unsafe.ndarray.empty_inferred(...). If it is arr = b[...], find array definition of b recursively. """ arr_def = get_definition(func_ir, arr) _make_debug_print("find_array_def")(arr, arr_def) if isinstance(arr_def, ir.Expr): if guard(_find_unsafe_empty_inferred, func_ir, arr_def): return arr_def elif arr_def.op == 'getitem': return find_array_def(arr_def.value) raise GuardException def fix_dependencies(expr, varlist): """Double check if all variables in varlist are defined before expr is used. Try to move constant definition when the check fails. Bails out by raising GuardException if it can't be moved. """ debug_print = _make_debug_print("fix_dependencies") for label, block in blocks.items(): scope = block.scope body = block.body defined = set() for i in range(len(body)): inst = body[i] if isinstance(inst, ir.Assign): defined.add(inst.target.name) if inst.value is expr: new_varlist = [] for var in varlist: # var must be defined before this inst, or live # and not later defined. if (var.name in defined or (var.name in livemap[label] and not (var.name in usedefs.defmap[label]))): debug_print(var.name, " already defined") new_varlist.append(var) else: debug_print(var.name, " not yet defined") var_def = get_definition(func_ir, var.name) if isinstance(var_def, ir.Const): loc = var.loc new_var = ir.Var(scope, mk_unique_var("new_var"), loc) new_const = ir.Const(var_def.value, loc) new_vardef = _new_definition(func_ir, new_var, new_const, loc) new_body = [] new_body.extend(body[:i]) new_body.append(new_vardef) new_body.extend(body[i:]) block.body = new_body new_varlist.append(new_var) else: raise GuardException return new_varlist # when expr is not found in block raise GuardException def fix_array_assign(stmt): """For assignment like lhs[idx] = rhs, where both lhs and rhs are arrays, do the following: 1. find the definition of rhs, which has to be a call to numba.unsafe.ndarray.empty_inferred 2. find the source array creation for lhs, insert an extra dimension of size of b. 3. replace the definition of rhs = numba.unsafe.ndarray.empty_inferred(...) with rhs = lhs[idx] """ require(isinstance(stmt, ir.SetItem)) require(isinstance(stmt.value, ir.Var)) debug_print = _make_debug_print("fix_array_assign") debug_print("found SetItem: ", stmt) lhs = stmt.target # Find the source array creation of lhs lhs_def = find_array_def(lhs) debug_print("found lhs_def: ", lhs_def) rhs_def = get_definition(func_ir, stmt.value) debug_print("found rhs_def: ", rhs_def) require(isinstance(rhs_def, ir.Expr)) if rhs_def.op == 'cast': rhs_def = get_definition(func_ir, rhs_def.value) require(isinstance(rhs_def, ir.Expr)) require(_find_unsafe_empty_inferred(func_ir, rhs_def)) # Find the array dimension of rhs dim_def = get_definition(func_ir, rhs_def.args[0]) require(isinstance(dim_def, ir.Expr) and dim_def.op == 'build_tuple') debug_print("dim_def = ", dim_def) extra_dims = [ get_definition(func_ir, x, lhs_only=True) for x in dim_def.items ] debug_print("extra_dims = ", extra_dims) # Expand size tuple when creating lhs_def with extra_dims size_tuple_def = get_definition(func_ir, lhs_def.args[0]) require(isinstance(size_tuple_def, ir.Expr) and size_tuple_def.op == 'build_tuple') debug_print("size_tuple_def = ", size_tuple_def) extra_dims = fix_dependencies(size_tuple_def, extra_dims) size_tuple_def.items += extra_dims # In-place modify rhs_def to be getitem rhs_def.op = 'getitem' rhs_def.fn = operator.getitem rhs_def.value = get_definition(func_ir, lhs, lhs_only=True) rhs_def.index = stmt.index del rhs_def._kws['func'] del rhs_def._kws['args'] del rhs_def._kws['vararg'] del rhs_def._kws['kws'] # success return True for label in find_topo_order(func_ir.blocks): block = func_ir.blocks[label] for stmt in block.body: if guard(fix_array_assign, stmt): block.body.remove(stmt) def _new_definition(func_ir, var, value, loc): func_ir._definitions[var.name] = [value] return ir.Assign(value=value, target=var, loc=loc) @rewrites.register_rewrite('after-inference') class RewriteArrayOfConsts(rewrites.Rewrite): '''The RewriteArrayOfConsts class is responsible for finding 1D array creations from a constant list, and rewriting it into direct initialization of array elements without creating the list. ''' def __init__(self, state, *args, **kws): self.typingctx = state.typingctx super(RewriteArrayOfConsts, self).__init__(*args, **kws) def match(self, func_ir, block, typemap, calltypes): if len(calltypes) == 0: return False self.crnt_block = block self.new_body = guard(_inline_const_arraycall, block, func_ir, self.typingctx, typemap, calltypes) return self.new_body is not None def apply(self): self.crnt_block.body = self.new_body return self.crnt_block def _inline_const_arraycall(block, func_ir, context, typemap, calltypes): """Look for array(list) call where list is a constant list created by build_list, and turn them into direct array creation and initialization, if the following conditions are met: 1. The build_list call immediate precedes the array call; 2. The list variable is no longer live after array call; If any condition check fails, no modification will be made. """ debug_print = _make_debug_print("inline_const_arraycall") scope = block.scope def inline_array(array_var, expr, stmts, list_vars, dels): """Check to see if the given "array_var" is created from a list of constants, and try to inline the list definition as array initialization. Extra statements produced with be appended to "stmts". """ callname = guard(find_callname, func_ir, expr) require(callname and callname[1] == 'numpy' and callname[0] == 'array') require(expr.args[0].name in list_vars) ret_type = calltypes[expr].return_type require(isinstance(ret_type, types.ArrayCompatible) and ret_type.ndim == 1) loc = expr.loc list_var = expr.args[0] # Get the type of the array to be created. array_typ = typemap[array_var.name] debug_print("inline array_var = ", array_var, " list_var = ", list_var) # Get the element type of the array to be created. dtype = array_typ.dtype # Get the sequence of operations to provide values to the new array. seq, _ = find_build_sequence(func_ir, list_var) size = len(seq) # Create a tuple to pass to empty below to specify the new array size. size_var = ir.Var(scope, mk_unique_var("size"), loc) size_tuple_var = ir.Var(scope, mk_unique_var("size_tuple"), loc) size_typ = types.intp size_tuple_typ = types.UniTuple(size_typ, 1) typemap[size_var.name] = size_typ typemap[size_tuple_var.name] = size_tuple_typ stmts.append(_new_definition(func_ir, size_var, ir.Const(size, loc=loc), loc)) stmts.append(_new_definition(func_ir, size_tuple_var, ir.Expr.build_tuple(items=[size_var], loc=loc), loc)) # The general approach is to create an empty array and then fill # the elements in one-by-one from their specificiation. # Get the numpy type to pass to empty. nptype = types.DType(dtype) # Create a variable to hold the numpy empty function. empty_func = ir.Var(scope, mk_unique_var("empty_func"), loc) fnty = get_np_ufunc_typ(np.empty) sig = context.resolve_function_type(fnty, (size_typ,), {'dtype':nptype}) typemap[empty_func.name] = fnty stmts.append(_new_definition(func_ir, empty_func, ir.Global('empty', np.empty, loc=loc), loc)) # We pass two arguments to empty, first the size tuple and second # the dtype of the new array. Here, we created typ_var which is # the dtype argument of the new array. typ_var in turn is created # by getattr of the dtype string on the numpy module. # Create var for numpy module. g_np_var = ir.Var(scope, mk_unique_var("$np_g_var"), loc) typemap[g_np_var.name] = types.misc.Module(np) g_np = ir.Global('np', np, loc) stmts.append(_new_definition(func_ir, g_np_var, g_np, loc)) # Create var for result of numpy.. typ_var = ir.Var(scope, mk_unique_var("$np_typ_var"), loc) typemap[typ_var.name] = nptype dtype_str = str(dtype) if dtype_str == 'bool': dtype_str = 'bool_' # Get dtype attribute of numpy module. np_typ_getattr = ir.Expr.getattr(g_np_var, dtype_str, loc) stmts.append(_new_definition(func_ir, typ_var, np_typ_getattr, loc)) # Create the call to numpy.empty passing the size tuple and dtype var. empty_call = ir.Expr.call(empty_func, [size_var, typ_var], {}, loc=loc) calltypes[empty_call] = typing.signature(array_typ, size_typ, nptype) stmts.append(_new_definition(func_ir, array_var, empty_call, loc)) # Fill in the new empty array one-by-one. for i in range(size): index_var = ir.Var(scope, mk_unique_var("index"), loc) index_typ = types.intp typemap[index_var.name] = index_typ stmts.append(_new_definition(func_ir, index_var, ir.Const(i, loc), loc)) setitem = ir.SetItem(array_var, index_var, seq[i], loc) calltypes[setitem] = typing.signature(types.none, array_typ, index_typ, dtype) stmts.append(setitem) stmts.extend(dels) return True class State(object): """ This class is used to hold the state in the following loop so as to make it easy to reset the state of the variables tracking the various statement kinds """ def __init__(self): # list_vars keep track of the variable created from the latest # build_list instruction, as well as its synonyms. self.list_vars = [] # dead_vars keep track of those in list_vars that are considered dead. self.dead_vars = [] # list_items keep track of the elements used in build_list. self.list_items = [] self.stmts = [] # dels keep track of the deletion of list_items, which will need to be # moved after array initialization. self.dels = [] # tracks if a modification has taken place self.modified = False def reset(self): """ Resets the internal state of the variables used for tracking """ self.list_vars = [] self.dead_vars = [] self.list_items = [] self.dels = [] def list_var_used(self, inst): """ Returns True if the list being analysed is used between the build_list and the array call. """ return any([x.name in self.list_vars for x in inst.list_vars()]) state = State() for inst in block.body: if isinstance(inst, ir.Assign): if isinstance(inst.value, ir.Var): if inst.value.name in state.list_vars: state.list_vars.append(inst.target.name) state.stmts.append(inst) continue elif isinstance(inst.value, ir.Expr): expr = inst.value if expr.op == 'build_list': # new build_list encountered, reset state state.reset() state.list_items = [x.name for x in expr.items] state.list_vars = [inst.target.name] state.stmts.append(inst) continue elif expr.op == 'call' and expr in calltypes: arr_var = inst.target if guard(inline_array, inst.target, expr, state.stmts, state.list_vars, state.dels): state.modified = True continue elif isinstance(inst, ir.Del): removed_var = inst.value if removed_var in state.list_items: state.dels.append(inst) continue elif removed_var in state.list_vars: # one of the list_vars is considered dead. state.dead_vars.append(removed_var) state.list_vars.remove(removed_var) state.stmts.append(inst) if state.list_vars == []: # if all list_vars are considered dead, we need to filter # them out from existing stmts to completely remove # build_list. # Note that if a translation didn't take place, dead_vars # will also be empty when we reach this point. body = [] for inst in state.stmts: if ((isinstance(inst, ir.Assign) and inst.target.name in state.dead_vars) or (isinstance(inst, ir.Del) and inst.value in state.dead_vars)): continue body.append(inst) state.stmts = body state.dead_vars = [] state.modified = True continue state.stmts.append(inst) # If the list is used in any capacity between build_list and array # call, then we must call off the translation for this list because # it could be mutated and list_items would no longer be applicable. if state.list_var_used(inst): state.reset() return state.stmts if state.modified else None