def test_starmap_async() -> None: """Test that a function can be starmapped over many iterables asynchronously.""" with ProcessPool(processes=3) as pool: assert ( sum(pool.starmap_async(summation, [range(10), range(10), range(10)]).get()) == 135 )
def test_imap() -> None: """Test that mapped function results can be iterated.""" with ProcessPool(processes=3) as pool: total = 0 for result in pool.imap(square, [2] * 6): total += result assert total == 24
def test_with_context(mocker: MockerFixture) -> None: """Test that the composed pool's context is managed as well.""" pool = ProcessPool(processes=3) mock = mocker.patch.object(pool, "_pool", autospec=True) with pool: pass mock.__enter__.assert_called_once() mock.__exit__.assert_called_once()
def test_starmap() -> None: """Test that a function can be starmapped over many iterables.""" with ProcessPool(processes=3) as pool: assert sum(pool.starmap(summation, [range(10), range(10), range(10)])) == 135
def test_imap_unordered() -> None: """Test that mapped function results can be iterated in any order.""" with ProcessPool(processes=3) as pool: assert sum(pool.imap_unordered(square, [2] * 6)) == 24
def test_map_async() -> None: """Test that a function can be mapped over an iterable of values asynchronously.""" with ProcessPool(processes=3) as pool: assert sum(pool.map_async(square, [2] * 6).get()) == 24
def test_map() -> None: """Test that a function can be mapped over an iterable of values.""" with ProcessPool(processes=3) as pool: assert sum(pool.map(square, [2] * 6)) == 24
def test_apply_async() -> None: """Test that a function can be applied asynchronously.""" with ProcessPool(processes=3) as pool: assert pool.apply_async(square, (3,)).get() == 9
def test_apply() -> None: """Test that a function can be applied.""" with ProcessPool(processes=3) as pool: assert pool.apply(square, (3,)) == 9
def test_close(mocker: MockerFixture) -> None: """Test that the composed pool is closed as well.""" pool = ProcessPool(processes=3) mock = mocker.patch.object(pool, "_pool", autospec=True) pool.close() mock.close.assert_called_once()
def test_init(attributes: dict) -> None: """Test that a process pool can be initialized with each of its arguments.""" ProcessPool(**attributes)
def flux_variability_analysis( model, reaction_list=None, loopless=False, fraction_of_optimum=1.0, pfba_factor=None, processes=None, ): """ Determine the minimum and maximum possible flux value for each reaction. Parameters ---------- model : cobra.Model The model for which to run the analysis. It will *not* be modified. reaction_list : list of cobra.Reaction or str, optional The reactions for which to obtain min/max fluxes. If None will use all reactions in the model (default). loopless : boolean, optional Whether to return only loopless solutions. This is significantly slower. Please also refer to the notes. fraction_of_optimum : float, optional Must be <= 1.0. Requires that the objective value is at least the fraction times maximum objective value. A value of 0.85 for instance means that the objective has to be at least at 85% percent of its maximum. pfba_factor : float, optional Add an additional constraint to the model that requires the total sum of absolute fluxes must not be larger than this value times the smallest possible sum of absolute fluxes, i.e., by setting the value to 1.1 the total sum of absolute fluxes must not be more than 10% larger than the pFBA solution. Since the pFBA solution is the one that optimally minimizes the total flux sum, the ``pfba_factor`` should, if set, be larger than one. Setting this value may lead to more realistic predictions of the effective flux bounds. processes : int, optional The number of parallel processes to run. If not explicitly passed, will be set from the global configuration singleton. Returns ------- pandas.DataFrame A data frame with reaction identifiers as the index and two columns: - maximum: indicating the highest possible flux - minimum: indicating the lowest possible flux Notes ----- This implements the fast version as described in [1]_. Please note that the flux distribution containing all minimal/maximal fluxes does not have to be a feasible solution for the model. Fluxes are minimized/maximized individually and a single minimal flux might require all others to be suboptimal. Using the loopless option will lead to a significant increase in computation time (about a factor of 100 for large models). However, the algorithm used here (see [2]_) is still more than 1000x faster than the "naive" version using ``add_loopless(model)``. Also note that if you have included constraints that force a loop (for instance by setting all fluxes in a loop to be non-zero) this loop will be included in the solution. References ---------- .. [1] Computationally efficient flux variability analysis. Gudmundsson S, Thiele I. BMC Bioinformatics. 2010 Sep 29;11:489. doi: 10.1186/1471-2105-11-489, PMID: 20920235 .. [2] CycleFreeFlux: efficient removal of thermodynamically infeasible loops from flux distributions. Desouki AA, Jarre F, Gelius-Dietrich G, Lercher MJ. Bioinformatics. 2015 Jul 1;31(13):2159-65. doi: 10.1093/bioinformatics/btv096. """ if reaction_list is None: reaction_ids = [r.id for r in model.reactions] else: reaction_ids = [ r.id for r in model.reactions.get_by_any(reaction_list) ] if processes is None: processes = configuration.processes num_reactions = len(reaction_ids) processes = min(processes, num_reactions) fva_result = DataFrame( { "minimum": zeros(num_reactions, dtype=float), "maximum": zeros(num_reactions, dtype=float), }, index=reaction_ids, ) prob = model.problem with model: # Safety check before setting up FVA. model.slim_optimize( error_value=None, message="There is no optimal solution for the chosen objective!", ) # Add the previous objective as a variable to the model then set it to # zero. This also uses the fraction to create the lower/upper bound for # the old objective. # TODO: Use utility function here (fix_objective_as_constraint)? if model.solver.objective.direction == "max": fva_old_objective = prob.Variable( "fva_old_objective", lb=fraction_of_optimum * model.solver.objective.value, ) else: fva_old_objective = prob.Variable( "fva_old_objective", ub=fraction_of_optimum * model.solver.objective.value, ) fva_old_obj_constraint = prob.Constraint( model.solver.objective.expression - fva_old_objective, lb=0, ub=0, name="fva_old_objective_constraint", ) model.add_cons_vars([fva_old_objective, fva_old_obj_constraint]) if pfba_factor is not None: if pfba_factor < 1.0: warn( "The 'pfba_factor' should be larger or equal to 1.", UserWarning, ) with model: add_pfba(model, fraction_of_optimum=0) ub = model.slim_optimize(error_value=None) flux_sum = prob.Variable("flux_sum", ub=pfba_factor * ub) flux_sum_constraint = prob.Constraint( model.solver.objective.expression - flux_sum, lb=0, ub=0, name="flux_sum_constraint", ) model.add_cons_vars([flux_sum, flux_sum_constraint]) model.objective = Zero # This will trigger the reset as well for what in ("minimum", "maximum"): if processes > 1: # We create and destroy a new pool here in order to set the # objective direction for all reactions. This creates a # slight overhead but seems the most clean. chunk_size = len(reaction_ids) // processes with ProcessPool( processes, initializer=_init_worker, initargs=(model, loopless, what[:3]), ) as pool: for rxn_id, value in pool.imap_unordered( _fva_step, reaction_ids, chunksize=chunk_size): fva_result.at[rxn_id, what] = value else: _init_worker(model, loopless, what[:3]) for rxn_id, value in map(_fva_step, reaction_ids): fva_result.at[rxn_id, what] = value return fva_result[["minimum", "maximum"]]
def _multi_deletion(model, entity, element_lists, method="fba", solution=None, processes=None, **kwargs): """ Provide a common interface for single or multiple knockouts. Parameters ---------- model : cobra.Model The metabolic model to perform deletions in. entity : 'gene' or 'reaction' The entity to knockout (``cobra.Gene`` or ``cobra.Reaction``). element_lists : list List of iterables ``cobra.Reaction``s or ``cobra.Gene``s (or their IDs) to be deleted. method: {"fba", "moma", "linear moma", "room", "linear room"}, optional Method used to predict the growth rate. solution : cobra.Solution, optional A previous solution to use as a reference for (linear) MOMA or ROOM. processes : int, optional The number of parallel processes to run. Can speed up the computations if the number of knockouts to perform is large. If not passed, will be set to the number of CPUs found. kwargs : Passed on to underlying simulation functions. Returns ------- pandas.DataFrame A representation of all combinations of entity deletions. The columns are 'growth' and 'status', where index : tuple(str) The gene or reaction identifiers that were knocked out. growth : float The growth rate of the adjusted model. status : str The solution's status. """ solver = sutil.interface_to_str(model.problem.__name__) if method == "moma" and solver not in sutil.qp_solvers: raise RuntimeError( "Cannot use MOMA since '{}' is not QP-capable." "Please choose a different solver or use FBA only.".format(solver)) if processes is None: processes = configuration.processes with model: if "moma" in method: add_moma(model, solution=solution, linear="linear" in method) elif "room" in method: add_room(model, solution=solution, linear="linear" in method, **kwargs) args = set([frozenset(comb) for comb in product(*element_lists)]) processes = min(processes, len(args)) def extract_knockout_results(result_iter): result = pd.DataFrame( [( set(ids), growth, status, ) for (ids, growth, status) in result_iter], columns=["ids", "growth", "status"], ) return result if processes > 1: worker = dict(gene=_gene_deletion_worker, reaction=_reaction_deletion_worker)[entity] chunk_size = len(args) // processes with ProcessPool(processes, initializer=_init_worker, initargs=(model, )) as pool: results = extract_knockout_results( pool.imap_unordered(worker, args, chunksize=chunk_size)) else: worker = dict(gene=_gene_deletion, reaction=_reaction_deletion)[entity] results = extract_knockout_results( map(partial(worker, model), args)) return results