Skip to content
forked from enjine-com/mcos

Implementation of Monte Carlo Optimization Selection from the paper "A Robust Estimator of the Efficient Frontier"

License

Notifications You must be signed in to change notification settings

StonemeisterW/mcos

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Monte Carlo Optimization Simulation

This library is the fulfillment of all your hopes and dreams, assuming your hopes and dreams consist entirely of an implementation of the Monte Carlo Optimization selection from Dr. Marcos Lopez de Prado's 2019 paper 'A Robust Estimator of the Efficient Frontier'. Those are some weirdly specific hopes and dreams. But you're our kind of people. MCOS allows you to compare the allocation error of different portfolio optimization methods, given a particular set of inputs.

WHY MCOS?

Optimizing a portfolio is like petting a cat: the same approach doesn't work for every situation. Cats and markets are both complex creatures and you need to be cautious before you dive in. It is naive to think that one method is best for everything until the end of time. MCOS allows you to quickly compare different optimization methods to find which is most robust in your particular case.

HOW?

After you have calculated the covariance matrix and the expected returns for your portfolio you feed them into the simulator. Using the optimization methods you've selected, the library then calculates the optimal weights. Then a Monte Carlo simulation is performed, where optimal allocations are computed on a large number of simulated covariances and returns. It then compares these allocations to the ideal and calculates the error.

GETTING STARTED

Install and update using pip:

pip install mcos

RUNNING SIMULATIONS

Before using the MCOS library, it is assumed that you have calculated both the covariance and the expected return vectors of your portfolio for all of the optimizers that you wish to compare.

The main entry point for the library is the mcos.py file. The entry point is mcos.simulate_observations(). This function completes the MCOS procedure for all of the specified optimizers, and returns to you a DataFrame with the results. It takes an observation simulator, the number of simulations you'd like to run, a list of all the optimizers you would like to compare, the type of metric you'd like to test, and an optional covariance transformer. All of these are described below.

SAMPLE SIMULATION

obs_sim = MuCovObservationSimulator(expected_returns, covariance, num_sims)
optimizers = [HRPOptimizer(), MarkowitzOptimizer(),NCOOptimizer(max_num_clusters, num_clustering_trials)]
error_estimator = VarianceErrorEstimator()
covariance_transformers = [DeNoiserCovarianceTransformer()]
num_sims = 50

results = mcos.simulate_optimizations(obs_sim, num_sims, optimizers, error_estimator, covariance_transformers)

DATA INPUTS

As mentioned above, the key input for the system is the expected return vector and covariance of the portfolio that you are trying to analyze. You choose the type of Observation Simulator and initialize it with the covariance and expected returns. Please feed the Simulator only numpy arrays, lest it get cranky and uncooperative. The expected return vector is a 1 dimensional array of expected returns for the portfolio constituents, while the covariance is a n x n matrix. You must also include the number of simulations that you wish to run. The observation simulators that are currently supported are:

  1. Standard - The chicken fingers of simulators. Plain, unexciting, but darn it, it gets the job done. Regular estimation of the covariance matrix.

  2. Ledoit-Wolf - If you prefer your covariance matrix shrunken, this is the one for you. Read all about it in this unfortunately titled paper

CONFIGURATION INPUTS

Along with selecting your choice of Observation Simulator, you can also specify the optimzers that you would like to compare. These are passed in as a list of Optimizer class objects. The Optimizers currently supported are:

  1. Markowitz Optimization – Modern Portfolio Theory: the original gangster of portfolio optimizations created in 1952.

    MarkowitzOptimizer()

  2. Nested Cluster Optimization – Optimization developed by Marcos Lopez de Prado and laid out in his 2019 paper “A Robust Estimator of the Efficient Frontier”. There are two optional variables that you can pass to this specific optimizer. They are the maximum number of clusters to use during clustering, and the number of times to perform the clustering.

    NCOOptimizer(max_num_clusters, num_clustering_trials)

  3. Risk Parity – Risk Parity builds on the work of Markowitz to create portfolios that focus on diversifying risk instead of diversifying capital. If you do not want equal risk distribution you are able to initialize the Risk Parirty Optimizer with an array of weights of your choosing.

    RiskParityOptimizer(weights_array)

  4. Hierarchical Risk Parity – Another triumph for Dr. Lopez de Prado, as he details an optimization method that does not require inverting a covariance matrix.

    HRPOptimizer()

Almost as important as your choice in optimizer is your choice in error estimator. In order to compare something you need the criteria by which to judge. In the quant world we can't just say “this thing is better than that thing”, we need to say “this this thing is better than that thing based on this measure”. The MCOS library is no different. When you call the simulate_observations() function you will have to pass it an instance of the AbstractErrorEstimator class. The current available selections for the Error Estimators are:

  1. Expected Outcome: Calculates the mean difference with respect to expected outcomes of the portfolios

    ExpectedOutcomeErrorEstimator()

  2. Variance: Calculate the mean difference in variance of the portfolios

    VarianceErrorEstimator()

  3. Sharpe Ratio: DONT USE THIS IT CAUSES THE WHOLE THING TO EXPLODE! Just kidding. As you can guess, this calculates the mean difference with respect to the Sharpe ratio of the portfolios

    SharpeRatioErrorEstimator()

You may also pass in an instance of a CovarianceTransformer if you so choose. This can be useful in helping remove some error from the simulation due to things such as noise. Currently we only have this single transformer available:

  1. Denoiser Transformer - as detailed in this paper by Dr. Lopez de Prado, this transformer helps shrinks the noise to aid in the simulation.

RETURN VALUES

The library will return to you a pandas DataFrame with the name of the optimizer, the mean of whichever error estimator you chose, and the standard deviation of the estimator.

AUTHORS

The library was constructed by the team at Enjine.

About

Implementation of Monte Carlo Optimization Selection from the paper "A Robust Estimator of the Efficient Frontier"

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages

  • Python 100.0%