Projects and Research /stor-i-student-sites/mark-holcroft Wed, 14 Jan 2026 15:50:59 +0000 en-GB hourly 1 https://wordpress.org/?v=6.9.4 /stor-i-student-sites/mark-holcroft/wp-content/uploads/sites/72/2025/01/cropped-Logo_Final-32x32.png Projects and Research /stor-i-student-sites/mark-holcroft 32 32 Robust Portfolio Optimisation Research Project /stor-i-student-sites/mark-holcroft/2025/04/23/robust-portfolio-optimisation/?utm_source=rss&utm_medium=rss&utm_campaign=robust-portfolio-optimisation /stor-i-student-sites/mark-holcroft/2025/04/23/robust-portfolio-optimisation/#respond Wed, 23 Apr 2025 14:53:13 +0000 /stor-i-student-sites/mark-holcroft/?p=240 Having done my MSci in Financial Mathematics, I was eager to spend some of my MRes year researching finance-related topics. I got this opportunity with my second research project, which was on robust portfolio optimisation.

Robustness in portfolio selection is not a novel concept, having been reviewed extensively in financial literature. It aims to address the downfalls of traditional portfolios by enforcing a lower bound on the losses that can be incurred under a range of market scenarios. The objective of the project was to assess whether classical frameworks for portfolio optimisation such as Markowitz and CVaR can be improved by adding a robust element.

Robust and non-robust versions of the CVaR and Markowitz models were created. They were each firstly trained on historical returns data for 55 stocks from between 1995 and 2025, and their split assessed using three criterion – their stock allocation, their division between industries, and their geographical spread. These are shown in the diagrams below:

Contrary to expectation, the robust models focus a larger proportion of capital into a smaller number of “safe-haven” assets. These tend to be well-established companies in traditionally safe sectors, with a focus on steady returns rather than outsized and volatile growth. Some of the companies to consistently receive a large proportion of capital include Johnson & Johnson (JNJ), The Commonwealth Bank of Australia (CBA.AX), and Nestle (NESN.SW).

The robust models show an expected exodus from risky sectors such as consumer discretionary, and an inflow into the sectors of healthcare and consumer staples which empirically weather financial shocks well.

The geographical split indicates that portfolios are dominated by American listed companies, representative of the large share of the global stock market which the US represents. The main change that the robust models exhibited is a movement from EU-based companies to Australian and Canadian entities. Although this could be indicative of factors such as a lower correlation between the latter and American stocks, it is likely this is random and based more on the specific chosen companies.

Having assessed the allocations, we next wanted to assess how well the portfolios performed. We did this firstly using an in- and out-of-sample method on the thirty years of financial data. This yielded the following graphs:

The best of the four models in the scatter plot is the robust CVaR, which has the greatest clustering in the top-left (high-return, low-variance). We are also able to pick up the 2008/2009 financial crash, shown by the four points in the bottom-right of the chart. The risk-return trade-off is similar for the four models as shown in the Sharpe ratio over time, although the CVaR maintains a slight edge over the other models in most time periods.

To further assess model performance, a copula was used to simulate returns data for testing on. A copula is a function that can model both marginal distributions as well as co-dependence structures. Forty years of data was sampled from the copula, and this was split into one-month, 21-day periods to assess short-term performance. The performances of each of the stocks are shown in the below histograms:

The robust models exhibited a greater concentration around the central values, resembling a t-distribution of the returns. There we also fewer values in the extremes, which is to be expected due to the extra precautions taken to ensure more security in the portfolio. Of greater interest though is the reduction in variances, with the robust models dominated by months of low variance, compared to the non-robust models having much longer tails and many months of high variance.

Th results of the report made it evident that robustness can have huge benefits when applied even to simple frameworks, and that the reduction in volatility does not need to come at a compromise to expected return.

The full report, including much more detail on data pre-processing, the models and general metrics achieved, can be found Here.

]]>
/stor-i-student-sites/mark-holcroft/2025/04/23/robust-portfolio-optimisation/feed/ 0
Multi-State Models in Healthcare Research Project /stor-i-student-sites/mark-holcroft/2025/04/23/multi-state-models-in-healthcare-research-project/?utm_source=rss&utm_medium=rss&utm_campaign=multi-state-models-in-healthcare-research-project /stor-i-student-sites/mark-holcroft/2025/04/23/multi-state-models-in-healthcare-research-project/#respond Wed, 23 Apr 2025 13:32:16 +0000 /stor-i-student-sites/mark-holcroft/?p=229 As part of my MRes year, I completed a research project looking at multi-state models and how they are used in healthcare. The aim was to create a report that could be read and understood by a general audience, mirroring those found in statistics magazines such as Impact.

The data used in the report was related to the severity of Cardiac Allograft Vasculopathy (CAV) of patients who had received a heart transplant. Post-transplant, patients went between “Well”, “Mild/Moderate”, “Severe” and “Death”, and it was of interest to see which patient covariates had the largest effect on patient health decline. The covariates considered were the patient age, the age of the heart donor, the patient sex, whether the patients were diagnosed with Ischaemic Heart Disease (IHD) prior to transplant, and the cumulative rejection episodes experienced by the patient. Models considering all combinations of the covariates were considered, and their Akaike’s Information Criterion plotted in Figure 1.

Based on this, we used a model including the age of the heart donor, the IHD diagnoses and the cumulative rejection episodes. For each, we investigated the hazard ratios between each pair of states, shown below.

The hazard ratios, with some exceptions, generally seemed to indicate that increases in the covariate values increase the hazard. Note that the hazard ratio for the cumulative rejection episodes and for the donor age is the increase in hazard for each episode/year, whereas IHD is binary. To test our hypothesis, we took a patient in the 20th percentile of each covariate, and a patient in the 80th percentile, and evaluated their progression between states. The graphs corresponding to the patient at the 20th percentile are shown below:

The graphs show a lower-risk profile associated with the above individual – it takes a long time for the patient to transition out of the well state, and their overall survival probability is higher than that of the overall group at all time point at a 5% level of confidence. The same graphs were also plotted for the 80th percentile individual.

The survival in this group is comparably much lower. Particularly, there is a much shorter time spent in the well state, with greater transitions to both mild/moderate and straight to death. The survival probability is also much reduced across all time points, indicating that these covariates are indeed responsible for increased risk to patients.

Such results are useful to patients and doctors as they allow treatment to be tailored to the specific needs of the patient – patients identified as being high-risk can be afforded more appointments and stronger medication, and the lowest risk patients can be allowed fewer side-effects and a fuller life.

The full article can be found Here.

]]>
/stor-i-student-sites/mark-holcroft/2025/04/23/multi-state-models-in-healthcare-research-project/feed/ 0
Deep Q-Learning Research Project /stor-i-student-sites/mark-holcroft/2024/12/02/deepqlearningtspproject/?utm_source=rss&utm_medium=rss&utm_campaign=deepqlearningtspproject /stor-i-student-sites/mark-holcroft/2024/12/02/deepqlearningtspproject/#comments Mon, 02 Dec 2024 08:22:13 +0000 /stor-i-student-sites/mark-holcroft/?p=1 For my 2024/2025 research project, I investigated whether Deep Q-Learning (DQL) could overcome the combinatorial explosion associated with the Travelling Salesman Problem. I ran DQL with different parameter tunings, finding that by adjusting the exploration decay, minimum exploration and learning rate could improve performance. I then compared with the Genetic Algorithm (GA) for both its solution cost and runtime. Some of the results are shown below:

Performance of DQL with various parameter combinations.
Several parameter combinations run again with a minimum exploration rate of 0.05.
Cost vs Iterations comparison of DQL and GA performance

The results found showed that DQL could indeed be applied to the TSP, but the results were inferior to that of the GA, both with regards to runtime and final cost.

My full report can be viewed .

]]>
/stor-i-student-sites/mark-holcroft/2024/12/02/deepqlearningtspproject/feed/ 1