Next Article in Journal
Dataset-Aware Preprocessing for Hippocampal Segmentation: Insights from Ablation and Transfer Learning
Previous Article in Journal
Statistical Analysis of the Induced Ailamujia Lifetime Distribution with Engineering and Bidomedical Applications
Previous Article in Special Issue
Temperature-Compensated Multi-Objective Framework for Core Loss Prediction and Optimization: Integrating Data-Driven Modeling and Evolutionary Strategies
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bi-Objective Portfolio Optimization Under ESG Volatility via a MOPSO-Deep Learning Algorithm

1
Department of Economics, Law, Cybersecurity, and Sports Sciences, University of Naples “Parthenope”, Guglielmo Pepe Street, 80035 Nola, Italy
2
Department of Economics, Business, Mathematics and Statistics, University of Trieste, A. Valerio Street, 4/1, 34127 Trieste, Italy
3
Department of Management and Quantitative Studies, University of Naples “Parthenope”, Generale Parisi Street, 13, 80132 Naples, Italy
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(20), 3308; https://doi.org/10.3390/math13203308
Submission received: 10 September 2025 / Revised: 8 October 2025 / Accepted: 13 October 2025 / Published: 16 October 2025
(This article belongs to the Special Issue Multi-Objective Optimization and Applications)

Abstract

In this paper, we tackle a bi-objective optimization problem in which we aim to maximize the portfolio diversification and, at the same time, minimize the portfolio volatility, where the ESG (Environmental, Social, and Governance) information is incorporated. More specifically, we extend the standard portfolio volatility framework based on the financial aspects to a new paradigm where the sustainable credits are taken into account. In the portfolio’s construction, we consider the classical constraints concerning budget and box requirements. To deal with these new asset allocation models, in this paper, we develop an improved Multi-Objective Particle Swarm Optimizer (MOPSO) embedded with ad hoc repair and projection operators to satisfy the constraints. Moreover, we implement a deep learning architecture to improve the quality of estimating the portfolio diversification objective. Finally, we conduct empirical tests on datasets from three different countries’ markets to illustrate the effectiveness of the proposed strategies, accounting for various levels of ESG volatility.

1. Introduction

Diversification constitutes a key feature of portfolio management and has long been central to the decision-making process under risk and uncertainty in economics and finance [1]. In the framework of portfolio theory (e.g., the classical mean–variance model [2] and expected utility theory), diversification is based on asset dependence. Ceteris paribus, the weaker the positive dependence among portfolio assets, the greater is the portfolio diversification, thereby reducing simultaneous adverse performances across assets [3].
Therefore, the notion of diversification is linked to the portfolio risk. Nevertheless, the formalization of diversification remains conceptually complex, and debate persists regarding the most appropriate measure [4]. Recent contributions highlight the importance of constructing well-diversified portfolios and have proposed different approaches, such as Risk Parity, and entropy-based measures (see, e.g., [5,6]). Recently, the academic literature has underlined the diversification benefits of sustainable investments, particularly when green assets are incorporated into traditional bond and equity portfolios. Hoepner [7], for instance, states that Environmental, Social, and Governance (ESG) criteria can strengthen portfolio diversification by mitigating specific asset risks.
ESG scores offer a comprehensive evaluation of a company’s performance across the Environmental (E), Social (S), and Governance (G) dimensions, providing a measure of exposure to non-financial risks and long-term opportunities related to sustainable practices. The environmental pillar reflects the impact on natural resources and emissions; the social pillar captures aspects such as labor standards and stakeholder relationships; while the governance pillar measures the quality of corporate management and shareholder rights.
According to a report from Morgan Stanley (https://www.morganstanley.com/content/dam/msdotcom/en/assets/pdfs/Sustainable-Reality_2022_Final_CRC-5440003.pdf, (accessed on 11 August 2025)), in August 2020, the market value of sustainable funds surpassed USD 1 trillion for the first time. Both institutional and retail investors increasingly allocate capital to ESG-linked securities, driven by social pressures, ethical motivations, and climate-related awareness. While early approaches relied just on screening criteria, recent strategies have shifted toward ESG integration, reflecting heightened sensitivity to environmental and governance issues [8] (see also references therein). In this setting, ESG-linked signals start to acquire an interest comparable to the traditional financial indicators. For example, ESG signals related to the volatility of ESG scores may track a new source of systematic risk.
Such considerations enter into a classical multi-objective paradigm, where competing criteria must be jointly optimized.
Since the seminal contribution of Harry Markowitz in the 1950s [2], introducing the mean–variance portfolio selection framework, multi-optimization models have become fundamental tools in financial decision-making. Indeed, the domains of finance and insurance present substantial challenges for the design and implementation of advanced optimization methodologies. These challenges arise from the uncertainty and dynamic nature of financial systems, the multiplicity of stakeholders involved, the complexity characterizing global financial markets, the wide variety of financial products, and the increasing availability of large-scale data for analysis. Within this context, multi-objective optimization models are particularly relevant, where a unique optimal solution does not exist, and decision-makers’ preferences play a crucial role in selecting among Pareto-efficient alternatives. An example is the classical portfolio selection problem, where risk and return typically serve as the two primary criteria. This bi-objective framework has been extensively studied in the literature and is widely employed in practice. Nevertheless, risk is a multi-faceted notion with different origins, including systematic, unsystematic, and systemic sources, which have motivated the development of diverse risk measures extending beyond return variance. In addition, organizational and investor-specific policies (e.g., liquidity requirements, diversification preferences) and emerging considerations such as corporate social responsibility and regulatory compliance increasingly influence financial decision-making. The coexistence of these heterogeneous and often conflicting dimensions gives rise to practical challenges, and therefore the importance of the multi-objective optimization (see, e.g., [9,10,11], and references therein).
One of the first contributions to sustainable portfolio optimization problems incorporates ESG considerations as constraints within the classical Markowitz mean–variance model (see [12]). A significant advancement has been achieved by Utz et al. [13], who introduced sustainability as an objective function of the optimization problem, extending the efficient frontier into the tri-dimensional variance-profit-sustainability space. In this direction, the research activity has increased noticeably, with some approaches aiming to minimize risk and maximize diversification, and others integrating ESG criteria into classical financial metrics such as the mean, Sharpe ratio, and tracking error (see, e.g., [14,15,16,17,18]).
The present study contributes to this growing literature by introducing a novel bi-objective portfolio optimization framework that aims to minimize an augmented volatility, where ESG signals related to the volatility of ESG scores are incorporated, and to maximize the portfolio diversification à la Carmichael–Koumou–Moran [19].
To address this optimization problem, we develop a hybrid Multi-Objective Particle Swarm Optimization (MOPSO) algorithm, enhanced with a deep learning architecture and a tailored constraint-handling mechanism. For further details, we refer to Section 2 and Section 4.
Therefore, the novelties of this manuscript are threefold:
(i)
We introduce as the objective function an augmented portfolio volatility, where the standard financial volatility is expanded with ESG volatility signals via a sustainable-risk-aversion parameter;
(ii)
We analyze the interplay between this new augmented volatility and the portfolio diversification à la Carmichael–Koumou–Moran;
(iii)
We develop an improved MOPSO solver with a suitable project-and-repair mechanism to satisfy the constraints, which is also enhanced by a Long Short-Term Memory (LSTM) neural network architecture to better estimate the diversification measure.
It should be noted that, among the common approaches in the literature related to the use of ESG scores (screening, constraints, or objective-function integration), our contribution is in line with the third category, as ESG information is directly incorporated into the objective function through the augmented volatility measure (see, e.g., [8,20], and references therein.)
Findings from empirical analysis indicate that the introduction of the ESG component tends to generate more diversified portfolios, compared with the case in which purely financial risk is considered.
This work could also give some hints for the sustainable credits literature. Indeed, the ESG factors not only contribute to long-term value creation but also generate positive externalities for the broader socioeconomic environment. However, the efficient incorporation of such information into market valuations remains uncertain. Zeidan and Onabolu [21] state that sustainability credit scoring systems, which operate analogously to conventional credit rating models, represent an effective mechanism to enhance their lending practices. Such a framework should be designed to capture the full spectrum of potential outcomes by integrating both positive and negative potential interactions between environmental and social determinants and financial performance, rather than focusing exclusively on downside risks (see, e.g., [22,23,24]).
The remainder of the paper is then structured as follows. In the next section, we review the literature concerning the MOPSO solver and the deep-learning architecture. Section 3 introduces the mathematical model and its justification, while Section 4 presents the characteristics of the MOPSO algorithm and the LSTM design. Section 5 provides results related to the monthly efficient frontiers, describing the best trade-off between the augmented volatility and the portfolio diversification, and the profitability analysis of different strategies depending on a sustainable-risk-aversion parameter tested on three datasets. Section 6 offers concluding remarks and suggestions for future research directions. Finally, in Appendix A one could find the normality checks for the data and the goodness-of-fit tests for the LSTM architecture, while Appendix B is devoted to a deeper analysis concerning the Pareto front outcomes and the optimal weight allocations.

2. Literature Review

This section is devoted to a brief overview of the literature concerning the MOPSO solver and the neural network (NN) architecture.

2.1. MOPSO Solver

This section is focused on the literature linked to the Particle Swarm Optimization (PSO) algorithm and its extension to the multi-objective framework. The PSO has been initially designed for single-objective optimization in [25], and it has gained popularity due to its simplicity in implementation, efficiency, and ability to handle complicated search spaces. Over time, PSO has been improved to overcome the stagnation and local trapping issues. As a result, mutation and repair mechanics have been introduced (see, e.g., [26,27,28,29,30]). For a more recent application of a hybrid version of PSO with neural network architecture, we refer to [20], where the neural component is employed to calibrate the three key PSO parameters ( c 1 , c 2 , w) and to generate estimates that directly feed into the objective function of the optimization problem.
Since real-world problems typically involve multiple decision objectives, the PSO paradigm has been extended to the Multi-Objective Particle Swarm Optimization (MOPSO). In this context, new strategies appear as the Pareto dominance, the leader selection mechanism, and the external archive based on the crowding distance, in order to effectively navigate in the multi-dimensional decision space.
The pioneering paper concerning the extension of PSO to multi-objective optimization problems goes back to [31]. In such a study, it has been shown that swarm-based approaches can handle multiple conflicting objectives effectively, laying the foundation for later adaptations. Afterward, in [32], the authors introduced a MOPSO algorithm incorporating Pareto dominance and an external archive to maintain solution diversity and improve convergence. The use of a double mutation strategy has been implemented in a subsequent paper [33]. In [34], a comparative analysis of MOPSO algorithms, highlighting challenges in maintaining solution diversity, is presented. The study introduces a MOPSO variant, which incorporates a velocity constraint mechanism to improve stability and diversity preservation. These findings highlight the importance of the clamping and repair mechanism in the MOPSO framework. Finally, in reference [35] is presented a detailed review of MOPSO advancements, categorizing modifications based on leader selection, archive repository, and velocity updates. They have stressed the fundamental role of mutation operators, adaptive mechanisms, and hybrid approaches in improving the standard MOPSO solver.

2.2. Neural Network Architecture

Neural networks (NN) represent a class of deep-learning-based models inspired by the functioning mechanism of the human brain. They are composed of interconnected computational units, called neurons, organized into multiple layers. Each neuron processes incoming data and transmits the transformed information to the following layers in order to provide an outcome that is as accurate as possible.
During the last decades, neural networks have become increasingly relevant in finance thanks to their ability to model non-linear relationships and to effectively handle large-scale datasets (see [36,37], among others). Furthermore, they are known as useful and potential tools for estimating quantities when no closed-form formula exists or can be applied, making them powerful approximation models. In addition to finance, neural networks have also been successfully applied in a wide range of fields, including computer vision [38] and natural language processing [39], making them a versatile tool for different goals.
In the literature, different types of architectures have been defined according to the structure of connections among units. Limiting our attention to the models at the basis of the architecture adopted in this work, the starting point is represented by Feed-Forward (FF) neural networks, in which information propagates only forward, from the input layer to the output layer [40]. In this type of network, each layer processes the information from the previous one, progressively extracting outcomes that are functional to the learning task. These networks represent the foundation for the evolution of neural architectures and a reference for numerous applications.
A natural extension is provided by Recurrent Neural Networks (RNNs), introduced by Elman [41]. Unlike the FF networks, they are characterized by cyclic connections that allow the output generated in previous steps to be reused as input. In this case, predictions take into account the past information, making these architectures particularly suited for processing temporal or sequential data. However, training a vanilla RNN is often limited by the vanishing gradient issue, which makes this architecture less effective in the presence of long-term dependencies.
To overcome these limitations, more sophisticated architectures have been developed, including the Long Short-Term Memory (LSTM) networks [42]. LSTM, compared with the RNNs, integrates additional memory cells that are able to store and process the long-term information [43]. In the literature, this type of network has shown promising results in complex domains, such as machine translation or modeling music and speech sequences, leading to their following application in other fields, including finance (another widely used variant is represented by Gated Recurrent Units (GRUs), introduced by Cho et al. [44], which simplify the structure of LSTM by reducing the number of parameters and, in some cases, show greater efficiency).

3. Mathematical Problem

This section is dedicated to the description of the theoretical framework for our optimization task, and we also present the motivation behind the problem.
The sustainable portfolio bi-objective optimization problem can be mathematically written at every fixed time t = 1 , , T as follows,
min w i , j = 1 N w i w j λ σ i j f i n + ( 1 λ ) σ i j E S G ( minimize the augmented volatility ) max w 1 2 i , j = 1 N d i j w i w j ( maximize the portfolio diversification )
such that the following constraints are satisfied:
i = 1 N w i = 1 ( budget )
0 w i 1 i = 1 , , n ( box ) .
The vector w = ( w 1 , , w N ) R N represents the N weights of a portfolio, where N indicates the number of assets of the investable universe and w i R is the proportion of wealth invested in the i-th asset, for all i = 1 , , N . Each i-th asset is associated to a real-valued return (per period) r i , and each pair of assets ( i , j ) are associated to a real-valued financial covariance σ i j f i n . The matrix Σ f i n R N × N is symmetric and the diagonal element σ i i is the financial variance of i-th assets. The matrix Σ E S G is the sustainable counterpart, where Σ E S G = σ i j E S G i , j = 1 N . To build such a matrix Σ E S G , we adopt an approach analogous to [45], in which a return-like quantity is derived using ESG data as a proxy for a company’s sustainability. Following [46], we treat sustainability-related returns as stochastic, since managers’ good practices are ex-ante impossible to predict. For the empirical implementation, we employ ESG ratings from Refinitiv, which are reported on a 0–100 scale. While this standardized scaling facilitates the evaluation of a firm’s relative sustainability performance, we assess whether a firm’s ESG rating outperforms or underperforms with respect to industry standards over successive periods. Accordingly, we define the sustainable-return counterpart as the rate of return in ESG performance for each asset i at time t. Specifically, following [45], let the standardized relative sustainability performance be
s p i ( t ) = E S G i ( t ) E S G ¯ ( t ) ,
where E S G ¯ ( t ) is the average ESG rating of all companies in the current rating universe across industries. Therefore, we can define the sustainable-return as
s r i ( t ) = s p i ( t ) s p i ( t 1 ) 1 .
A positive sustainable-return indicates that the firm has improved its relative ESG rating over a given period, thereby reflecting progress in the adoption of ESG-oriented practices. Hence, analogously to the financial counterpart, we build the sustainable-covariance matrix, denoted by Σ E S G , which captures the dynamics of ESG-related risk over time. Formally, for each asset i the expected return of ESG at time t is computed over the past L observations as
s r ¯ i ( t ) = 1 L = 1 L s r i ( t ) .
The ESG covariance (it should be noted that if i = j , σ i i E S G , ( t ) corresponds to the ESG variance of asset i) between assets i and j at the same time is defined by
σ i j E S G , ( t ) = 1 L = 1 L s r i ( t ) s r ¯ i ( t ) s r j ( t ) s r ¯ j ( t ) .
Accordingly, the ESG covariance matrix at time t is
Σ E S G , ( t ) = σ i j E S G , ( t ) i , j = 1 N
and the ESG volatility of asset i is given by the standard deviation
σ i E S G , ( t ) = σ i i E S G , ( t ) .
For the sake of simplicity, in the following we drop the explicit time index and simply write Σ E S G = σ i j E S G i , j = 1 N .
By construction, ESG volatility satisfies the standard properties of a variability index (see, e.g., [47] among others): (i) non-negativity and (ii) being equal to zero if and only if all ESG returns are identical across the sample. Furthermore, it also satisfies homogeneity under linear transformations, namely σ ( Y ) = | α | σ ( X ) if Y = α X + β .
Given this framework, the augmented volatility to be minimized in (1) is then given by the combination of the two matrices Σ f i n and Σ E S G , where λ [ 0 , 1 ] is the sustainable-risk-aversion parameter which describes the agent’s appetite for the ESG volatility information. It is worth stressing that, since ESG returns are defined in analogy with financial returns, both Σ f i n and Σ E S G are expressed in homogeneous units of variability. Hence, λ acts as a preference parameter between financial and sustainable risk.
As a second aim, we aim to maximize the diversification of the portfolio, which is encoded through the dissimilarity matrix D = ( d i j ) i , j = 1 N , where d i j measures the degree of dissimilarity between asset i and asset j. Following the approach proposed in [19], this dissimilarity is defined as
d i j = ( 1 ρ i j ) k = 1 N CVaR α ( r k ) N · CVaR α ( r i ) 1 2 · k = 1 N CVaR α ( r k ) N · CVaR α ( r j ) 1 2
with α = 0.05 and ρ i j is the Pearson correlation coefficient between asset i and asset j (this formulation implies that pairs of assets with low (or negative) correlations are associated with higher dissimilarity values, reflecting a high potential for diversification, as also stated in [19]). In (4), the notion of Conditional Value-at-Risk (CVaR) is introduced as
CVaR α ( r i ) = E [ r i r i VaR α ( r i ) ]
where VaR α ( r i ) = inf { R R : P ( r i R ) α } .
It is worth noting that the CVaR is a coherent risk measure (see, e.g., [48]) capturing the average tail loss, and r i has a continuous and strictly increasing distribution (see also [49]).
Finally, in (1) we consider the standard constraints given by the budget and box requirements. The budget constraint (2) translates the requirement that all the available capital has to be invested. At the same time, the weight of asset i has to be in the range [ 0 , 1 ] in order to exclude short sales.
Remark 1.  
It is straightforward to show that the augmented matrix presented in the objective function of problem (1) is positive semidefinite since it is a convex combination of positive semidefinite matrices.
Remark 2.  
We underline that our optimization process is rebalanced month-by-month for a total of T = 60 rebalancing phases. Rebalancing occurs on the last business day of each month, and trades are executed at the closing prices.

3.1. Motivation of the Problem

One of the two objectives of the optimization problem defined in (1) is minimizing the augmented volatility, which incorporates the uncertainty component associated with the variability of ESG scores. To motivate this choice, we explore the correlation structure among the ESG of the different constituents of a financial index. To better understand this exploration, in Figure 1 we show, for three financial indices—the Dow Jones Industrial (USA), the Euro Stoxx 50 (Europe), and the FTSE 100 (UK), provided by Refinitiv (www.refinitiv.com)—two heatmaps: one relating to the correlation matrix of the daily returns of assets (left side), and the latter referring to the correlation matrix of the ESG scores (right side). Pearson correlation coefficients of the two matrices are calculated on daily data over the period 2006–2020, and they express the degree of linear co-movement for each pair of securities, both from a financial perspective and from a sustainability point of view.
For returns (maps on the left), as expected, correlations between couples of assets are mainly positive and medium-high, confirming that stocks belonging to the same market tend to move together in response to macroeconomic shocks and systemic factors.
On the other hand, the three maps relating to ESG variability are characterized by greater heterogeneity. It should be noted that there exist several blocks associated to positive correlations, which indicate similar trends for different pairs of stocks. Nevertheless, several correlations are almost absent or even negative, indicating that an improvement in the ESG profile of a company corresponds to a worsening in the same profile for another one. This phenomenon can be attributed to the different ways in which companies balance the environmental (E), corporate behavior (S), and governance (G) strategies to improve their ESG profile. As a result, ESG scores do not evolve independently over time, but follow a dynamic structure of interconnection among them. To take this dynamics into account, we introduce an augmented volatility, which combines the financial risk component with the sustainable one.
In the previous analysis, the heatmaps were created from the three different datasets, which cover the United States (Dow Jones Industrial), Europe (Euro Stoxx 50), and the United Kingdom (FTSE 100).
We focus on these three different markets that comprise the largest and most liquid stocks in their respective regions, making them suitable for portfolio construction and risk assessment. Furthermore, their different market structures and ESG disclosure practices enable us to test whether our bi-objective optimizer behaves consistently in each different geographical context.
Each dataset consists of daily prices and ESG ratings (provided by Refinitiv) covering the period from the 6th October 2006 to the 31st December 2020 for a total of 3715 observations. More details about the datasets can be found in Table 1.
The empirical panel is balanced since only full-history assets are considered. Datasets are publicly available at [50].
More specifically, we employ an out-of-sample window, from January 2016 to December 2020, to point out the effects of the market changes on the behavior of the investments. Furthermore, the in-sample window is represented by the period from January 2012 to December 2015, updated in each rebalancing phase.

3.2. Scenario Generation Technique

To simulate the rate of returns, we use a bootstrap procedure [51]. More specifically, we consider a stationary bootstrap that employs a random block size (i.e., a set of consecutive scenarios) to bring some robustness with respect to the standard block bootstrap (which instead employs fixed block size). We start by selecting the optimal average block size B * through the technique presented in [52], which is fully automatic. Then, we randomly extract, from the in-sample data, a block of length B * . We repeat this procedure until the extracted sample reaches the desired size H (we remark that H represents the number of business days in a month), adjusting the last block length if necessary.
Finally, we a simulated bootstrap sample of H rates of simple-return for each asset i, denoted as r ^ i ( t ) , with i = 1 , , N and t = 1 , , H . We calculate the simulated H-step ahead rate of simple-return of the i-th asset by r ^ i ( H ) = t = 1 H 1 + r ^ i ( t ) and the simulated H-step ahead rate of simple-return of portfolio w in the following way, r ^ w ( H ) = i = 1 N w i r ^ i ( H ) . We repeat this procedure S times to have an estimate of the empirical distribution.
In this paper, the number of scenarios is set at 10,000. For all the details, we refer the interested reader to [52].

4. Overview of the Solver

We describe below the hybrid solver developed to tackle the bi-objective problem presented in (1). Firstly, we illustrate the structure of the MOPSO solver (implemented in MATLAB R2025a) with some improvements. Next, we design the LSTM architecture needed to effectively estimate the dissimilarity matrix.

4.1. MOPSO Description

Multi-Objective Particle Swarm Optimization (MOPSO) was introduced in an unpublished document and subsequently proposed by Moore and Chapman [31] as an extension of the standard PSO algorithm. MOPSO is specifically designed to tackle optimization problems involving multiple conflicting objectives, a common occurrence in real-world scenarios. Unlike single-objective optimization, where the goal is to find a single optimal solution, multi-objective optimization seeks to find a set of solutions that represent the trade-offs among competing objectives, the so-called Pareto front. We present in Figure 2 the flowchart of the algorithm.
While in the standard PSO algorithm, each particle has one global best position to track and move toward to, in MOPSO, each particle can have multiple potential leaders, each representing a different Pareto optimal solution. The particle selects one of these leaders at each iteration to update its position and velocity, allowing the swarm to explore multiple optimal regions of the solution space simultaneously. These multiple leaders are stored in an external archive. The archive serves as a repository for the best-found solutions, which are stored based on the Pareto dominance principle.
The archive plays a crucial role in ensuring that the solutions found by the swarm remain diverse and representative of the entire Pareto front. In this respect, we employ the Pareto dominance principle and the crowding distance mechanism (see [32]).
In addition to modifying the swarm dynamics, MOPSO also corrects the positions and velocities of the particles that exceed the boundaries. The velocity update rule is designed to balance exploration (searching for new regions of the solution space) with exploitation (refining already discovered solutions). By doing so, MOPSO is able to effectively navigate complex, multi-dimensional solution spaces and identify high-quality solutions that lie along the Pareto front.
Finally, to improve the basic MOPSO, we implement a mutation strategy that divides the swarm into three subgroups. The first subgroup is not affected by the mutation process, the second one is mutated through a uniform mutation, and the last one is enhanced via the mutation operator developed by Coello et al. in [32].
Algorithm 1 describes the pseudocode of the MOPSO algorithm.
Algorithm 1: Multi-Objective Particle Swarm Optimization (MOPSO)
1:
Input: Number of particles N p o p , search space bounds [ 0 , 1 ] , maximum iterations m a x _ i t e r , inertia weight ω , social coefficient c 1 , and cognitive coefficient c 2
2:
Output: Pareto front archive with non-dominated solutions
3:
Initialize N p o p particles with random positions w p and zero velocities v p
4:
Project the particles in the feasible region via (5) and (6)
5:
Evaluate objective functions f 1 ( w p ) , f 2 ( w p ) in (1) for each particle p
6:
Set personal best position w b e s t p = w p and cost c o s t b e s t p = [ f 1 ( w p ) , f 2 ( w p ) ]
7:
Initialize archive to store non-dominated solutions
8:
for  i t e r = 1 to m a x _ i t e r  do
9:
   for each particle p do
10:
     Select global best g b e s t randomly from archive if not empty, otherwise from particles
11:
     Update velocity component-wise: v i p = ω · v i p + c 1 · r 1 · ( g b e s t i w i p ) + c 2 · r 2 · ( w b e s t , i p w i p )
12:
     Update position component-wise: w i p = w i p + v i p
13:
     Apply the mutation mechanism with percentage m u t %
14:
     Correct the positions and velocities of the particles that exceed the boundaries and project them into the feasible region
15:
     Evaluate objectives: f 1 ( w p ) , f 2 ( w p )
16:
     if current solution dominates personal best then
17:
        Update personal best: w b e s t p = w p and c o s t b e s t p = [ f 1 ( w p ) , f 2 ( w p ) ]
18:
     end if
19:
     Update archive through the Pareto principle:
20:
     Remove dominated solutions by w p
21:
     if  w p is not dominated by any archive member then
22:
        Add w p to archive
23:
     end if
24:
   end for
25:
   if archive size exceeds A m a x  then
26:
     Reduce archive by crowding distance mechanism to maintain A m a x solutions
27:
   end if
28:
end for
29:
Return: Archive containing Pareto optimal solutions
The initialization phase ensures that the swarm begins with a well-distributed set of candidate solutions with zero velocities. In line 3, each of the N p o p particles, for i = 1 , , N , is initialized with a random position w i p within the problem’s search bounds [ 0 , 1 ] (in our context) and assigned a velocity v i p = 0 . In line 4, after defining the initial positions, the particles are projected into the feasible region through a project-and-repair mechanism (see [20]). For further details of such a process, we refer to Section “Project-and-Repair Mechanism”.
In line 5, the objective functions f 1 ( w p ) and f 2 ( w p ) of (1) are evaluated for each particle via the use of the LSTM framework (see Section 4.2). These function values determine the quality of the initial solutions.
In line 6, the personal best position w b e s t p for each particle is set to its initial position, and its corresponding objective values are stored as c o s t b e s t p = [ f 1 ( w p ) , f 2 ( w p ) ] . The personal best is taken as a reference for future updates, ensuring that each particle tracks its own progress. In line 7, an archive is initialized to store non-dominated solutions found throughout the optimization process. This archive is employed as an external memory to approximate the Pareto front by retaining the best trade-off solutions encountered.
The main loop of the algorithm executes the optimization process, iterating through the swarm until the maximum number of iterations m a x _ i t e r is reached, which is the stopping criteria adopted in this paper. For each iteration, the algorithm processes all the particles sequentially and performs the following steps: in line 10, the global best g b e s t is selected randomly from the archive if it contains solutions; otherwise, it is chosen from the swarm. In line 11, the velocity is updated via the formula v i p ( i t e r + 1 ) = ω · v i p ( i t e r ) + c 1 · r 1 · g b e s t i ( i t e r ) w i p ( i t e r ) + c 2 · r 2 · w b e s t , i p ( i t e r ) w i p ( i t e r ) , combining the particle’s previous velocity ω · v i p ( i t e r ) , its attraction toward the global best position g b e s t i ( i t e r ) , and its attraction toward its own best position found so far w b e s t , i p ( i t e r ) . In addition, ω , c 1 , c 2 are fixed parameters and r 1 , r 2 are random numbers generated at every iteration uniformly distributed in [ 0 , 1 ] . In line 12, the position of the particle is updated by adding the new velocity to its current position. To enhance the ability of our solver, we introduced a mutation mechanism in line 13. If the updated position exceeds the search bounds, it is corrected via the project-and-repair process to ensure feasibility (line 14). After updating positions, in line 15, the objective functions are re-evaluated for the new positions. In lines 16–18, if the new position dominates the previous personal best, the particle’s personal best position is updated accordingly. This ensures that each particle preserves the best trade-off it has encountered during the optimization process. In lines 19–23, the archive is updated to maintain only non-dominated solutions. First, solutions dominated by the newly evaluated position w p are removed. Then, if w p itself is not dominated by any existing archive member, it is added to the archive. Since the archive size can grow indefinitely, in lines 25–27, if it exceeds a prefixed limit A m a x , solutions are removed via the crowding distance evaluation. This mechanism controls memory usage while preserving solution diversity. All of the process is repeated until the maximum number of iterations is reached.
In line 29, the algorithm returns the final archive, which contains a well-distributed set of non-dominated solutions forming the desired Pareto front. These solutions represent the best trade-offs found among the objectives.
Remark 3. 
We remark that the augmented volatility objective function f 1 ( w ) , which involves both the financial covariance matrix Σ f i n and the ESG covariance matrix Σ E S G , is evaluated through the shrinkage estimator proposed in [53]. In this way, shrinkage regularization is applied to the entire augmented covariance structure. On the other hand, the portfolio diversification f 2 ( w ) is estimated thanks to the deep-learning architecture presented in Section “A Multivariate Quantile-LSTM for CVaR Estimate”.
Remark 4.  
We stress that the implementation of the LSTM paradigm in the MOPSO solver to enhance the estimate of the objective function is new in this context.

Project-and-Repair Mechanism

Since we introduced a mutation mechanism in line 13, it may happen that the particles no longer satisfy the constraints (2) and (3). Therefore, we implement a project-and-repair technique which, as a first step, employs the following standard projection:
w i p = min max w i p , 0 , 1
where p = 1 , , N p o p and i = 1 , , N . It is worth noting that the previous operator projects each component of the vector on the interval [ 0 , 1 ] satisfying the (3) requirement (see, e.g., [54]).
After the previous projection, the candidate solutions may give a sum different from 1. Then, as a second step of the project-and-repair mechanism, we use the repair transformations presented in [55] to manage constraint (2). In particular, for each particle p = 1 , , N p o p , we adjust the candidate solution w p as
w i p = w i p j = 1 N w j p if j = 1 N w j p > 1 w i p if j = 1 N w j p = 1 1 ( 1 w i p ) ( N 1 ) N j = 1 N w j p if j = 1 N w j p < 1
for all i = 1 , , N .
Remark 5. 
It is worth noting that the repair operator in (6) adjusts the candidate solutions such that they satisfy the sum-to-one constraint (2) while maintaining the proportions among the assets’ weight.

4.2. LSTM Design

Long Short-Term Memory (LSTM) networks represent a subfamily of Recurrent Neural Networks (RNNs), designed to overcome the vanishing gradient problem (the vanishing gradient issue consists of the tendency of the gradient to vanish or, conversely, to explode), typical of the RNNs. For this reason, LSTM results are particularly efficient when the time intervals between input and learning signals are very long.
We consider a multivariate time series z t Z R d 0 , 0 < t T , where d 0 N denotes the input dimension. Given d 1 N the dimension of the hidden layer (thus, ( d 0 , d 1 ) N 2 represent the number of units of the input and hidden layers, respectively), the dynamics of an RNN layer can be expressed as
h t ( z ) = ϕ ( W z t + U h t 1 ( z ) + b ) , t = 1 , , T
h 0 ( z ) = 0
where W R d 1 × d 0 is the weight matrix that projects the current input into the hidden state space; U R d 1 × d 1 is the weight matrix associated with the output generated from the previous step; b R d 1 is the bias vector; and ϕ : R R is an elementary activation function (typically tanh or ReLU), applied component-wise to vectors. In this way, the hidden state h t incorporates both the information from the current input z t and the memory from the previous step h t 1 , allowing the network to learn from the past. Nevertheless, as already mentioned above, the calibration of the weights ( W , U , b ) often encounters the challenge of vanishing gradients.
Unlike the RNN layer, an LSTM unit presents additional memory cells able to store and process long-term sequential data.
Formally, we denote by W ( k ) R d 1 × d 0 , U ( k ) R d 1 × d 1 , and b ( k ) R d 1 the corresponding weight matrices and bias vector, for each subnet k { i , f , C , o } . The mechanism of an LSTM layer, for t = 1 , , T , can be formalized as follows:
f t = σ W ( f ) z t + U ( f ) h t 1 + b ( f )
i t = σ W ( i ) z t + U ( i ) h t 1 + b ( i )
C ˜ t = tanh W ( C ) z t + U ( C ) h t 1 + b ( C )
C t = f t C t 1 + i t C ˜ t C 0 = 0
o t = σ W ( o ) z t + U ( o ) h t 1 + b ( o )
h t = o t tanh C t h 0 = 0
where σ ( · ) : R ( 0 , 1 ) and tanh ( · ) : R ( 1 , 1 ) are the sigmoid and the hyperbolic tangent activation functions, respectively, and C t is the memory cell state at time t.
Figure 3 represents the LSTM architecture, designed to dynamically and appropriately process the flow of information along the timeline. Information is handled by three gates: forget f t , input i t , and output o t gates. These gates define which elements of the internal memory should be retained, updated, or provided externally, respectively.
The process starts with the forget gate (see (9)), establishing which parts of the previous state can be considered irrelevant and therefore deleted. Consequently, it also determines the portion of information that will be retained for subsequent steps. The input gate (see (10)) selects the share of information that enters the internal state. In parallel, a tanh layer produces the candidate values C ˜ t (see (11)) to be integrated into the cell state together with C t 1 . At this point, the memory is updated as a linear combination of C t 1 and C ˜ t through the weights f t and i t , respectively, as stated in (12). Then, the output gate (see (13)) controls which part of the state C t should be returned. The final output of the LSTM layer, h t , emerges from the interaction between o t and the cell state (see (14)), processed by a tanh function.
From a computational perspective, each LSTM unit involves the estimation of 4 × d 1 × ( d 0 + d 1 + 1 ) parameters. Indeed, for each subnet k { i , f , C , o } , there are: the input weights W ( k ) (dimensions d 1 × d 0 ), the hidden state weights U ( k ) (dimensions d 1 × d 1 ), and the bias term b ( k ) (dimensions d 1 × 1 ). Considering the four gates (input, forget, cell, and output), the total number of parameters grows accordingly.

A Multivariate Quantile-LSTM for CVaR Estimate

Following the approach proposed in [56], a multivariate quantile LSTM (q-LSTM) is designed to simultaneously estimate the CVaR at the α quantile for each asset in the portfolio. These estimates are subsequently provided as input for the computation of dissimilarities between each couple of assets, as stated in (4). As extensively discussed in [56], this choice is motivated by the lack of a closed-form solution when some strong assumptions—such as Gaussian distributions of returns and constant conditional correlations between each asset and the portfolio over time—do not hold.
Figure 4 illustrates the architecture of the q-LSTM model, which combines an LSTM unit with fully connected (FCN) layers.
For each month t in the out-of-sample window, the LSTM unit (orange block) processes the time series of returns and generates a hidden state, represented by a latent vector. This vector serves as input to the branch responsible for the estimate of the distress quantile. The yellow block corresponds to the use of the ReLU activation function, while the blue block indicates the use of the tanh function. Initially, the LSTM output is fed into a fully connected layer (pink block) with a ReLU activation function, which produces the estimate of the Value-at-Risk at the α level for each asset (the green circle). Subsequently, the VaR values are passed as input to an additional fully connected layer with tanh activation, which yields the final output, i.e., the CVaR vector for the same quantile level α . This architecture enables the joint modeling of multivariate VaR and CVaR, ensuring internal consistency between the two measures and reducing problems associated with the quantile crossing.
Denoted by L N a lookback period, we consider the lagged time series r ( t ) = ( r i ( t ) ) 1 i N , for = 1 , , L . The hidden state h α , t R d α (with d α N ), generated from the LSTM unit, is used to compute the VaR at the α quantile. Let VaR α ( t ) = VaR α ( t ) ( r i ) 1 i N be defined as the vector of all VaR at the α quantile at time t, and it is given by
VaR α ( t ) = ϕ α ( VaR ) W α ( VaR ) h α , t + b α ( VaR )
where W α ( VaR ) R N × d α is the weight matrix, b α ( VaR ) R N is the bias vector, and ϕ α ( VaR ) is the ReLU activation function, applied with an ad hoc correction through the multiplication by ( 1 ) . Subsequently, via an FCN layer, the CVaR α ( t ) = CVaR α ( t ) ( r i ) 1 i N at time t is then generated as
CVaR α ( t ) = ϕ α ( CVaR ) W α ( CVaR ) VaR α ( t ) + b α ( CVaR )
where W α ( CVaR ) R N × N , b α ( CVaR ) R N , and ϕ α ( CVaR ) is the tanh activation function.
The chosen hyperparameter configuration is summarized in Table 2 (see bottom side). Regarding the number of neurons per layer, after some preliminary tests, we set the value to u α = 50 for the distress quantile. The dropout rate has been set to 10 % to prevent overfitting. This choice reflects that, for lower values of α , the model can better capture extreme events. We set the batch size to 8 and the number of epochs to 500 for the VaR calculation. Since the architecture is integrated within an MOPSO framework, a limited number of epochs has been chosen to limit computational times. Empirical tests confirm that this configuration is adequate and avoids overfitting (see Appendix A).
The network parameters are estimated using the backpropagation (BP) algorithm. For quantile-based forecasts, CVaR is evaluated at the quantile level α = 0.05 , as standard in the literature (see, e.g., [57]). Consistently, the pinball loss function is used, which measures the discrepancy between the 0.05 -quantile forecasts and the observed values. Weight optimization is performed via the Adam algorithm, particularly suited to deep learning because it dynamically adapts the learning rate based on the gradient’s first- and second-order moments. This adaptation allows for faster and more stable convergence, enabling the LSTM network to capture complex patterns in the data and minimize the loss function considered (see, e.g., [56] and references therein).
Remark 6. 
We remark that the inputs of the LSTM architecture are represented by the rate of returns, generated by the bootstrap technique, while the outputs are the CVaRs. Finally, the MOPSO algorithm designs the optimal weight allocation that solves the bi-objective optimization problem (1).

5. Experimental Results

In this section, we describe the parameter settings employed in the numerical analysis and present the efficient frontiers obtained month by month that show the best trade-offs between the augmented volatility and the portfolio diversification (see Section 5.1). Finally, we assess a profitability analysis for the efficient portfolio strategy that reaches the maximum Sharpe ratio for the three different sustainable-risk-aversion parameters λ { 0 , 0.5 , 1 } , for each dataset.
In Table 2, we introduce the parameter setting for the MOPSO solver and the LSTM architecture. Such a setting is based on some preliminary tests and is also coherent with the literature (see, e.g., [20,32,33,58]).
We recall that we employ an out-of-sample window, from January 2016 to December 2020, to point out the effects of the market changes on the behavior of the investments. Moreover, the in-sample window is represented by the period from January 2012 to December 2015, updated in each rebalancing phase. We execute the analysis on an Intel Core i5 with 8 GB RAM and a macOS 11.6.8 system using MATLAB R2025a.

5.1. Efficient Frontiers

We construct, for the three markets considered—Dow Jones Industrial, Euro Stoxx 50, and FTSE 100—the frontiers of efficient portfolios, i.e., the Pareto-efficient set of portfolios in the augmented volatility-diversification plane (we indicate the augmented volatility by σ ). The frontiers are generated on the last day of December for each year in the out-of-sample window (2016–2020), and they are collected in Figure 5, distinguishing the cases according to the sustainable-risk-aversion parameter λ . We recall that λ { 0 , 0.5 , 1 } , corresponding, respectively, to sustainable (green frontier), balanced (blue frontier), and purely financial (black frontier) volatility cases.
The results reveal some interesting shared features. First, we can observe that over time, diversification requires a modest increase in volatility, especially for λ = 0 and λ = 0.5 , as illustrated by vertical gaps on the left tails. This feature indicates that sustainable or balanced portfolios achieve better diversification without excessive risk costs. This behavior is repeated across the three different markets.
Looking at the differences among markets, as the number of assets increases, obtaining more diversified portfolios requires a more significant increase in volatility. This effect, once again, is more pronounced in the cases of λ = 0 and λ = 0.5 , while it is attenuated for λ = 1 . This occurs because the introduction of the sustainability component in volatility encourages a more homogeneous distribution of portfolio weights. Graphically, this greater diversification appears as a longer right tail with horizontal gaps (see, for instance, Euro Stoxx 50 and FTSE 100 in 2016–2017).
Moreover, the black curves ( λ = 1 ), where only financial volatility is considered, show a steeper shape and a shift to the right. This shape suggests that, without an ESG component, diversification requires greater increases in pure risk. However, increasing the financial volatility tends to generate portfolios that are more concentrated compared with the other two cases.
Finally, we refer to Appendix B for a deeper analysis concerning the Pareto front outcomes and the optimal weight allocations. Specifically, we included an illustrative example where it should be noted that the optimal allocations are different depending on the sustainable-risk-aversion parameter.

5.2. Profitability Analysis

We consider a frictionless market in which no short selling is allowed, and all investors act as price takers. We indicate by w t the optimal portfolio at the ex-post month t, with t = 1 , , 60 . The mean time of execution of each monthly rebalancing phase is reported in Table 3 for each dataset.
Since the investment plan is time dependent, we assume an initial wealth W 0 = 10 , 000 , 000 (in each local currency) and we explicitly evaluate the magnitude of the trading through the cost function c ( w t , w t ) introduced in [59]. In particular, concerning the transaction cost function, following [59], we considered the general complex structure that was offered by OCBC Securities, a broking firm in Singapore (www.iocbc.com), with five different trading segments, whose characteristics are reported in Table 4. For more details, we refer interested readers to Sections 3 and 5 of [59].

5.3. Ex-Post Performance Measures

In this subsection, we present the ex-post measures considered to evaluate the profitability of the investment strategies with λ { 0 , 0.5 , 1 } . We denote by r w , t e x p o s t the ex-post portfolio rate of return realized at time t, with t = 1 , , 60 . First of all, we introduce the so-called ex-post Sharpe ratio, defined as
S R = μ e x p o s t σ e x p o s t
where μ e x p o s t and σ e x p o s t are the mean and the standard deviation of the ex-post portfolio rates of return, respectively.
Further, to measure the profitability of the investment at time t, we compute the net wealth given by the following formula,
W t = W t 1 1 + r w , t e x p o s t c ( w t , w t ) .
In the previous equation, w t = ( w t , 1 , , w t , N ) represents the re-normalized portfolio weight before the rebalancing at time t (see [60]), which is defined for t = 2 , , 60 as
w t , i = w t 1 , i r t 1 , i g r o s s j = 1 N w t 1 , j r t 1 , j g r o s s
where the denominator is the gross portfolio return at month t 1 . The gross return of asset i at month t is defined as r t 1 , i gross = p i , t p i , t 1 , where p i , t is the price (adjusted for splits and dividends) of the i-th asset at the end of month t.
At time t = 1 , we set w t = 0 .
In this way, the turnover is calculated as
T R = i = 1 N | w t , i w t , i | .
Remark 7. 
According to the change in the portfolio weights between consecutive rebalancing periods, w t , i w t , i represents the trades of rebalancing the i-th asset at time t. The changes in positions during each rebalancing period are related to the transaction costs.
In addition, we evaluate the so-called compound annual growth rate, calculated as
C A G R = W 60 W 0 12 60 1
where W 0 represents the initial wealth and W 60 is the final wealth of the last month.
At this point, in order to assess the capacity of a strategy to avoid high losses, we introduce the drawdown measure [61]
D D t = min 0 , W t W p e a k W p e a k
where W p e a k is the maximum amount of wealth reached by the strategy until time t. In particular, we consider the mean (mean DD), the standard deviation (std DD), and the maximum (max DD) of the drawdown measure over time.
Finally, to understand the ex-post portfolio diversification, we exploit the diversification index [62], defined as
D I = 1 H I = 1 i = 1 N w i 2
where H I represents the Herfindahl Index.

5.4. Ex-Post Performance Analysis

In the ex-post analysis, we investigate the profitability of different allocation strategies concerning the agent’s sustainable-risk-aversion parameter λ . We recall that the selected efficient portfolio is the one that reaches the maximum Sharpe ratio on the Pareto front.
The empirical results are summarized in Table 5. We recall that for λ = 0 , the agent is completely sustainable-risk-adverse; if λ = 1 , one considers only the financial volatility, while for λ = 0.5 , there is a balance between the financial and sustainable volatilities.
Such a table shows the results for the ex-post performance measures presented in the previous Section 5.3 for the three different levels of λ { 0 , 0.5 , 1 } and for the three different datasets. Moreover, we compare such strategies with the traditional Equally Weighted (EW) portfolio allocation strategy (see [63]). Specifically, the last two rows, nnz assets (avg) and nnz assets (min), indicate, respectively, the average (it is worth underlining that the average number is rounded to the lower integer) and the minimum number of non-zero assets across the 60 out-of-sample months. In addition, c r e b represents the costs due to the rebalancing phases, calculated via the Table 4.
Finally, in Figure 6 we present the net wealth trend for all the datasets.
Remark 8. 
It is worth mentioning that the allocation strategies presented are active investing and not index-linked strategies.
In general, from the results of Table 5, we can infer that strategies with λ = 0 have higher profits but result in higher volatility values. The balanced strategy with λ = 0.5 shows better profit/risk trade-offs, presenting higher Sharpe ratio measures and generally higher diversification benefits. Finally, allocations with λ = 1 present lower overall portfolio risk. In particular, one has lower risk measures (standard deviation and drawdown). As a consequence, the turnover is better controlled as well as the rebalancing costs. The results are preserved for all three datasets, although they are more pronounced as the number of assets increases. Regarding stability, the interquartile ranges of turnover (IQRsTR) indicate the lowest dispersion at λ = 0 for the Dow Jones Industrial and Euro Stoxx 50, and at λ = 0.5 for the FTSE 100, confirming that turnover stability is dataset-specific.
The previous analysis could be justified considering that the financial volatility seems to be heavier than the sustainable one, and it influences the allocation process.
In any case, the introduction of the volatility information gives good results, showing a balanced strategy (with λ = 0.5 ) for a moderate agent who looks for a trade-off between profit and risk in the different cases analyzed, taking into account the diversification benefits.
Such an analysis is graphically corroborated by the findings in Figure 6.
Remark 9. 
It is worth stressing that all the strategies proposed outperform the Equally Weighted (EW) portfolio allocation.

6. Conclusions and Future Works

In this paper, we tackled a bi-objective optimization problem in which we aim to maximize the portfolio diversification and, at the same time, minimize the augmented portfolio volatility, where the ESG information is incorporated. Moreover, we developed an improved MOPSO embedded with a repair-and-projection mechanism to satisfy the constraints, and we implemented a deep learning architecture to improve the quality of estimating the portfolio diversification goal. Finally, we conducted empirical tests on datasets from three different countries’ markets to illustrate the effectiveness of the proposed strategies, accounting for various sustainable-risk-aversion appetites.
For future work, we plan to extend the bi-objective paradigm to a tri-objective framework where a profit measure is incorporated in the optimization process. Furthermore, we would also like to develop suitable constraint-handling techniques to keep the transaction costs under control. Finally, another interesting direction might be to examine how the results vary across the three different markets when exchange-rate effects are taken into account and to explore the use of the proposed methodology for hedging purposes.

Author Contributions

Conceptualization, I.L.A., G.B., G.S. and S.S.; methodology, I.L.A., G.B., G.S. and S.S.; software, I.L.A., G.S. and S.S.; validation, I.L.A., G.S. and S.S.; writing—original draft, I.L.A., G.B., G.S. and S.S.; supervision, G.B. All authors have equal contributions. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The original data presented in the study are openly available in [50] at [https://www.mdpi.com/2071-1050/14/4/2050], (accessed on 11 August 2025) and [20] at [https://link.springer.com/article/10.1007/s10203-025-00523-y], (accessed on 11 August 2025).

Acknowledgments

I. L. Aprea, G. Sbaiz, and S. Scognamiglio are members of the INdAM (Italian Institute for Advanced Mathematics) group. The research of I. L. Aprea was partially supported by the project Environmental Humanities and sustainable growth models operated by University of Naples “Parthenope”. The research of G. Sbaiz was partially supported by the project of the Regional Programme (PR) FSE+ 2021/2027 of the Friuli Venezia Giulia Autonomous Region—PPO 2023—Specific Programme 22/23.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Normality Check and Goodness-of-Fit Tests

In this appendix, we present the results of normality tests of returns, confirming the non-Gaussian distribution and motivating the use of an ad hoc tool for CVaR estimation. We then illustrate and discuss the goodness-of-fit of the q-LSTM model estimates. The results are provided for the three datasets referred to throughout the paper.
We start from the Dow Jones Industrial dataset, and in Table A1 we report, for all the constituents, the p-values of three tests: Anderson–Darling, Jarque–Bera, and Lilliefors. All the tests share the null hypothesis of normal distributions against an alternative of non-normality. To simplify the reading, we highlight in bold the values above the 5 percent threshold. It should be noted that, except for a few assets (Asset 7, Asset 11, Asset 12, and Asset 15), for every asset, there is at least one test for which the null hypothesis is rejected at the 5 % significance level. However, with a significance level of 10 % , according to the Jarque–Bera test, also Assets 7, 12, and 15 do not pass the test. Therefore, overall, the constituents share a non-Gaussian behavior.
Table A1. Normality tests for the Dow Jones Industrial returns. The bold formatting indicates the significant values in the table to facilitate the reading of the results.
Table A1. Normality tests for the Dow Jones Industrial returns. The bold formatting indicates the significant values in the table to facilitate the reading of the results.
AssetAnderson–
Darling
Jarque–
Bera
LillieforsAssetAnderson–
Darling
Jarque–
Bera
Lilliefors
Asset 10.00370.00100.0436Asset 150.39180.08130.4802
Asset 20.07510.14480.0217Asset 160.07060.00320.1204
Asset 30.00290.00220.0455Asset 170.00830.00200.0473
Asset 40.11090.01880.1035Asset 180.07990.01300.0274
Asset 50.18390.00270.2755Asset 190.00050.00100.0010
Asset 60.00050.00100.0010Asset 200.04260.02080.0742
Asset 70.11340.05170.2490Asset 210.00050.00100.0010
Asset 80.09260.02280.5000Asset 220.15990.02640.4203
Asset 90.02050.00890.1658Asset 230.00070.00100.0010
Asset 100.01040.00100.1288Asset 240.00050.00100.0113
Asset 110.70150.50000.5000Asset 250.00050.00100.0224
Asset 120.13670.05920.1356Asset 260.00050.01100.0010
Asset 130.00190.00640.0641Asset 270.01040.00530.0497
Asset 140.29440.00100.5000Asset 280.56940.50000.5000
Mutatis mutandis, similar considerations can be stated when we consider the other two datasets, Euro Stoxx 50 and FSTE 100. The results of tests are collected in Table A2 and Table A3, respectively.
Even in these cases, most assets show very low p-values, confirming the rejection of the normality assumption. Nevertheless, in the FTSE 100, given the greater number of stocks, there are more assets for which the normality hypothesis cannot be rejected. It should be underlined that, even in these markets, however, a part of these stocks (which are incompatible with normality at the 5 % level) could be considered non-normally distributed if observed at a looser level ( 10 % ).
Table A4 reports the evaluation metrics of the quantile forecast for each asset of the Dow Jones Industrial. Each error measure is calculated as the average over the number of months in the out-of-sample window. The pinball loss values, calculated at 5 % , are relatively low for most assets, confirming the model’s ability to predict the extreme quantiles. In particular, a few assets show higher losses, indicating a lower accuracy in forecasting extreme events. The relative loss results are also consistent with these observations, confirming the model’s robustness. Finally, the hit rate—theoretically expected to be around 5 % —shows an average value of 0.0720 across the 28 assets, a satisfactory result as it is very close to the reference benchmark. Overall, the model shows good performance.
Table A2. Normality tests for the Euro Stoxx 50 returns. The bold formatting indicates the significant values in the table to facilitate the reading of the results.
Table A2. Normality tests for the Euro Stoxx 50 returns. The bold formatting indicates the significant values in the table to facilitate the reading of the results.
AssetAnderson–
Darling
Jarque–
Bera
LillieforsAssetAnderson–
Darling
Jarque–
Bera
Lilliefors
Asset 10.01530.04050.0851Asset 240.00050.00100.0010
Asset 20.09700.01340.0658Asset 250.00050.00100.0010
Asset 30.00050.00100.0228Asset 260.26330.27590.3484
Asset 40.07940.00810.1816Asset 270.53350.01890.1399
Asset 50.22880.39990.1541Asset 280.15730.01270.3688
Asset 60.26170.01100.5000Asset 290.01390.00100.0404
Asset 70.00050.00100.0029Asset 300.31850.11090.4791
Asset 80.00050.00100.0188Asset 310.01280.00750.0072
Asset 90.00480.00100.1069Asset 320.01710.00230.1261
Asset 100.20240.11550.1550Asset 330.07630.00160.5000
Asset 110.00070.00690.0042Asset 340.00050.00100.0093
Asset 120.05020.01130.0337Asset 350.02680.00600.3609
Asset 130.00590.00100.0226Asset 360.00220.00200.0234
Asset 140.15070.08400.3124Asset 370.01510.00730.0433
Asset 150.05730.00980.0347Asset 380.00080.00180.0096
Asset 160.00050.00100.0010Asset 390.00050.00100.0010
Asset 170.00180.00100.0114Asset 400.12780.00320.1551
Asset 180.00050.00110.0010Asset 410.00050.00100.0010
Asset 190.00800.00100.1514Asset 420.00050.00100.0040
Asset 200.00050.00100.0010Asset 430.00160.00100.0011
Asset 210.00090.00900.0035Asset 440.01050.00930.0025
Asset 220.16590.04330.0086Asset 450.00050.00100.0010
Asset 230.01540.00100.0608
Table A3. Normality tests for the FTSE 100 returns. The bold formatting indicates the significant values in the table to facilitate the reading of the results.
Table A3. Normality tests for the FTSE 100 returns. The bold formatting indicates the significant values in the table to facilitate the reading of the results.
AssetAnderson–
Darling
Jarque–
Bera
LillieforsAssetAnderson–
Darling
Jarque–
Bera
Lilliefors
Asset 10.00050.00100.0077Asset 410.84910.13070.5000
Asset 20.13950.33620.5000Asset 420.04440.02720.1430
Asset 30.74010.24020.5000Asset 430.00050.00100.0059
Asset 40.26910.12200.5000Asset 440.05800.01160.1319
Asset 50.41240.29820.2738Asset 450.05670.00810.1842
Asset 60.00190.00100.0598Asset 460.00050.00100.0010
Asset 70.84640.50000.5000Asset 470.02000.00100.1629
Asset 80.00090.00100.0047Asset 480.00050.00100.0681
Asset 90.00900.01140.0450Asset 490.06980.01550.0790
Asset 100.00050.00100.0607Asset 500.00050.00100.0010
Asset 110.12880.01060.2018Asset 510.00110.00100.0352
Asset 120.38730.50000.3881Asset 520.00050.00100.0010
Asset 130.09860.08440.0321Asset 530.27250.17980.2450
Asset 140.04030.00100.1467Asset 540.01250.01110.0416
Asset 150.47040.14940.5000Asset 550.10140.01040.2645
Asset 160.00150.00100.0108Asset 560.19740.00670.3843
Asset 170.00050.00100.0010Asset 570.95190.50000.5000
Asset 180.00240.00290.0079Asset 580.61180.29890.5000
Asset 190.02410.04840.0711Asset 590.00050.00100.0010
Asset 200.08070.01230.0856Asset 600.33050.46050.3275
Asset 210.00050.00100.0010Asset 610.00050.00100.0010
Asset 220.00050.00100.0010Asset 620.00160.00100.0786
Asset 230.03140.02320.0329Asset 630.00050.00100.0010
Asset 240.00050.00100.0035Asset 640.00050.00100.0010
Asset 250.02740.00100.0450Asset 650.00210.00440.0855
Asset 260.00050.00100.0010Asset 660.25770.50000.1634
Asset 270.40400.08970.5000Asset 670.01750.00390.0455
Asset 280.00050.00100.0010Asset 680.00050.00100.0010
Asset 290.57510.13000.3636Asset 690.02840.02870.0022
Asset 300.03830.00120.1135Asset 700.02490.04090.0542
Asset 310.00050.00100.0010Asset 710.00050.00100.0010
Asset 320.00050.00100.0010Asset 720.06360.30410.0642
Asset 330.06210.00180.3543Asset 730.01350.00810.2031
Asset 340.09450.08620.1916Asset 740.00050.00100.0010
Asset 350.00050.00100.0010Asset 750.02160.00100.1127
Asset 360.00050.00100.0010Asset 760.00050.00100.0137
Asset 370.48750.05960.2852Asset 770.00160.00100.0159
Asset 380.63830.39060.5000Asset 780.02540.00100.3607
Asset 390.00110.00130.0175Asset 790.01570.00430.0691
Asset 400.05500.01870.1491Asset 800.02410.05720.0395
Table A4. Forecast evaluation metrics for each asset of the Dow Jones Industrial.
Table A4. Forecast evaluation metrics for each asset of the Dow Jones Industrial.
AssetPinball
Loss
Relative
Loss
Hit
Rate
AssetPinball
Loss
Relative
Loss
Hit
Rate
Asset 10.00270.00130.1000Asset 150.00190.00050.0333
Asset 20.00240.00100.0833Asset 160.00260.00120.1333
Asset 30.00190.00040.0667Asset 170.00150.00020.0500
Asset 40.00160.00030.0500Asset 180.00190.00060.0500
Asset 50.00130.00000.0167Asset 190.00180.00040.0500
Asset 60.00180.00040.0500Asset 200.00150.00020.0167
Asset 70.00180.00060.0333Asset 210.00260.00120.0833
Asset 80.00200.00070.0833Asset 220.00190.00030.0333
Asset 90.00260.00100.0833Asset 230.00260.00140.1000
Asset 100.00170.00030.0167Asset 240.00340.00190.1833
Asset 110.00130.00000.0000Asset 250.00310.00160.0667
Asset 120.00170.00040.0500Asset 260.00170.00040.0333
Asset 130.00390.00260.1167Asset 270.00210.00060.1000
Asset 140.00420.00310.1667Asset 280.00390.00250.1667
Similarly, Table A6 and Table A8 report the results for the Euro Stoxx 50 and FTSE 100, respectively. The pinball loss and relative loss values remain generally low, with some exceptions on more complex stocks. The average hit rate is 0.0315 for the Euro Stoxx 50, which is close to the theoretical benchmark of 0.05, which is quite satisfactory. The situation is more complex for the FTSE 100, where the average hit rate ( 0.1492 ) deviates more from the benchmark. This effect can be attributed to the higher number of assets, making maintaining accurate forecasts for all the stocks more difficult. In summary, the model provides robust and reliable estimates across the different markets.
Table A5. Forecast evaluation metrics for each asset of the Euro Stoxx 50.
Table A5. Forecast evaluation metrics for each asset of the Euro Stoxx 50.
AssetPinball
Loss
Relative
Loss
Hit
Rate
AssetPinball
Loss
Relative
Loss
Hit
Rate
Asset 10.00200.00010.0333Asset 240.00200.00000.0167
Asset 20.00220.00030.0333Asset 250.00310.00090.0500
Asset 30.00180.00000.0167Asset 260.00220.00030.0667
Asset 40.00180.00010.0167Asset 270.00180.00000.0000
Asset 50.00230.00040.0667Asset 280.00220.00050.0333
Asset 60.00190.00010.0167Asset 290.00170.00000.0167
Asset 70.00220.00020.0500Asset 300.00170.00000.0000
Asset 80.00210.00030.0333Asset 310.00210.00030.0167
Asset 90.00280.00090.0833Asset 320.00200.00010.0333
Asset 100.00230.00050.0333Asset 330.00280.00090.0333
Asset 110.00190.00010.0333Asset 340.00180.00000.0000
Asset 120.00180.00000.0000Asset 350.00260.00070.0500
Asset 130.00200.00020.0333Asset 360.00190.00010.0333
Asset 140.00170.00000.0167Asset 370.00230.00040.0167
Asset 150.00180.00000.0000Asset 380.00200.00030.0167
Asset 160.00210.00020.0333Asset 390.00220.00050.0333
Asset 170.00280.00080.0833Asset 400.00180.00000.0167
Asset 180.00240.00050.0333Asset 410.00280.00100.0333
Asset 190.00210.00010.0167Asset 420.00490.00340.0333
Asset 200.00300.00120.0833Asset 430.00230.00050.0333
Asset 210.00180.00000.0167Asset 440.00230.00040.0500
Asset 220.00190.00000.0333Asset 450.00200.00010.0167
Asset 230.00200.00030.0500
Table A6. Forecast evaluation metrics for each asset of the FTSE 100.
Table A6. Forecast evaluation metrics for each asset of the FTSE 100.
AssetPinball
Loss
Relative
Loss
Hit
Rate
AssetPinball
Loss
Relative
Loss
Hit
Rate
Asset 10.00180.00070.1000Asset 410.00200.00100.1167
Asset 20.00300.00210.1500Asset 420.00390.00270.2000
Asset 30.00190.00100.1333Asset 430.00210.00090.1000
Asset 40.00130.00030.0500Asset 440.00250.00150.1667
Asset 50.00310.00190.1333Asset 450.00170.00070.0333
Asset 60.00420.00310.2333Asset 460.00410.00330.2000
Asset 70.00160.00070.1333Asset 470.00440.00310.2167
Asset 80.00180.00070.1167Asset 480.00240.00140.1000
Asset 90.00210.00110.1333Asset 490.00290.00190.1500
Asset 100.00280.00180.1500Asset 500.00240.00120.1167
Asset 110.00250.00170.1667Asset 510.00410.00290.2167
Asset 120.00260.00150.1167Asset 520.00270.00180.1333
Asset 130.00150.00050.1000Asset 530.00140.00030.0833
Asset 140.00250.00130.1167Asset 540.00190.00090.1667
Asset 150.00200.00090.1167Asset 550.00190.00080.1333
Asset 160.00320.00210.1500Asset 560.00300.00200.1833
Asset 170.00390.00270.2000Asset 570.00220.00120.1167
Asset 180.00160.00070.0833Asset 580.00240.00150.1500
Asset 190.00290.00200.1833Asset 590.00310.00200.2000
Asset 200.00290.00180.1500Asset 600.00250.00160.1167
Asset 210.00510.00390.2333Asset 610.00350.00230.2000
Asset 220.00440.00310.2000Asset 620.00330.00240.1333
Asset 230.00170.00070.1000Asset 630.00450.00320.2500
Asset 240.00160.00050.1000Asset 640.00480.00370.1500
Asset 250.00370.00260.1667Asset 650.00280.00180.1667
Asset 260.00150.00040.0167Asset 660.00410.00330.1500
Asset 270.00300.00200.2000Asset 670.00310.00200.1833
Asset 280.00230.00140.0667Asset 680.00390.00260.2000
Asset 290.00200.00090.1167Asset 690.00180.00080.1000
Asset 300.00230.00100.0833Asset 700.00450.00330.2333
Asset 310.00270.00170.2000Asset 710.00340.00220.1667
Asset 320.00330.00210.1667Asset 720.00290.00200.1167
Asset 330.00270.00160.1500Asset 730.00440.00340.1667
Asset 340.00180.00080.1333Asset 740.00630.00490.2333
Asset 350.00290.00180.1667Asset 750.00180.00070.0667
Asset 360.01530.01380.3333Asset 760.00160.00070.0833
Asset 370.00240.00140.1167Asset 770.00230.00110.1333
Asset 380.00210.00100.2167Asset 780.00270.00160.1333
Asset 390.00300.00190.1833Asset 790.00280.00190.1500
Asset 400.00280.00180.1333Asset 800.00390.00260.2167

Appendix B. Pareto Set Metrics and Optimal Weights

In this appendix, we present a deeper analysis concerning the Pareto set outcomes derived by the hybrid MOPSO implementation. The analysis is performed with a seed fixed as s e e d = 123 in the MATLAB implementation over 30 independent runs and with maximum repetitions equals to 500.
In Table A7, we report, for each dataset on 31st December 2018 and with λ = 0.5 , the maximum, minimum, and mean of the objective functions f 1 , f 2 (we recall that f 1 represents the augmented volatility and f 2 the diversification). Moreover, we show the maximum, minimum, and mean of the cardinality in the Pareto front and for two different metrics, i.e., the spacing metric [64] and the spread metric [65]. It is worth noting that the mean cardinality is rounded to the lower integer.
Table A7. Pareto metrics on the three datasets.
Table A7. Pareto metrics on the three datasets.
MetricMinimumMaximumMean
Dow Jones Industrial
cardinality136171153
f 1 value0.017320.022550.01908
f 2 value0.223180.331250.30367
spacing3 · 10 4 0.002067 · 10 4
spread0.801871.011280.97610
Euro Stoxx 50
cardinality149200185
f 1 value0.021630.037090.02564
f 2 value0.228040.318790.29308
spacing2 · 10 4 5 · 10 4 4 · 10 4
spread0.747520.972090.94567
FTSE 100
cardinality88124105
f 1 value0.018140.031080.02075
f 2 value0.338000.399830.38327
spacing2 · 10 4 0.001635 · 10 4
spread0.976431.016080.98890
In general, from Table A7, one can notice a numerical stability behavior for the solver presented.
In the next Table A8, we show how the optimal weights vary in dependence of λ , with λ { 0 , 0.5 , 1 } , on the 31 December 2018. Results are reported for the Dow Jones Industrial. It is worth noting that the optimal allocations are different depending on the sustainable-risk-aversion parameter.
Table A8. Optimal weights for the three different λ -strategies of the Dow Jones Industrial.
Table A8. Optimal weights for the three different λ -strategies of the Dow Jones Industrial.
Asset λ = 0 λ = 0.5 λ = 1 Asset λ = 0 λ = 0.5 λ = 1
Asset 10.00720.02250.0212Asset 150.02490.07330.0230
Asset 20.00520.00170Asset 160.06950.01550.0094
Asset 3000.0009Asset 170.04750.05700.0296
Asset 40.007200.0436Asset 180.04370.09850.1115
Asset 50.14850.14050.0688Asset 190.00110.00010.0013
Asset 600.00850.0291Asset 20000.0003
Asset 70.02500.08070.0905Asset 210.00020.00200.0632
Asset 8000.0010Asset 220.01310.00430
Asset 90.01630.05110.0755Asset 230.04490.03080.0235
Asset 100.10730.09780.3238Asset 240.02850.02810
Asset 110.03700.07470.0633Asset 250.14070.09700
Asset 120.00040.00080.0006Asset 260.008200
Asset 130.16740.03420Asset 270.00090.00380.0178
Asset 140.04650.04870.0001Asset 280.00860.02850.0020

References

  1. Hagin, R.L. Investment Management: Portfolio Diversification, Risk, and Timing–Fact and Fiction; John Wiley & Sons: Hoboken, NJ, USA, 2004; Volume 235. [Google Scholar]
  2. Markowitz, H. Modern portfolio theory. J. Financ. 1952, 7, 77–91. [Google Scholar]
  3. Koumou, G.B. Diversification and portfolio theory: A review. Financ. Mark. Portf. Manag. 2020, 34, 267–312. [Google Scholar] [CrossRef]
  4. Meucci, A. Managing diversification. Risk 2009, 22, 74–79. [Google Scholar]
  5. Pola, G. On entropy and portfolio diversification. J. Asset Manag. 2016, 17, 218–228. [Google Scholar] [CrossRef]
  6. Ararat, Ç.; Cesarone, F.; Pınar, M.Ç.; Ricci, J.M. MAD risk parity portfolios. Ann. Oper. Res. 2024, 336, 899–924. [Google Scholar] [CrossRef]
  7. Hoepner, A.G. Corporate Social Responsibility and Investment Portfolio Diversification. 2010. Available online: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=1599334 (accessed on 11 August 2025).
  8. Sbaiz, G. Sustainable information into portfolio optimization models. J. Appliedmath 2023, 1, 125. [Google Scholar] [CrossRef]
  9. Brauers, W.K. Optimization Methods for a Stakeholder Society: A Revolution in Economic Thinking by Multi-Objective Optimization; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2003; Volume 73. [Google Scholar]
  10. Coello, C.A.C. Evolutionary multi-objective optimization in finance. In Handbook of Research on Nature-Inspired Computing for Economics and Management; IGI Global: Hershey, PA, USA, 2007; pp. 74–89. [Google Scholar]
  11. Doumpos, M.; Zopounidis, C. Multi-objective optimization models in finance and investments. J. Glob. Optim. 2020, 76, 243–244. [Google Scholar] [CrossRef]
  12. Fabretti, A.; Herzel, S. Delegated portfolio management with socially responsible investment constraints. Eur. J. Financ. 2012, 18, 293–309. [Google Scholar] [CrossRef]
  13. Utz, S.; Wimmer, M.; Hirschberger, M.; Steuer, R.E. Tri-criterion inverse portfolio optimization with application to socially responsible mutual funds. Eur. J. Oper. Res. 2014, 234, 491–498. [Google Scholar] [CrossRef]
  14. Coqueret, G. Diversified minimum-variance portfolios. Ann. Financ. 2015, 11, 221–241. [Google Scholar] [CrossRef]
  15. Gasser, S.M.; Rammerstorfer, M.; Weinmayer, K. Markowitz revisited: Social portfolio engineering. Eur. J. Oper. Res. 2017, 258, 1181–1190. [Google Scholar] [CrossRef]
  16. Pedersen, L.H.; Fitzgibbons, S.; Pomorski, L. Responsible investing: The ESG-efficient frontier. J. Financ. Econ. 2021, 142, 572–597. [Google Scholar] [CrossRef]
  17. Steuer, R.E.; Utz, S. Non-contour efficient fronts for identifying most preferred portfolios in sustainability investing. Eur. J. Oper. Res. 2023, 306, 742–753. [Google Scholar] [CrossRef]
  18. Ling, A.; Li, J.; Wen, L.; Zhang, Y. When trackers are aware of ESG: Do ESG ratings matter to tracking error portfolio performance? Econ. Model. 2023, 125, 106346. [Google Scholar] [CrossRef]
  19. Carmichael, B.; Koumou, G.B.; Moran, K. Unifying portfolio diversification measures using Rao’s quadratic entropy. J. Quant. Econ. 2023, 21, 769–802. [Google Scholar] [CrossRef]
  20. Aprea, I.L.; Sbaiz, G. A neural network-particle swarm solver for sustainable portfolio optimization problems. In Decisions in Economics and Finance; Springer: Berlin/Heidelberg, Germany, 2025; pp. 1–40. [Google Scholar]
  21. Zeidan, R.; Onabolu, S. The generalized sustainability credit rating system. Braz. Rev. Financ. 2023, 21, 21–47. [Google Scholar] [CrossRef]
  22. Poveda, C.A.; Lipsett, M.G. A review of sustainability assessment and sustainability/environmental rating systems and credit weighting tools. J. Sustain. Dev. 2011, 4, 36. [Google Scholar] [CrossRef]
  23. Zeidan, R.; Boechat, C.; Fleury, A. Developing a sustainability credit score system. J. Bus. Ethics 2015, 127, 283–296. [Google Scholar] [CrossRef]
  24. Gutiérrez-Nieto, B.; Serrano-Cinca, C.; Camón-Cala, J. A credit score system for socially responsible lending. J. Bus. Ethics 2016, 133, 691–701. [Google Scholar] [CrossRef]
  25. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the ICNN’95-International Conference on Neural Networks, Perth, WA, Australia, 27 November–1 December 1995; Volume 4, pp. 1942–1948. [Google Scholar]
  26. Mostaghim, S.; Teich, J. Strategies for finding good local guides in multi-objective particle swarm optimization (MOPSO). In Proceedings of the 2003 IEEE Swarm Intelligence Symposium. SIS’03 (Cat. No. 03EX706), Indianapolis, IN, USA, 26–26 April 2003; pp. 26–33. [Google Scholar]
  27. Zheng, X.; Liu, H. A hybrid vertical mutation and self-adaptation based MOPSO. Comput. Math. Appl. 2009, 57, 2030–2038. [Google Scholar] [CrossRef]
  28. Lalwani, S.; Singhal, S.; Kumar, R.; Gupta, N. A comprehensive survey: Applications of multi-objective particle swarm optimization (MOPSO) algorithm. Trans. Comb. 2013, 2, 39–101. [Google Scholar]
  29. Sun, Y.; Gao, Y. A multi-objective particle swarm optimization algorithm based on gaussian mutation and an improved learning strategy. Mathematics 2019, 7, 148. [Google Scholar] [CrossRef]
  30. Chen, J.; Liu, Y.; Luo, Y.; Ouyang, A.; Yang, J.; Bai, W. Multi-objective particle swarm optimization algorithm for task allocation and archived guided mutation strategies. Sci. Rep. 2025, 15, 15821. [Google Scholar] [CrossRef] [PubMed]
  31. Moore, J. Application of Particle Swarm to Multiobjective Optimization. Technical Report. 1999. Available online: https://cir.nii.ac.jp/crid/1571980076089705344 (accessed on 11 August 2025).
  32. Coello, C.A.C.; Pulido, G.T.; Lechuga, M.S. Handling multiple objectives with particle swarm optimization. IEEE Trans. Evol. Comput. 2004, 8, 256–279. [Google Scholar] [CrossRef]
  33. Sierra, M.R.; Coello Coello, C.A. Improving PSO-based multi-objective optimization using crowding, mutation and ϵ-dominance. In Proceedings of the International Conference on Evolutionary Multi-Criterion Optimization, Guanajuato, Mexico, 9–11 March 2005; Springer: Berlin/Heidelberg, Germany, 2005; pp. 505–519. [Google Scholar]
  34. Durillo, J.J.; García-Nieto, J.; Nebro, A.J.; Coello, C.A.C.; Luna, F.; Alba, E. Multi-objective particle swarm optimizers: An experimental comparison. In Proceedings of the Evolutionary Multi-Criterion Optimization: 5th International Conference, EMO 2009, Nantes, France, 7–10 April 2009; Springer: Berlin/Heidelberg, Germany, 2009; pp. 495–509. [Google Scholar]
  35. Reyes-Sierra, M.; Coello, C.C. Multi-objective particle swarm optimizers: A survey of the state-of-the-art. Int. J. Comput. Intell. Res. 2006, 2, 287–308. [Google Scholar] [CrossRef]
  36. Chalvatzis, C.; Hristu-Varsakelis, D. High-performance stock index trading via neural networks and trees. Appl. Soft Comput. 2020, 96, 106567. [Google Scholar] [CrossRef]
  37. Heaton, J.B.; Polson, N.G.; Witte, J.H. Deep learning for finance: Deep portfolios. Appl. Stoch. Model. Bus. Ind. 2017, 33, 3–12. [Google Scholar] [CrossRef]
  38. Garcia-Garcia, A.; Orts-Escolano, S.; Oprea, S.; Villena-Martinez, V.; Martinez-Gonzalez, P.; Garcia-Rodriguez, J. A survey on deep learning techniques for image and video semantic segmentation. Appl. Soft Comput. 2018, 70, 41–65. [Google Scholar] [CrossRef]
  39. Nassif, A.B.; Elnagar, A.; Shahin, I.; Henno, S. Deep learning for Arabic subjective sentiment analysis: Challenges and research opportunities. Appl. Soft Comput. 2021, 98, 106836. [Google Scholar] [CrossRef]
  40. Rosenblatt, F. The perceptron: A probabilistic model for information storage and organization in the brain. Psychol. Rev. 1958, 65, 386. [Google Scholar] [CrossRef]
  41. Elman, J.L. Finding structure in time. Cogn. Sci. 1990, 14, 179–211. [Google Scholar] [CrossRef]
  42. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  43. Md, A.Q.; Kapoor, S.; AV, C.J.; Sivaraman, A.K.; Tee, K.F.; Sabireen, H.; Janakiraman, N. Novel optimization approach for stock price forecasting using multi-layered sequential LSTM. Appl. Soft Comput. 2022, 134, 109830. [Google Scholar] [CrossRef]
  44. Cho, K.; Van Merriënboer, B.; Bahdanau, D.; Bengio, Y. On the properties of neural machine translation: Encoder-decoder approaches. arXiv 2014, arXiv:1409.1259. [Google Scholar] [CrossRef]
  45. Aslan, A.; Posch, P.N. How do investors value sustainability? A utility-based preference optimization. Sustainability 2022, 14, 15963. [Google Scholar] [CrossRef]
  46. Dorfleitner, G.; Utz, S. Safety first portfolio choice based on financial and sustainability returns. Eur. J. Oper. Res. 2012, 221, 155–164. [Google Scholar] [CrossRef]
  47. Borra, S.; Hill, A.D.C.M. Statistica-Metodologie per le Scienze Economiche e Sociali/2e; McGraw-Hill: New York, NY, USA, 2014. [Google Scholar]
  48. Rockafellar, R.T.; Uryasev, S. Optimization of conditional value-at-risk. J. Risk 2000, 2, 21–42. [Google Scholar] [CrossRef]
  49. Leitner, J. A short note on second-order stochastic dominance preserving coherent risk measures. Math. Financ. Int. J. Math. Stat. Financ. Econ. 2005, 15, 649–651. [Google Scholar] [CrossRef]
  50. Cesarone, F.; Martino, M.L.; Carleo, A. Does ESG impact really enhance portfolio profitability? Sustainability 2022, 14, 2050. [Google Scholar] [CrossRef]
  51. Politis, D.N.; Romano, J.P. The stationary bootstrap. J. Am. Stat. Assoc. 1994, 89, 1303–1313. [Google Scholar] [CrossRef]
  52. Politis, D.N.; White, H. Automatic block-length selection for the dependent bootstrap. Econom. Rev. 2004, 23, 53–70. [Google Scholar] [CrossRef]
  53. Ledoit, O.; Wolf, M. Honey, I shrunk the sample covariance matrix. J. Portf. Manag. 2004, 30, 110–119. [Google Scholar] [CrossRef]
  54. Beck, A. First-Order Methods in Optimization; SIAM: Philadelphia, PA, USA, 2017. [Google Scholar]
  55. Meghwani, S.S.; Thakur, M. Multi–criteria algorithms for portfolio optimization under practical constraints. Swarm Evol. Comput. 2017, 37, 104–125. [Google Scholar] [CrossRef]
  56. Aprea, I.L.; Scognamiglio, S.; Zanetti, P. Systemic risk measurement: A quantile long short-term memory network approach. Appl. Soft Comput. 2024, 152, 111224. [Google Scholar] [CrossRef]
  57. Tobias, A.; Brunnermeier, M.K. CoVaR. Am. Econ. Rev. 2016, 106, 1705. [Google Scholar] [CrossRef]
  58. Goodfellow, I.; Bengio, Y.; Courville, A.; Bengio, Y. Deep Learning; MIT Press Cambridge: Cambridge, MA, USA, 2016; Volume 1. [Google Scholar]
  59. Beraldi, P.; Violi, A.; Ferrara, M.; Ciancio, C.; Pansera, B.A. Dealing with complex transaction costs in portfolio management. Ann. Oper. Res. 2021, 299, 7–22. [Google Scholar] [CrossRef]
  60. Shen, W.; Wang, J.; Ma, S. Doubly regularized portfolio with risk minimization. In Proceedings of the AAAI Conference on Artificial Intelligence, Québec City, QC, Canada, 27–31 July 2014; Volume 28. [Google Scholar]
  61. Chekhlov, A.; Uryasev, S.; Zabarankin, M. Drawdown measure in portfolio optimization. Int. J. Theor. Appl. Financ. 2005, 8, 13–58. [Google Scholar] [CrossRef]
  62. Woerheide, W.; Persson, D. An index of portfolio diversification. Financ. Serv. Rev. 1992, 2, 73–85. [Google Scholar] [CrossRef]
  63. DeMiguel, V.; Garlappi, L.; Uppal, R. Optimal versus naive diversification: How inefficient is the 1/N portfolio strategy? Rev. Financ. Stud. 2009, 22, 1915–1953. [Google Scholar] [CrossRef]
  64. Schott, J.R. Fault Tolerant Design Using Single and Multicriteria Genetic Algorithm Optimization. Master’s Thesis, Massachusetts Institute of Technology, Cambridge, MA, USA, 1995. [Google Scholar]
  65. Wang, Y.N.; Wu, L.H.; Yuan, X.F. Multi-objective self-adaptive differential evolution with elitist archive and crowding entropy-based diversity measure. Soft Comput. 2010, 14, 193–209. [Google Scholar] [CrossRef]
Figure 1. Heatmaps of correlation matrices of daily returns (left) and correlation matrices of ESG scores (right) for the constituents of the Dow Jones Industrials, Euro Stoxx 50, and FTSE 100 over 2006–2020.
Figure 1. Heatmaps of correlation matrices of daily returns (left) and correlation matrices of ESG scores (right) for the constituents of the Dow Jones Industrials, Euro Stoxx 50, and FTSE 100 over 2006–2020.
Mathematics 13 03308 g001
Figure 2. A graphical representation of the MOPSO solver.
Figure 2. A graphical representation of the MOPSO solver.
Mathematics 13 03308 g002
Figure 3. Architecture of the LSTM unit.
Figure 3. Architecture of the LSTM unit.
Mathematics 13 03308 g003
Figure 4. Diagram of the q-LSTM network for the CVaR estimation.
Figure 4. Diagram of the q-LSTM network for the CVaR estimation.
Mathematics 13 03308 g004
Figure 5. Efficient frontiers in the augmented volatility-diversification plane, for the three datasets on the last day of December for each year in the out-of-sample window—from 2016 (first row) to 2020 (last row). The x-axis reports the augmented volatility σ , i.e., f 1 ( w ) , while the y-axis represents the diversification index value, i.e., f 2 ( w ) . The colors indicate different levels of the sustainable-risk-aversion parameter.
Figure 5. Efficient frontiers in the augmented volatility-diversification plane, for the three datasets on the last day of December for each year in the out-of-sample window—from 2016 (first row) to 2020 (last row). The x-axis reports the augmented volatility σ , i.e., f 1 ( w ) , while the y-axis represents the diversification index value, i.e., f 2 ( w ) . The colors indicate different levels of the sustainable-risk-aversion parameter.
Mathematics 13 03308 g005
Figure 6. Evolution of the net wealth for the three different λ levels and the Equally Weighted (EW) allocation, using the assets in the three datasets.
Figure 6. Evolution of the net wealth for the three different λ levels and the Equally Weighted (EW) allocation, using the assets in the three datasets.
Mathematics 13 03308 g006aMathematics 13 03308 g006b
Table 1. The three datasets with the corresponding number of observations, number of market constituents (N), and the corresponding currency.
Table 1. The three datasets with the corresponding number of observations, number of market constituents (N), and the corresponding currency.
Dataset NameDaily ObservationsAssets (N)Currency
Dow Jones Industrial371528$
Euro Stoxx 50371545
FTSE 100371580£
Table 2. Parameter setting for the MOPSO solver and the LSTM architecture.
Table 2. Parameter setting for the MOPSO solver and the LSTM architecture.
ParameterValue
MOPSO solver
N p o p 600
m a x _ i t e r 500
A m a x 200
ω 0.4
c 1 2
c 2 2
m u t % 50%
LSTM architecture
α 0.05
u α 50
dropoutα 10 %
batch-sizeα8
epochsα500
Activation Functions
VaR α ReLU ×(−1)
CVaR α tanh
Table 3. Mean time of execution of each monthly allocation problem for each dataset.
Table 3. Mean time of execution of each monthly allocation problem for each dataset.
Dataset NameComputational Time (s)
Dow Jones Industrial26.2567
Euro Stoxx 5028.4240
FTSE 10035.9011
Table 4. Structure of transaction costs as in [59].
Table 4. Structure of transaction costs as in [59].
Trading Segment t s (Local Currency)Fixed Fee (Local Currency)Proportional Cost (%)
0 t s 8000 400
8000 < t s 50 , 000 00.5
50 , 000 < t s 100 , 000 00.4
100 , 000 < t s 200 , 000 00.25
t s > 200 , 000 4000
Table 5. Ex-post performance measures of the proposed portfolio allocation model for different allocation strategies on the three datasets. The bold formatting indicates the significant values in the table to facilitate the reading of the results.
Table 5. Ex-post performance measures of the proposed portfolio allocation model for different allocation strategies on the three datasets. The bold formatting indicates the significant values in the table to facilitate the reading of the results.
λ Value λ = 0 λ = 0.5 λ = 1 EW
Dow Jones Industrial
CAGR0.31510.28830.29660.1294
SR0.59540.61270.66150.2716
μ e x p o s t 0.02450.02260.02300.0113
σ e x p o s t 0.04130.03700.03490.0418
mean DD−0.0121−0.0109−0.0100−0.0247
std DD0.02710.02430.02340.0405
max DD−0.1359−0.1114−0.1097−0.2039
TR0.65840.66480.6378
IQRTR0.37720.49640.4810
c r e b 8618.78393.57965.6
DI0.79510.79830.8022
nnz assets (avg)22212128
nnz assets (min)12111328
Euro Stoxx 50
CAGR0.37110.33000.27260.0671
SR0.62010.69290.62070.1343
μ e x p o s t 0.02830.02540.02150.0071
σ e x p o s t 0.04570.03670.03470.0533
mean DD−0.0104−0.0075−0.0070−0.0514
std DD0.02830.02120.01680.0627
max DD−0.1753−0.1336−0.1009−0.2744
TR0.75220.81650.6536
IQRTR0.41750.64600.4758
c r e b 103439937.48174.1
DI0.85520.85340.8461
nnz assets (avg)30302745
nnz assets (min)13121045
FTSE 100
CAGR0.31010.24840.24470.0649
SR0.58160.58390.58440.1517
μ e x p o s t 0.02460.02020.01990.0066
σ e x p o s t 0.04230.03470.03410.0439
mean DD−0.0151−0.0113−0.0098−0.0464
std DD0.03220.02360.02060.0678
max DD−0.1636−0.1276−0.1139−0.2634
TR0.88930.84010.8099
IQRTR0.37520.33490.4565
c r e b 155021502213627
DI0.93800.95180.9490
nnz assets (avg)56545180
nnz assets (min)19201880
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Aprea, I.L.; Bosi, G.; Sbaiz, G.; Scognamiglio, S. Bi-Objective Portfolio Optimization Under ESG Volatility via a MOPSO-Deep Learning Algorithm. Mathematics 2025, 13, 3308. https://doi.org/10.3390/math13203308

AMA Style

Aprea IL, Bosi G, Sbaiz G, Scognamiglio S. Bi-Objective Portfolio Optimization Under ESG Volatility via a MOPSO-Deep Learning Algorithm. Mathematics. 2025; 13(20):3308. https://doi.org/10.3390/math13203308

Chicago/Turabian Style

Aprea, Imma Lory, Gianni Bosi, Gabriele Sbaiz, and Salvatore Scognamiglio. 2025. "Bi-Objective Portfolio Optimization Under ESG Volatility via a MOPSO-Deep Learning Algorithm" Mathematics 13, no. 20: 3308. https://doi.org/10.3390/math13203308

APA Style

Aprea, I. L., Bosi, G., Sbaiz, G., & Scognamiglio, S. (2025). Bi-Objective Portfolio Optimization Under ESG Volatility via a MOPSO-Deep Learning Algorithm. Mathematics, 13(20), 3308. https://doi.org/10.3390/math13203308

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop