2. Background and Theory
This section reviews previous research on cryptocurrency markets and outlines the historical development of the Fractal Market Hypothesis (FMH). It also introduces key metrics used to characterise the underlying nature of the market under investigation. Finally, methods for assessing the direction and stability of time-series trends are derived, based on the assumption of a self-affine field following a Lévy distribution.
2.1. Previous Research
There is now a substantial body of academic literature examining the price dynamics and return characteristics of cryptocurrencies. Much of this research has focused on determining whether high-market-cap cryptocurrencies, such as Bitcoin (BTC) and Ethereum (ETH), exhibit long-term dependence in their price signals or their derivatives. Most studies report evidence of cryptocurrency market inefficiency, suggesting the existence of profitable trading opportunities. However, a parallel line of research indicates that BTC is gradually becoming more efficient as the market evolves.
Urquhart (2016) analysed daily BTC price data from 2010 to 2016 to test for informational efficiency [
22]. Using the Ljung–Box test, Runs test, Hurst exponent, and Bartels test, he assessed long-term memory and concluded that the market was inefficient, with the Hurst exponent showing strong anti-persistence. Nevertheless, by around 2013, BTC began to show signs of maturity.
Building on this work, Nadarajah and Chu (2017) [
23] examined similar datasets from 2010 to 2016, split into two three-year samples, and applied additional methods such as the generalised spectral and Portmanteau tests. Their findings supported Urquhart’s conclusion of return independence. Bariviera (2017) [
24] studied long-term dependence in BTC returns and volatility using daily data from 2011 to 2017 and De-trended Fluctuation Analysis (DFA) with sliding windows to estimate the Hurst exponent. Results indicated that BTC returns showed persistence that declined toward efficiency after 2014, although volatility remained persistently long-range up to 2017. Further work by Bariviera et al. (2017) [
25] demonstrated that alternative time scales did not affect long-term dependence. Lahmiri et al. (2018) [
26] employed Fractionally Integrated GARCH models and rejected the Efficient Market Hypothesis (EMH), finding consistent evidence of market memory under various distributional assumptions.
Following the dramatic price surges of 2017, Al-Yahyaee et al. (2018) [
27] used daily data from 2010 to 2017 to compare BTC efficiency with that of other asset classes, including foreign exchange markets. This was the first application of Multifractal Detrended Fluctuation Analysis (MF-DFA), developed by Kantelhardt in 2002, to the BTC market. Their results showed that BTC exhibited the strongest long-term persistence and the lowest efficiency among the markets examined. Alvarez-Ramirez et al. (2018) [
28], also using DFA, reported alternating periods of efficiency and inefficiency, with the latter associated with anti-persistence. In the same year, Jiang et al. (2018) [
29] tested for long-term dependence using the Hurst exponent on a 14-day rolling basis, along with the Ljung–Box and Automatic Variance Ratio (AVR) tests, again finding strong evidence of inefficiency and long memory. Zhang et al. (2018) [
30] analysed daily data from 2013 to 2018 for a basket of cryptocurrencies using kurtosis, skewness, autocorrelation, and DFA, showing that BTC was gradually approaching market efficiency (Hurst exponent = 0.5), although volatility continued to display clear long-term dependence. Later in 2018, Caporale [
31] applied Harold Hurst’s Re-scaled Range (R/S) analysis to the largest cryptocurrencies by market capitalisation, finding that BTC’s Hurst exponent was trending toward efficiency but remained unstable, implying that trend-following strategies could still generate excess profits.
Research in 2019 further strengthened these conclusions. Celeste et al. [
32] used R/S analysis and continuous wavelet transforms to examine fractal dynamics in BTC and ETH prices, reporting strong market memory and increasing memory effects in ETH. Hu et al. (2019) [
33] rejected the EMH using a panel framework that revealed cross-sectional dependence among high-cap cryptocurrencies. Chu (2019) [
34] tested the Adaptive Market Hypothesis for BTC and ETH with hourly data from July 2017 to September 2018, employing the Ljung–Box and Kolmogorov–Smirnov tests. His results showed time-varying efficiency in both BTC and ETH and suggested that market sentiment and news events were not necessarily decisive factors in determining efficiency.
More recent studies have confirmed and extended these findings. Al-Yahyaee et al. (2020) [
35] analysed multifractal characteristics of several cryptocurrencies using a rolling MF-DFA and quantile regression approach. They found long-term memory effects and time-varying inefficiency, concluding that high liquidity and low volatility enhance efficiency and create arbitrage opportunities for active traders. During the COVID-19 pandemic, Kakinaka and Umeno (2021) [
36] examined daily data from January 2019 to December 2020 using MF-DFA, observing an increase in short-term multifractality but a decline in long-term multifractal behaviour. David et al. (2021) [
37] applied Auto-Regressive Integrated Moving Average (ARIMA), Auto-Regressive Fractionally Integrated Average models, and DFA to daily BTC data from July 2016 to March 2019. Their analysis of the Hurst exponent, fractal dimension, and Lyapunov exponents revealed persistent long-term dependence and chaotic dynamics.
Complementary work has advanced the use of high-frequency data. Drożdż et al. (2018) [
38] applied advanced MF-DFA methods to capture subtle variations in multifractality that were not detectable in earlier studies. Watorek et al. (2021) [
39] confirmed these findings using surrogate datasets to verify that observed multifractal structures were genuine and not artefacts of noise or methodology. This combination of high-frequency analysis and rigorous validation strengthens the evidence for complex scaling dynamics in cryptocurrency markets.
Taken together, this extensive body of research strongly challenges the EMH and supports the view that BTC, while becoming more efficient over time, continues to exhibit long-term memory effects. Such memory implies that future prices depend partly on past values, allowing investors to predict returns and reduce market risk. Because the Fractal Market Hypothesis (FMH) explicitly incorporates memory within a stochastic price field, it provides a natural and powerful framework for modelling cryptocurrency markets in future research.
2.2. Trading with Cryptocurrencies
Cryptocurrencies can be traded on a variety of platforms, each offering distinct advantages. Examples such as Coinbase and Binance are dedicated exclusively to cryptocurrency markets, whereas platforms like eToro also facilitate trading in traditional assets such as stocks and commodities. Each platform applies its own fee structure for executing trades. For instance, Coinbase charges a transaction fee of approximately
[
40], while eToro charges around
[
41]. However, these platform fees are not the only costs associated with cryptocurrency trading. An intrinsic cost, known as a ‘Gas Fee’, is required to add a transaction to the blockchain, as discussed in the introduction. Gas fees depend heavily on current blockchain traffic and the complexity of the transaction, and they can fluctuate widely, making trades significantly less cost-effective and potentially disrupting a trading strategy.
For example, while eToro charges a higher base transaction fee than Coinbase, it does not require payment of a gas fee. Given that the average gas fee is approximately $20, with peaks that can reach several thousand dollars during periods of network congestion, eToro can offer a more stable and predictable cost structure for traders.
At present (i.e. at the time of undertaken the research discussed in this article) the Ethereum blockchain is transitioning to Ethereum 2.0, an upgrade designed to introduce a more energy-efficient mechanism for validating transactions, thereby reducing gas fees. If this transition is completed as projected by the end of 2022, it could significantly influence platform preferences among traders. Currently, eToro remains the only widely used and internationally trusted platform that supports crypto-based options and derivatives.
For the purposes of this study, the eToro model is adopted, under which no gas fee is applied. Accordingly, all Return on Investment (ROI) calculations are presented on a pre-fee basis and therefore represent gross profit.
2.3. The Fractal Market Hypothesis
Research dating back to the 1960s suggested that commodity price-change distributions were too sharply peaked to be Gaussian, a phenomenon later termed leptokurtosis. This characteristic was also observed by Mandelbrot [
42,
43]. These and other studies indicate that the assumption of normally distributed price changes is inadequate for capturing the fundamental statistics of Financial Time Series (FTS), motivating the development of the Fractal Market Hypothesis (FMH).
The FMH was formally introduced by Edgar Peters in his 1994 book
Fractal Market Analysis: Applying Chaos Theory to Investment and Economics [
44], building on earlier work by Mandelbrot and Ralph Elliott. It draws heavily on fractal geometry, a field first proposed by Felix Hausdorff in 1918 [
45] and popularised by Mandelbrot in his 1982 book
The Fractal Geometry of Nature [
46]. Fractal geometry studies how fractured objects exhibit self-similarity, a property in which geometric features are preserved at all scales [
47]. This property is closely related to chaotic signals, which can be generated through the iteration of strictly nonlinear functions. Modelling FTS as chaotic signals has found practical applications in various trading scenarios [
48].
FTS are now recognised as examples of stochastic self-affine fields. The self-affine properties of financial series were first recorded by Ralph Elliott in 1938 [
49]. Unlike conventional calculus, fractional calculus introduces the concept of memory into derivatives. Within this framework (i.e., modelling FMH signals as fractional differential equations), the FMH provides a model for financial time series that incorporates memory, meaning that the price of a given currency or commodity depends on past values. This represents a major advancement over the EMH, which assumes that markets are efficient and therefore unpredictable. Moreover, because FTS appear statistically similar at different scales, one can assume that the probability density functions (PDFs) of price values are approximately scale-invariant.
Many self-affine functions exist. Early examples include the ‘Lévy Curve’, introduced by Paul Lévy in his 1938 paper
Plane or Space Curves and Surfaces Consisting of Parts Similar to the Whole [
50]. Early FMH developments also relied on the Lévy distribution. Mandelbrot observed that cotton price changes followed a Lévy distribution [
51] with a Lévy index (
) of 1.7 [
43]. The Lévy index quantifies the deviation of a distribution from Gaussian behaviour. Within the FMH,
, where lower values indicate heavier tails and a higher peak, higher values indicate the opposite, and
corresponds to a Gaussian distribution. As
decreases further from 2, the probability of rare but extreme events—so-called ‘Lévy flights’—increases.
Other metrics capture a signal’s fractal characteristics. Mandelbrot, applying principles of Brownian motion to asset prices, proposed that markets exhibiting long-term dependence tend to maintain trends. He also developed a metric to quantify a field’s persistence (trend continuation) or anti-persistence, known as the Hurst exponent (
H). This metric is named after Harold Edwin Hurst, who, in 1906, discovered that the range of water levels in the Nile River scaled according to a fractional power law [
52]. The Hurst exponent measures long-term memory:
indicates anti-persistent behaviour (negative correlation), whereas
indicates persistent behaviour (positive correlation). In the context of FTS, a high
H implies that prices are biased in a particular trend direction over time, meaning that in a Random Walk Model (RWM), price deviations from the origin are larger.
The parameters
and
H are related to the fractal dimension (
), which quantifies a field’s complexity and self-affinity [
53]. These can also be estimated using the Spectral Decay Coefficient (
), as discussed later in this report.
Because fractals are inherently iterative and related to chaos theory, there is a natural connection between chaos and the FMH. Another important parameter is the Lyapunov exponent (), which characterises the level of chaotic behaviour in a system, i.e., the transition from regular to chaotic dynamics. Within this framework, can be used to analyse the evolution of a financial time series and support trend prediction.
Taken together, these metrics—H, , , and —provide a robust framework for analysing irregular financial time series within the context of the FMH.
2.4. The FMH vs. EMH
The Fractal Market Hypothesis and the Efficient Market Hypothesis offer contrasting perspectives on how financial markets function. The EMH posits that asset prices fully reflect all available information, rendering it impossible for investors to consistently achieve abnormal returns through prediction or arbitrage. It assumes rational agents, normally distributed returns, and market equilibrium driven by information efficiency. In this framework, price dynamics are largely random and memoryless, conforming to a Gaussian model.
By contrast, the FMH, draws on fractal geometry and chaos theory to argue that financial markets exhibit long-range dependence, heterogeneity of investor horizons, and self-similar structures across time scales. Rather than assuming equilibrium, FMH emphasises persistence and clustering in volatility, as well as the influence of structural changes and collective behaviour. It suggests that markets remain stable when investors with diverse time horizons interact, but instability arises when this diversity diminishes, leading to crises or bubbles.
In the context of cryptocurrency markets, FMH offers a more realistic framework than EMH. Cryptocurrencies are highly volatile, exhibit frequent periods of speculative bubbles, and are shaped by heterogeneous participants ranging from algorithmic traders to retail investors with divergent horizons. Moreover, empirical studies demonstrate heavy tails, autocorrelation in volatility, and fractal-like scaling in crypto price series, all of which contradict Gaussian assumptions underpinning EMH. FMH, by accommodating these nonlinear and long-memory features, provides a stronger theoretical basis for analysing cryptocurrency markets and assessing risks in environments characterised by structural instability and speculative dynamics.
2.4.1. Characteristic Indices
The observation that price changes in an FTS are inefficient [
9] allows the introduction of a number of further indices to analyse the variability in stochastic field distributions. These act as a metric of the data’s deviation from the normal distribution and characterise the fields susceptibility to undergo Lévy flights and exhibit long-term market memory. Moreover, these metrics can all be defined through linear relationships of the spectral decay coefficient,
, for a self-affine stochastic fields’ power spectrum. Therefore, definitions of these indices can be reduced to the calculation of the
.
The power spectrum
of a self-affine signal decays according to a power law [
54].
where
is the spectral frequency (specifically the temporal angular frequency) and
c is a real constant. Using logarithms, the above equation can be re-written as
from which the value of
can be obtained by determining the gradient of a linear regression applied to the power spectral log-log plot. In computational terms, application of the least squares regression formula provides a value of
when applied to the entire field. The Matlab code for calculating
is shown in
Appendix A.8. Using this result, the following standard algebraic formulae for the following measures can be obtained:
where
H is the Hurst Exponent [
55].
where
is the Fractal Dimension of the signal [
46].
where
is the Levy Index [
56]. These measures and their meanings with respect to stochastic field distributions and the relevant cryptocurrencies are discussed in
Section 2.3.
An FTS commonly consists of a set of discrete values, representing the price of a currency or commodity. There is value in developing deterministic models to analyse economic systems, providing such a quantitative analysis can aid the management of risk. However, large markets are usually functions of random variables characterised by external influences which are complex and therefore difficult to define. These influences on the system are compounded due to their non-linear nature caused by feedback, a market reacting to itself, or its sensitivity to shocks under ‘market memory’ conditions.
For these reasons, the models are often relatively complicated. By dispensing with the efficient market view of economic systems in favour of a fractal base, an index to predict the future price trend behaviour can be developed as outlined in the following discussion. The goal is to develop a single measure capable of predicting the start of an upward (bull) or downward (bear) trend in a cryptocurrency price.
A standard model for a signal
, in this case a continuous, time-dependent currency price, is embodied in the equation
where ⊗ denotes the convolution integral,
In Equation (
4),
is the Impulse Response Function,
is the Source Information Function and
is the stochastic function of time—residual and additive noise [
57]. An inherent problem of signal processing is the extraction of the relevant information from the ‘noisy’ signal when only an estimate of
is possible. This remains the case with financial signal processing. However, in this case,
consists of a wide density of global transactions that occur throughout time. The wide and effectively random nature of these transactions produce such a broad spectrum that they can be interpreted as ‘White Noise’, even if this ‘Noise’ represents real transactions. Under this assumption, Equation (
4) can be modified to
where
now represents the white noise of the ‘system’.
Consider a financial price signal where the system noise is in fact a valid price. Then the Signal-to-Noise Ratio can be considered to be high enough that residual noise can be ignored. Due to the human element of the transaction system, the convolution is where the feedback, or transaction history, is introduced into the model. This is a linear stationary model. However, financial signals are inherently non-stationary. Thus, a metric relating to a stationary model must be applied on a moving window basis to create an ‘index signal’ which represents the time evolution of the metric. This signal can then be used to indicate the instantaneous nature of the signal but not used to ‘predict’ the future, which is a common misconception in financial signal processing.
Under the assumptions of the fractal market displaying a Lévy distribution, the standard Fourier space signal signature,
, can be defined as (using convolution theorem)
where
is the noise present in the system and
is the transfer function of the signal given by
The temporal form of the transfer function is obtained by taking the Inverse Fourier Transform (IFT) giving [
58]
where
is the Gamma function and
—a metric that quantifies the self-affine characteristics of the signal—and
is the Heaviside step function given by
The signal is thus given by (using the convolution theorem)
The metric is related to the Fractal Dimension (), Hurst Exponent (H) and Levy Index (), and hence, by association, measures the signals persistence or anti-persistence, i.e., whether the signal is convergent (bear market trend) or divergent (bull market trend).
The self-affine characteristics of the signal
for an arbitrary scaling factor,
, can be analysed as follows. Suppose we let
and compute
for the same convolutional transform but for function
instead of
. Given Equation (
7), this yields
and thus, we can write,
However, because
is being taken to be a stochastic function, there will be differences between the time signatures observed over different time scales but, on the basis of Equation (
8), the distribution of amplitudes will remain the same over all scales. In other words, we can consider
to be statistically self-similar, a property that is compounded in the equation
This is the defining characteristic of a self-affine stochastic function and therefore shows that Equation (
7) is a valid representation of a fractal signal—a self-affine stochastic field.
This equation provides a method of calculating the metric (by applying regression method to the power spectrum for as discussed earlier) over an entire data set. However, this is not useful for estimating the instantaneous self-affine characteristic of the signal over time.
Suppose we consider a small ‘look-back’ window of data, where , is used to compute the index on a rolling basis. In this case, the resulting time varying metric will provide an indicator of the base field’s persistence such as a currency exchange rate.
Applying this approach to Equation (
7), for the case when
, the convolution integral can be approximated as
where
In practice, given Equation (
9),
T is the windowed data length, which is a fraction of the full data complement. Defining
, creates a single calculable measure when
and the signal
is convergent, indicating an impending decay in value of the underlying currency. When
there is an indication of a rise in price.
Applying logarithms to Equation (
9), we obtained the linear equation
and thus, once again, using the least squares regression algorithm, the log-log line gradient is given by the value of
. Implemented in code, this is a faster computational procedure than performing rolling convolutions, especially when the FTS consists of many data points. The Matlab function for computing
is provided in
Appendix A.3.
2.4.2. The Lyapunov Exponent
To develop another metric capable of indicating the time evolution of a stochastic field, other properties of an FTS can be utilised. For example, suppose we model an FTS as
where
is the amplitude of the signal (the price value) for the
iteration and
is a small
-order error. This is the basis for the RWM whether it be consistent for classical or fractional Brownian diffusion. However, self-affine stochastic fields do not exhibit Gaussian properties, being more susceptible to rare and extreme events and displaying longer distribution tails, which is indicative of a Lévy distribution. Under this model, fractal fields exhibit fractional diffusion where the error in each time step can have wide and fluctuating values.
Chaos theory is a field of applied mathematics where seemingly insignificant differences in initial conditions produce significant system outputs changes. This makes long-term deterministic models and predictions impossible [
59]. Fractional diffusion is in itself a chaotic iterative process where the next step is the result of some form of non-linear function.
A basic form for a chaotic iterative signal is
where
is some non-linear function that depends critically on the value of a parameter or parameter set. In coupled systems, a single non-linear element will render the entire system as chaotic. These systems are Iterated Function Systems (IFS). Introduced in 1981 by John E. Hutchinson, they provide methods of constituting self-affine fractal geometries and signals [
60].
In regard to Equation (
10), the stability of an iterative process must be quantified. This can be done by considering the following iteration:
In this case, the error
at any iteration
n is a measure of the rate of separation of values between two iterations. The Lyapunov exponent (
) is based on modelling this error as
Formally, this exponent is the measure of the rate of separation of infinitesimally close trajectories [
61]. In this context, it can be used to examine how sensitive a signal is to an initial condition, i.e., how chaotic (unstable) the signal is. This is because, if
, the error at each iteration will increase rapidly. Similarly, if
, the error will undergo exponential decay as the iteration progresses. From Equation (
11), it is clear (summing over
) that we can write the Lyapunov exponent in the form
In the case of an FTS,
is taken to the price value
.
For a cryptocurrency time series,
is calculated computationally as the sum of the log changes in price. From Equation (
12), it is clear that if
, then the iterative process is stable; i.e., it converges. Moreover, if
, the process will diverge. Based on this observation, a positive value of
denotes a signal with a chaotic characteristic and a negative value signifies a convergent, stable system. As values of
increase, so does the rate at which the signal will converge or diverge [
62].
Using this simple formula applied to an FTS assumed to be of the form of Equation (
5), the Lyapunov exponent can be calculated for a small ‘look-back’ window of size
N to once again create a metric signal that will represent a time evolution of the nature of the time series. As
is merely a normalisation factor, if
, then
and visa versa. Therefore, a change in polarity of the metric is a sign of the gradient of the time series, and thereby an indicator of the growth or decay of the currency price. The Matlab code used to implement this calculation for an input time series is presented in
Appendix A.1.
2.4.3. The Lyapunov-to-Volatility Index
The accuracy of an FMH trend indicator, be it the Lyapunov or Beta indexes, relies on the size of the moving window used during the calculation. This moving window introduces a trading delay into the system directly proportional to its size. In the case of and , transitioning from positive to negative, or visa-versa, is an indication of a transition in price trend. This transition manifests as the metric’s graphical zero-crossings. However, this in itself, is not a measure of the stability of this trend.
The volatility,
, is a measure of FTS stability. A trend indicator can be scaled using the volatility to produce not only a predictor of future trends, but also as a measure of the stability of that trend. In the case of the Lyapunov exponent, the resulting index is the Lyapunov-to-Volatility Ratio;
where the volatility is defined as
where
is the price value. The Matlab function for computing the volatility is presented in
Appendix A.2.
The same scaling can be applied to , i.e., , which now provides a different method, from a different theoretical base, to indicate changing price trends and their stability. These can be compared during system testing to observe any changes in accuracy or returns.
Although the LVR can now double as a stability indicator, the position of the zero-crossings, which recommend long or short positions, will depend on the accuracy of . This itself depends on the inherent ‘noise’ present within the FTS. An FMH model assumes stationarity, which is not the case in Lévy distributed stochastic fields. Their tendency to exhibit Lévy flights create ‘micro-trends’ that can have a dramatic effect on the system. This can yield errors in the exact position of the zero-crossings (buy/sell positions), especially with respect to small term trend deviations.
To mitigate this effect, a moving average filter can be applied to the FTS,
, reducing the effect of the noise. This can be represented in the following continuous form:
where
calculates the mean of the windowed data of length
W. This method is also applied to the resulting LVR signal so that
In this case, however, the windowed data is of length
T, the financial index calculation period. Using the LVR, the zero-crossings are simply defined by the following Kronecker delta functions:
where
t is the current position of the time series and
is a small time step. Essentially,
represents the end of a bear trend and the beginning of a bull trend and the opposite transition is represented by
. Combined with the previous Matlab functions, this provides a method of trend prediction that can be used to analyse BTC-USD/ETH-USD FTS and recommend long or short trades. Exploration into the optimisation of the
W and
T parameters for a specific FTS is conducted in
Section 5.1.
4. Time Series Modelling Using Symbolic Regression
A period of trend stability empirically states that tomorrow’s price will be similar to today’s. Thus, formulating equations to represent the data up to that point in time (i.e., within the ‘stable period’) would allow discretised progressions into the future for a short number of time steps. This process can continue for as long as the stability period persists on a moving window basis.
This methodology suggests the use of a non-linear trend matching algorithm. In this respect, Symbolic Regression is a method of Machine Learning (ML) that iterates combinations of mathematical expressions to find non-linear models of data sets. Randomly generated equations using primitive mathematical functions are iteratively increased in complexity until the regression error is close to 0, or terminated by user subject to a given tolerance, for a pre-defined set of historical data.
4.1. Symbolic Regression
Symbolic Regression (SR) is a data-driven approach that identifies mathematical expressions describing relationships between input and output variables. It is a machine learning approach that aims to discover underlying mathematical expressions that best describe a dataset. Unlike traditional regression methods, which require the researcher to specify a model structure in advance, SR searches over a space of symbolic mathematical expressions (e.g., polynomials, logarithms, trigonometric functions) to identify both the functional form and parameters that optimally fit the data [
65,
66]. This is often achieved using evolutionary algorithms such as genetic programming, though recent work also integrates deep learning and reinforcement learning techniques to improve efficiency.
Key aspects of SR include:
- (i)
Representation of Models: Models are represented as expression trees, where nodes correspond to mathematical operators (e.g., +, −, ×, ÷) and functions (e.g., sin, cos, exp, log), while leaves correspond to variables or constants.
- (ii)
Search Space: The search space includes all possible mathematical functions, expressions and combinations thereof within a specified complexity limit.
- (iii)
Fitness Evaluation: Fitness is based on criteria such as accuracy (e.g., mean squared error) and simplicity (e.g., number of nodes in the expression tree).
- (iv)
Parsimony: Balancing accuracy and simplicity prevents overfitting.
One major advantage of symbolic regression lies in its interpretability. Whereas models such as neural networks or ensemble methods (e.g., random forests) typically operate as “black boxes,” SR produces explicit mathematical formulas that can be easily interpreted, analysed, and validated by domain experts. This feature is particularly valuable in scientific research and engineering, where understanding the governing dynamics of a system is as important as accurate prediction. Furthermore, SR can capture nonlinear relationships without presupposing the form of the dependency, making it highly flexible and capable of uncovering novel insights.
However, symbolic regression also presents notable limitations. The search space of possible symbolic expressions is vast and often computationally expensive to explore, especially for high-dimensional data. This can lead to long training times and difficulties in scaling SR methods compared to more established algorithms such as neural networks, which are highly optimised for large-scale data. Moreover, SR is prone to overfitting, as complex symbolic models may capture noise rather than true structure, particularly in small or noisy datasets. By contrast, decision trees or neural networks often incorporate regularisation techniques to mitigate this risk.
SR provides a transparent and flexible alternative to traditional machine learning approaches, excelling in domains where interpretability and discovery of governing equations are paramount, but facing challenges in scalability and robustness. The applications of this approach using a specific system—TuringBot—is discussed further detail in the following section.
4.2. Symbolic Regression Using TuringBot
Symbolic Regression is based on applying biologically inspired techniques such as genetic algorithms or evolutionary strategies. These methods evolve a population of candidate transformation rules over successive generations to maximise a fitness function. By introducing mutations, crossovers, and selections, the approach can explore a vast space of mathematical configurations. This approach is particularly useful in generating non-linear function to simulate complicated time series. In this section, we consider the use of the TuringBot to implement this approach in practice.
In this work, we use the
TuringBot [
67], which is a symbolic regression tool for generating trend fits using non-linear equations. Based on Python’s mathematical libraries [
68], the TuringBot uses simulated annealing, a probabilistic technique for approximating the global maxima and minima of a data field [
69]. For this reason, we now provide a brief overview of the TuringBot system.
4.3. TuringBot
TuringBot [
67] is an innovative AI-powered platform designed to streamline the process of algorithm creation and optimisation. By leveraging cutting-edge artificial intelligence, TuringBot enables users to generate, test, and refine algorithms for a variety of applications, such as data analysis, automation, and machine learning, without requiring extensive programming expertise. Its user-friendly interface and advanced capabilities make it an invaluable tool for professionals, researchers, and students seeking efficient solutions to complex computational problems. TuringBot stands out as a versatile and accessible resource in the ever-evolving landscape of artificial intelligence and technology. TuringBot.com is a symbolic regression software tool (version 3.1.4) designed to automatically discover analytical formulas that best fit a given dataset. Symbolic regression differs from traditional regression techniques by not assuming a predefined model structure; instead, it searches the space of possible mathematical expressions to find the one that best explains or ‘fits’ the data. TuringBot was developed to make this process efficient, intuitive, and accessible, especially for those without a background in machine learning or advanced data science.
Launched in the late 2010s, TuringBot was created in response to the growing need for interpretable artificial intelligence models. While most machine learning tools rely on complex neural networks or ensemble models that are difficult to understand and verify, TuringBot emphasised simplicity, transparency, and mathematical interpretability. By combining evolutionary algorithms with equation simplification techniques, TuringBot is able to produce human-readable formulas that describe complex datasets. This makes it particularly valuable in scientific, engineering, and academic contexts, where understanding the model structure is as important as prediction accuracy. The software is available for Windows and Linux and comes with a straightforward graphical user interface, allowing users to load data, configure parameters, and generate formulas with minimal setup. TuringBot also offers command-line integration for advanced users and supports exporting results for further analysis. As of the mid-2020s, TuringBot has gained a user base across various domains, from physics and finance to biology and control systems. It continues to evolve with improvements in computational performance and integration with modern workflows. The name “TuringBot” pays homage to Alan Turing, reflecting the tool’s focus on combining algorithmic intelligence with human-understandable outputs.
In an era increasingly focused on explainable AI, TuringBot stands out as a lightweight, focused solution for data modelling grounded in classical mathematical reasoning. In this context, the system uses symbolic regression to evolve a formula that represents a simulation of the data (a time series) that is provided, subject to a Root Mean Square Error (RMSE) between the data and the formula together with other ‘solution information’ and ‘Search Options’.
TuringBot automates the search for analytical models from data. At its core, the system employs a genetic programming-based algorithm that evolves candidate mathematical expressions through iterative recombination, mutation, and selection processes. This evolutionary approach allows the system to explore a wide variety of functional forms, including polynomials, logarithmic, exponential, and trigonometric structures, without requiring the user to predefine the shape of the model.
The fitness criterion used by TuringBot is primarily based on measures of predictive accuracy, most often the Mean Squared Error (MSE) between the model’s predictions and observed target values. By minimising this error, the algorithm promotes the survival of symbolic expressions that better capture the underlying data structure. In addition to error-based metrics, the software also applies parsimony pressure, a complexity penalty that discourages overly large or convoluted formulas. This balance between accuracy and simplicity reflects the principle of Occam’s razor and reduces the likelihood of overfitting.
Hyper-parameter tuning is handled through a combination of automated search and user-defined configuration. Parameters such as population size, mutation rate, crossover rate, and maximum tree depth play a central role in guiding the exploration of the symbolic space. In practice, TuringBot provides flexible controls that allow users to adjust these settings depending on the complexity of the dataset and the trade-off between exploration and exploitation. Iterative experimentation, often supported by built-in heuristics, ensures that hyper-parameters converge toward values that produce both accurate and interpretable symbolic models.
4.4. Comparison of TuringBot with Standard Machine Learning Models for Time Series Prediction
Time series prediction has traditionally been approached using statistical or machine learning models such as ARIMA, LSTM neural networks, and tree-based algorithms like XGBoost. In recent years, symbolic regression tools such as TuringBot have emerged as promising alternatives, offering distinct advantages in interpretability and adaptability. A comparison of these approaches highlights their relative strengths and limitations.
4.4.1. ARIMA
Autoregressive Integrated Moving Average (ARIMA) is a widely used classical model for time series forecasting. It performs well when data exhibit strong autocorrelations and stationarity. However, ARIMA requires significant pre-processing, including stationarity checks and differencing, and struggles to model nonlinear dynamics. In contrast, TuringBot, through symbolic regression, can capture nonlinearities directly by searching for functional relationships, reducing the reliance on strict assumptions about the data.
4.4.2. LSTM
Long Short-Term Memory (LSTM) networks, a deep learning extension of recurrent neural networks, are designed to model long-range dependencies in sequential data. LSTMs have achieved state-of-the-art results in complex prediction tasks, particularly when large datasets are available. Nevertheless, they are computationally expensive, require extensive hyper-parameter tuning, and produce models that are essentially “black boxes.” TuringBot offers a counterpoint: its outputs are interpretable mathematical expressions that reveal underlying dynamics, which is particularly valuable in domains like finance where transparency is critical. However, TuringBot may underperform compared to LSTMs in highly noisy or large-scale datasets where neural networks excel.
4.4.3. XGBoost
XGBoost, a gradient boosting algorithm, is known for its efficiency and predictive power across structured data problems. For time series, XGBoost is often used after feature engineering (e.g., lags, rolling statistics). While highly accurate, XGBoost models lack inherent interpretability and require manual feature design. TuringBot automates the discovery of functional relationships, potentially reducing the need for extensive feature engineering. Yet, XGBoost can handle large datasets more efficiently than symbolic regression, which may struggle with scalability.
4.4.4. Discussion
Table 2 provides a comparison of TuringBot with some standard models for time series prediction focusing on their strengths and limitations.
Overall, TuringBot distinguishes itself by combining predictive modelling with interpretability, a rare feature among traditional models. While ARIMA, LSTM, and XGBoost may outperform it in specific scenarios—such as stationary linear series (ARIMA), long memory dependencies (LSTM), or high-dimensional data (XGBoost)—TuringBot’s ability to generate concise symbolic formulas provides unique insights into system behaviour. For applications where understanding model structure is as valuable as forecasting accuracy, TuringBot offers a compelling alternative to standard time series models.
In this work, we are interested in using the system to simulate a cryptocurrency time series of the types discussed earlier in the paper. For this purpose, a range of mathematical operations and functions are available including basic operations (addition, multiplication and division), trigonometric, exponential, hyperbolic, logical, history and ‘other’ functions. While all such functions can be applied, their applicability is problem specific. This is an issue that needs to be ‘tempered’, given that a TuringBot generated function will require translation to a specific programming language. This requirement necessitates attention regarding compatibility with the mathematical libraries that are available to implement the function in a specific language and the computational time required to compute such (nonlinear) functions. There is also an issue of how many data points should be used for the evolutionary process itself. In this context, it is noted that the demo version used in the case studies provided in the work only allows a limited number of data points to be used; i.e., quoting from the demo version (2024): ‘Only the first 50 lines—data points—of an input file are considered’.
4.5. Example Case Study
Figure 10 shows an example of a set of 50 BTC-USD data points from 10 August 2021 to 5 September 2021 (red) and the trend match outcome of the TuringBot ML system (blue). These results were acquired by manually entering 50 data points and running the system. Here, the equation of the line was achieved after ∼103,000,000 iterations with an RMS error of
and mean absolute error of
.
The non-linear equation for the ‘best fit’ shown in
Figure 10 is given by
Figure 10 shows no future predictions; it is purely a trend match for a period of relative ‘trend stability’. The principal point is that Equation (
15) can be used to evolve a small number of time steps into the future to estimate price fluctuation. By coupling the results of doing this with the scale of the LVR, for example, a confidence measure can be associated with the short-term forecast that are obtained. This is because a large positive or negative values of the LVR reflects regions in the time series where the log-term volatility is low, thereby providing confidence the forecast that is achieved. This is the principal associated with the financial time series analysis provided in the following section.
5. Bitcoin and Ethereum—Financial Time Series Analysis
With the derivation of market indexes and subsequent implementation in Matlab, an analysis of BTC and ETH FTS can be performed. Using the
function, a collection of series were created for two time scales, yearly opening day prices ranging from February 2016–February 2022, and monthly opening hour prices from November–February. A List of these series for BTC-USD is given in
Table 3, the same collection was created for ETH-USD.
Using these, the analysis can be done on separate time scales, to not only determine the systems overall performance, but also whether the system can capitalise on the self-affine nature of the crypto-markets. For the purpose of the report, the BVR was used for analysis with the corresponding LVR results presented later in the paper for comparison.
5.1. Daily Backtesting and Optimsiation
The first step in the analysis was to not only find the optimum parameter combinations for highest accuracy and profit, but to define a rule for selecting these parameters from a set range. Starting by running the optimisation function for the BTCUSD2021 time series, the range of parameter combinations that resulted in a 100% evaluator accuracy was observed in a three dimensional mesh plot. For the remainder of this section, the Filtering Window Width and Financial Calculation Window will be referred to as
W and
T, respectively.
Figure 11 shows the resulting mesh graphic produced by the backtesting system.
It displays a broad range of W, T combinations. Due to the delay caused by the filtering process, it is logical to choose low values of W. From the mesh, there are W values in excess of 90 data points, which is equivalent to over 3 months of delay in the analysis. However, small values (i.e., ) may not smooth the data enough to make the system perform well. This is confirmed by the lack of a accuracy result with .
The effect of different sizes of
T, however, is not yet fully clear. Backtesting for two
combinations, one with a low
T value and one with a high value, is presented in
Figure 12 for BTCUSD2021. It shows that for
the metric signal becomes a binary representation, alternating between
. This gives the trader no indication of a fluctuation in
leading to a trade position.
High
T values result in sinusoidal fluctuations in
making it hard to define periods of high stability and fast movements in trends. This is due to the assumption of stationarity within the windowed data, used to approximate the convolution integral in Equation (
9), not being feasible for such a large value of
T. Given that
W values should minimise delay whilst providing enough filtering to reduce noise, a suitable limit for the values of
T would be
.
From this comparison, T values should aim to be . W values should aim to be small enough to reduce system delay whilst maintaining a smooth enough price signal for good system performance.
To further study the optimum
range, the array of
combinations that returned 100% accuracy (
) from the
function can be used to generate a new mesh plot of ROIs. As presented in
Figure 13, this proves that not all optimum positions result in profitable trade positions. High-value combinations of
generally result in a loss over the year. However, from the topology in
Figure 13, it is clear that low values of
result in profitable trades irrespective of
W.
This provides evidence to suggest that the highest returns are achieved when a low value of
T is chosen for the smallest
W value, in this instance
(For
combinations only). To see how the returns for
positions compare to non-optimum positions, i.e.,
combinations that produced less than 100% accuracy, a separate mesh plot was generated where all combinations are considered. This is presented in
Figure 14, where yellow represents high ROIs and dark blue low (negative). For the mesh topology, this mesh provides evidence that the
combinations do create high returns relative to all combinations. Interestingly, it also displays that very small
W,
T values create large losses and parameter sets where
also create losses.
A significant discovery extracted from this mesh plot is that the highest possible returns do not occur at the optimum positions. The highest ROI from
Figure 14 is selected and the surrounding peaks do rise higher. This can be interpreted as the result of micro-trends in the time series where non-optimum parameters have fortuitously recommended trades during a local peak or trough that has yet to influence the windowed data. The aim of the system is to ensure accuracy and confidence in the trading strategy; therefore, optimum evaluator parameter combinations are preferred to highest profit achieving combinations. This priority definition warrants another evaluation of the optimum positions.
Figure 15 presents a different perspective of the data displayed in
Figure 11 where only the 100% accuracy combination are displayed on a two dimensional ‘Top Down’ view.
Figure 15 identifies a small grid at the bottom left hand corner where
W values are within the lowest range and
T values are
. Knowing that this grid achieves the highest ROIs as shown in
Figure 14, it is therefore preliminarily proposed that this ‘Grid of Choice’ (GOC) represents the best range of
to be chosen from. To provide more evidence of this theory, the same approach was taken for the other yearly time series of BTC-USD. Each year displayed the same properties lending weight to the GOC theory and evidencing that the assumption of a fractal stochastic field is constant throughout the data.
Figure 16 shows the resulting ‘Top Down’ optimum parameter mesh (left) and the ROI mesh (right) for BTC-USD 2020-21. It shows results consistent with that of BTC-USD 2021-22. The ridge of high (yellow) returns visible in
Figure 16 are in some cases
higher than the returns achieved under
. However, most of these parameter combinations violate the required conditions, in this case
and high
W values. This could be attributed to the fact that in 2020 BTC-USD had an almost constant upward trend.
With optimum parameters for 2021-22 (
), backtesting was performed.
Figure 17 shows the graphical output from the function with,
in green,
in blue, the filtered data in red and raw price signal in black. The same colour format will be used for all backtesting outputs in this report. It displays the 7 trades, resulting in a
return in a year when Bitcoin’s value against the dollar widely fluctuated and lost value overall.
The nature of the trading delay is clear, with the filtered data (red) lagging behind the raw price data (black). The signal shows that there are periods of general trend stability in both bear and bull directions. By inspection, it can be seen that although some trade indications occur in the trough of the filtered data, when applied to the raw signal, the difference between the two price signals results in an overall loss for that transaction. This is a result of the micro trends Bitcoin displays coupled with the trading delay and inherent volatility.
The backtesting was completed on the same basis for the rest of the BTC-USD financial time series as well as ETH-USD. Results are displayed in
Table 4. From these results, it is not always possible to achieve 100% accuracy. However, this does not lead to a loss for the year. It should be noted that during the analysis of ETH-USD data, the correlation between optimum parameters and high returns, including the GOC, was observed to provide further evidence in favour of the parameter selection theory.
5.2. Hourly Backtesting and Optimisation
Given that the BTC-USD market has been shown to be a self-affine fractal signal exhibiting scale invariance, backtesting over a different scale, in this case hourly prices, should return similar results. However, as can be seen from
Table 3 the monthly time series have double the number of data points. For this reason, the field is expected to have a higher level of detail and therefore noise. To inspect this,
Figure 18 shows a 20-day extraction from the January to February 2022 time series. The high volatility and wild price fluctuations are more prevalent than for the daily data, with micro-trends occurring faster with bigger relative movements.
Due to the increased noise content, a higher level of filtering was expected to maintain an acceptable level of accuracy and therefore confidence in the recommended trade positions. This results in a GOC where T levels remain consistent but W values rise significantly.
As with the daily time series, the first step is to run the
function for the BTCUSD January–February 2022 field.
Figure 19 shows the resulting mesh plot for all parameter combinations. Compared to the daily data, the general topology is far lower and as excepted, 100% accuracy is achieved with much higher values of
W, in this case
.
T values remain consistently low, an expected result due to the self-affinity of the underlying price signal.
Taking a further look at the ‘Top Down’ view of the optimisation mesh plot in
Figure 20, few accurate combinations of parameters exist. However, even with the sparsity of the results, there is still a clear grid containing the small range of
W values for low
T values. This is consistent with the expected findings. Producing the mesh plot for parameter returns,
Figure 21, confirms that the GOC remains a source of strong returns.
Analysing the mesh plot of ROIs, the topology suggests an ideal location for parameters with returns around
being a peak surrounded by low and negative results. This lends further evidence that the GOC is a valid theory. An interesting outcome is the plateau of high returns for high
T values, irrespective of what filtering is applied. Many of these
combinations are invalid due to
or
, the ideal range of filtering for this data. An explanation for this could be that for high values of
T, the metric signal
becomes heavily sinusoidal, containing low frequencies. This could result in low numbers of trades operating at heavily delayed trade positions that are fortuitously executed. Other anomalous peaks in returns surrounding the origin also violate the
rule. Such small filtering sizes increases the expected number of trades to high and infeasible values, due to the fees synonymous with trading cryptocurrencies. A comparison of these two invalid
combinations is shown in
Figure 22. In the
backtest, a 0% accuracy still gives a positive return, confirming the anomalous nature of these combinations.
The sharp and focused nature of the
peak suggests that the effect of micro trends in the hourly time series is greater. Returns are reduced rapidly at small deviations from optimum combinations. The backtesting output for
is shown in
Figure 23. 9 trades are executed resulting in a
ROI. The increased volatility in the time series is reflected in the corresponding volatility in
.
When applied to the other monthly time series, an interesting observation is the increase of filtering required as the fields evolve in time, suggesting that both BTC and ETH are entering a phase of high volatility.
Table 5 displays the results for each monthly financial time series used in backtesting.
5.3. Analysis Using LVR
The backtests performed in previous sections were repeated for yearly financial time series using the LVR to observe any changes in results, The equivalent graphical LVR output for BTCUSD2021 is displayed in
Figure 24. Results were consistent with the BVR metric, confirming that both ratios are valid for trend analysis. The LVR produced a metric signal with a greater amplitude than
which provides more flexibility to change the conditions on which the trading positions are recommended. Currently trades are considered viable only when
crosses the axis. However, if the
signal was re-programmed to produce a delta peak when the signal reaches a certain threshold, say
, this could reduce trading delay. A full set of ROI results for all BTC and ETH time series is presented in
Table 6.
5.4. Returns on Investment—Pre-Prediction
A full comparison of
and
returns compared to the standard ‘Buy and Hold’ (B& H) strategy, where the price change over the whole time series is taken, is shown in
Table 7.
It shows that the proposed system outperforms B&H for every financial time series under consideration, both for BTC and ETH coins, with the exception of BTCUSD 2016-17 LVR. In bear dominant years of high market loss, such as BTC 2018-19, the system was capable of producing a positive return. In other cases, where the year saw high overall gains, the system was able to improve further.
These returns are high when compared to other stock market indexes, with average returns considered to be , including; S & P Commodity index returning an average between 2009 and 2019, and the S & P 500 an average of between 2005 and 2019. Overall, returns for the hourly data sets also proved to beat the B&H strategy.
5.5. Short-Term Price Prediction Using Machine Learning
As discussed in
Section 3.4, periods of high trend stability, indicated by a ‘strong’ BVR amplitude, signify the opportunity for short-term prediction using Symbolic Regression (SR). As this period continues, non-linear formulas can be re-generated every day based on historical opening prices on a rolling window basis. Once generated for a time point
, the formula can be evolved for short-term future time horizons
,
,
,… where
n is the total number of data points used to create the formula. The hypothesis is that data points within the stable trend period can be used to generate non-linear formulas capable of guiding a trader to the optimum position execution, with the prices preceding the period being volatile and therefore detrimental to the SR algorithm.
Applying this to the BTCUSD2021 time series, a high-BVR period can be defined as
or
and designated
. The BVR signal reaches this threshold on the 21st of August, indicating the ability to utilise SR. Advancing forward 34 days to September 10th, still within the
, the TuringBot is used to generate a trend formula using the previous 34 days opening prices (
). The resulting solution is
Using this equation, data points
,
,
,… can estimate price fluctuations over a short time horizon.
Figure 25 shows the actual BTC-USD price data from 21 August to 20 September in black, the trend ‘fit’ for historical data up to 10 September, shown in blue, and then future estimated prices for +5 days to 15 September in red and +10 to 20 September in green. Each price estimation figure will use the same format.
Observing the SR output graph, it is clear that the prediction provides no useful guidance. It does not predict the large drop in price on day 36, nor the increase in the preceding days. An optimal profit would have been achieved by exiting the long position on day 36 at
$52,697. However, the prediction estimates
$48,347, a price gap of nearly
. During the backtesting for this data set, a short position was recommended on 20 September at
$47,144. Using
Figure 25 as a reference, no increased profit would have been created. On 16 September (
)
drops below
and
ceases. This proves that no additional profit could have been made by the ML system. Formulas were generated for days 30–34, to see if an earlier prediction would have yielded better results. In every case, the estimated future prices gave no accurate guidance and failed to optimise the sell position.
Figure 26 contains prediction plots for
(top) and
(bottom). The latter of which shows widely fluctuating prices and provides no confidence in its accuracy.
As a test,
Figure 27 shows the non-linear formula generated at 7 September (Day 34) using the preceding 31 data points within
and an additional 3 points from before the stable period began (i.e.,
). This test goes against the ML hypothesis.
This output does provide credible guidance, indicating an exit of the long position on 10 September (Day 37) for
$49,940. Compared to the zero-crossing recommendation on 20 September, this new trade position increased profit by
$2800 or
, a significant increase in profit. To examine this further,
Figure 28 shows a formula generated for 2 August to 7 September, now using 20 data points from outside the
(
).
This result is based on using the following equation:
giving a very accurate prediction. It correctly estimates the short price rise before the sharp fall. An indicated exit on 11 September at
$51,800 increases profit by
$4656 or 10%. The interesting observation here is that more precise price estimations came from extending the ‘look-back’ window beyond the
. Due to the manual nature of the TuringBot, this process was only completed for the BTCUSD2021 time series and not for all time series.
6. Discussion
The analysis of both the cryptocurrencies considered in this research, has shown clear indications of non-normality (i.e., non-Gaussian behaviour). This is a defining characteristic of a fractal stochastic field. The peaked and broader side bands of the PDF for these financial signals deviate from a Gaussian PDF model, violating a core principle of the EMH. Disregarding the assumption of an efficient cryptocurrency market allows various indicators to be utilised to determine the financial fields nature. All these indicators were calculated as linear functions associated with spectral decay of the signal which was obtained through linear regression of the log power spectral plot. However, this method can lead to inaccuracies due to the erratic nature of the log power spectrum. As seen in
Figure 29, the gradient of the log-log regression line could have a range of values.
In order to obtain an accurate value of
, precise calculations of the spectrum and optimum region for fitting the regression line are required, which in most cases is not available [
70]. As all of the indicators considered are linearly related, more precise methods of calculating their values are available, such as the ‘Higuchi method’ for determination of
and the algorithms for computing the Hurst exponent [
71,
72]. However, in the case of crypto-markets, the collection of indicators used showed such high deviations from normality that their inaccuracy would have made no difference to the conclusions in association with the application of a self-affine field model.
Each index showed a different aspect of deviation. The Hurst exponent (H) showed a level of anti-persistence not consistent with RWMs. The Levy index () indicated a peaked PDF, indicative of a Lévy distribution, not a Gaussian distribution. The value of the fractal dimension () showed that the field has a narrower spectrum consistent with a self-affine signal. The ACF showed clear signs of data correlation, long-term market memory. This provided an overwhelming amount of evidence that the standard market hypothesis is not applicable and therefore the potential inaccuracy in the computation of can be ignored.
During backtesting and optimisation, a range of ideal values for the W, T parameters was alluded to. The need to reduce trading delay, whilst minimising the noise in the signal, proved to be dependent on the time scale of the time series, with larger filtering windows being required for the more detailed hourly data fields. Financial calculations had to be undertaken over small time windows relative to the length of data being analysed, whilst staying above the limit of 2. Unlike W, T values did not increase with an increase in scaling. Observing the positions, the location of the GOC can be acquired. This is related to the smallest range of W values for which T is minimised. This GOC was consistent, through all fields of the same currency and scale, leading to the guidelines for the parameter choice as follows: W—Smallest values for which noise is sufficiently removed; T—. Although this range, for the two parameters, did not always result in the most profitable trades, the high accuracy allowed high confidence in a profitable strategy. The fluctuation in ROI for neighbouring combinations can be explained by rapid micro-trends making an ‘ideal’ position recommendation impossible given the level of trading delay.
Another method to help overcome the trading delay is to redefine the conditions for which the position indicator produces a Kronecker delta. Currently, positions are only recommended when either or change polarity. However, if this was altered so that the Kronecker deltas were produced when a given threshold is passed, the system would react faster to changing trends. The implications of this are that price changes opposed to the current trend direction require less impact on the filter window to produce a Kronecker delta indicator.
The only down side to this change, is the increased risk due to parabolic metric flights caused by fast trend sweeps large enough to influence BVR/LVR results into recommending buy and sell positions in quick succession. However, looking at backtesting graphical outputs, it can be seen that a reduction in trading delay occurs more frequently. This change also depends on the metric used. The LVR produces a signal with a higher amplitude, giving more flexibility to choose a threshold range. Performances for both metrics were similar, as were returns in all backtests. As the metrics are derived from a different theoretical basis, one from a fractal model and the other from a chaotic model, their similar effectiveness further proves that cryptocurrency exchanges adhere to the FMH.
The self-affine behaviour of the cypto-markets under consideration allow for different time scales to be analysed. For example,
Figure 30 shows an extraction from the daily backtesting output for BTC-USD 2021-22 for January–February (left) and compares it with the same backtesting time period for hourly data.
The hourly data (right) shows a number of trades being recommended with a return over a period of overall loss in value. This is in stark contrast to the daily backtest, which showed no positions, therefore resulting in a loss. The most interesting comparison, is the opposite trade positions recommended for the last few data points, with daily data suggesting a purchase and hourly data suggesting a short market position.
The higher detail in the hourly data would be expected to produce more accurate positions. However, hourly data is particularly susceptible to micro-trends, requiring a large increase in the filtering window. This makes system outputs, based on hourly data, riskier; as evidenced from the reduced returns. Even though actual price changes over a month are less than over a year, the relative movement over the time period as a percentage is directly comparable over any scale.
Coupled with an increased number of trades per unit time, which has an effect due to the fees charged by the trading platform, the hourly trading system has its limitations. A combination of both strategies, whereby long-term trend analysis positions are complimented by short-term data, will increase the likelihood that the most profitable position can be achieved.
The application of ML to aid optimum trade executions provides price estimations that were highly inaccurate when using ‘look-back’ data contained within the . A range of non-linear equations, of increasing length, were created for sequential time steps within the FTS, each of which failed to predict the impending change of a trend, suggesting that increasing the window length within the has no effect on accuracy. However, accurate price estimations were produced when including a range of ‘unstable’ data points that proceed the window. During testing of look-back windows, that included values outside of the , it was observed that longer window lengths produce more precise results that correctly predict the magnitude and direction of the next 5–8 days within an accuracy of ∼90%. For example, using a window consisting of 50 data points (), including 20 unstable values, increased the profit for a single trade by ∼10%.
On re-evaluation of the approach, it becomes clear that any non-linear function based on a window of ‘stable’ data will only continue to display this trend when evolved forward in time. This is clearly a fundamental flaw, as, by definition, this approach will not achieve the goal of predicting a future change in trend. Extending the window beyond the creates an equation that accounts for both the low volatility trend and the high volatility movements, therefore giving a more accurate representation of the data and a better basis for future estimation. Tests on a window length of ∼30 produced vastly improved results, suggesting that increasing the window length is not the primary way to improve estimations. Nevertheless, increases did improve prediction accuracy by ∼20%. Thus, it can be concluded that an increase in the window length, using stable and unstable data, increases future price prediction accuracy.
Considering that the BVR and LVR are both calculated using their own rolling windows of length T, this extension should be at least T steps beyond the .
The manual nature of the TuringBot results in highly inefficient calculations. The lack of SR integration within the system is a significant flaw, if ML is to be recommended as a strategy to reduce trading delay. A proprietary SR algorithm or existing library will greatly increase usability. However, this is outside the scope of this work. Extensive manual testing of the ML method was not able to be completed over the time frame available for this work. A beneficial evolution would be an implementation in Python where large ML and evolutionary computing libraries are currently available. However, more recently, TuringBot.com has released an API for their software, allowing remote access to the SR algorithm from within the system code. This provides another approach to increasing efficiency.
During code development, using truncation to preserve data length and vectorisation to improve performance proved vital to conducting an analysis, the slow nature of the original functions making the program non-viable for continued and efficient trading. The creation of the function made the system universal, where any .csv database of financial data could be uploaded and converted into a compatible time series.
Optimisation of the code resulted in a system that could generate an output in ∼0.0001 seconds. This was a significant improvement, making its use in conjunction with a live trading system viable. This decrease in computational time allows a continuous live data stream to be used, where fast tick times of a few seconds are implementable. However, this level of data requires a marked increase in filtering or new methods of creating price data. Opening daily prices can be replaced by an average of the underlying prices of the minimum time step.
The example results presented in
Table 8 provide a useful comparison of percentage return on investment (ROI) for BTC-USD and ETH-USD when using LVR and BVR indicators against a simple Buy and Hold strategy. These figures clearly demonstrate the potential for improved profitability when applying indicator-based strategies. However, ROI alone is a limited measure: it captures absolute gains or losses but neglects the variability of outcomes, prediction accuracy, and the risk-adjusted performance of a strategy. For a more robust evaluation, the incorporation of performance metrics such as the Root Mean Square Error (RMSE), Mean Absolute Percentage Error (MAPE), and the Sharpe Ratio is essential.
RMSE provides a measure of the deviation between predicted and actual values in the underlying financial time series. By penalising larger errors more heavily, RMSE highlights whether indicator-driven forecasts reliably track asset price movements. Similarly, MAPE offers an interpretable, scale-independent measure of forecast accuracy by quantifying the average percentage error. Together, these metrics ensure that strategies are not only profitable but also accurate and consistent in predicting price dynamics.
The Sharpe Ratio, in contrast, extends evaluation beyond forecast accuracy to financial viability. By calculating excess returns per unit of volatility, the Sharpe Ratio assesses how efficiently risk is converted into returns. A strategy with a higher ROI but excessive volatility may ultimately be less attractive than one with modest returns but a superior Sharpe Ratio. While the ROI results in
Table 8 highlight profitability, supplementing them with RMSE, MAPE, and Sharpe Ratios provides a more comprehensive assessment. These metrics collectively capture predictive precision, error magnitude, and risk-adjusted efficiency, offering a balanced framework for evaluating trading strategies in volatile cryptocurrency markets.
7. Symbolic Regression and Fractal Dynamics: A Critical Appraisal
The study of complex systems characterised by fractal dynamics presents substantial methodological challenges for traditional machine learning models. Financial markets, turbulent flows, ecological systems, and even certain physiological processes often display self-similarity, long-range dependence, and nonlinear feedback loops that resist representation by simple parametric models. Symbolic regression (SR), which searches for closed-form mathematical expressions directly from data, has emerged as a promising alternative to black-box methods such as neural networks or decision trees. This discussion evaluates why SR may outperform traditional models in the context of fractal dynamics, focusing on interpretability, functional discovery, scalability, and limitations.
7.1. Capturing Nonlinear and Self-Similar Structures
Fractal dynamics are typically defined by power-law relationships and recursive self-similar patterns across time scales. Traditional machine learning models, while capable of fitting highly nonlinear data, often do so in a statistical rather than structural manner. For example, a neural network may approximate a power-law curve through nonlinear activations, but it does not explicitly reveal the mathematical law underlying the data. In contrast, SR is specifically designed to uncover analytic functions, such as power laws or fractional differential equations, that are directly interpretable in fractal contexts. This functional discovery is crucial because fractal systems are often governed by compact mathematical expressions whose elegance is obscured by black-box models.
7.2. Interpretability and Theoretical Integration
A defining strength of SR lies in its interpretability. Whereas decision trees or neural networks produce models that must be treated as predictive artefacts, SR delivers explicit formulas. In fractal dynamics, interpretability is particularly important, as researchers seek not merely to predict but to understand the governing processes. For example, SR may reveal a scaling law consistent with fractional Brownian motion or uncover equations consistent with multifractal cascades. Such outcomes allow direct integration with existing theories of chaos and fractals, enabling models to serve as both predictive tools and theoretical insights. Traditional models rarely provide this level of epistemic value.
7.3. Adaptability to Multi-Scale Phenomena
Fractal systems exhibit multi-scale behaviour, where statistical properties at small scales resemble those at large scales. Symbolic regression naturally accommodates this property by exploring mathematical forms that explicitly capture scaling and recursive dynamics. In contrast, neural networks and ensemble methods are constrained by their architecture. While deep networks can approximate multi-scale patterns, they require large amounts of data and extensive training, often resulting in models that are computationally costly and prone to overfitting. SR, by seeking parsimonious mathematical rules, can often generalise more effectively across scales.
7.4. Robustness in Data-Limited Regimes
A further advantage of SR in fractal contexts is its efficiency in data-limited environments. Many fractal systems are difficult to observe over long horizons, and the available datasets may be noisy or sparse. Traditional machine learning methods thrive on large datasets, where their statistical approximations average out irregularities. By contrast, SR can leverage the inherent self-similarity of fractals to infer governing laws from relatively small samples, provided the symbolic search space is well managed. The explicit inclusion of parsimony criteria also reduces the risk of overfitting to noise, which is especially problematic in high-frequency fractal signals such as market tick data or turbulence measurements.
7.5. Limitations and Challenges
Despite these strengths, SR is not without limitations. The search space for symbolic expressions is combinatorially large, and naive implementations can be computationally intractable for high-dimensional data. While genetic programming and other heuristic algorithms alleviate this issue, they do not guarantee convergence to the globally optimal expression. Moreover, SR remains vulnerable to overfitting when model complexity is insufficiently penalised, especially in fractal systems where noise can mimic structural patterns. Traditional models such as random forests or gradient boosting often exhibit greater robustness to noise by design.
Another challenge is the interpretive bias introduced by SR’s reliance on a predefined library of functions. If fractal dynamics in a given system are best described by a function outside the candidate set (for example, certain fractional calculus operators), the algorithm may converge to an approximation that captures behaviour without revealing the true generative law. Neural networks, by contrast, are universal approximations and may perform better when theoretical constraints are weak or unknown.
7.6. Comparative Value in Practice
The relative value of SR over traditional models depends on the research objective. If prediction accuracy is the sole criterion, ensemble methods or deep networks may sometimes outperform SR, particularly on high-dimensional datasets with weak theoretical grounding. However, when the objective is to uncover interpretable, generalisable laws consistent with fractal dynamics, SR is superior. Its ability to generate human-readable models offers explanatory power, a feature increasingly demanded in both academic and applied domains such as finance, physics, and environmental science.
8. Conclusions: Summary, Discussion and Future Directions
Symbolic Regression (SR) offers a powerful approach for modelling fractal dynamics, with strengths in interpretability, adaptability to scaling phenomena, and robustness in data-limited environments. Although computational challenges and the risk of overfitting remain, SR’s ability to uncover underlying mathematical relationships sets it apart from traditional black-box methods. In situations where understanding the generative structure of fractal systems is as important as making accurate predictions, SR provides a methodological advantage that can complement, and in some cases outperform, neural networks or decision trees. Its primary value lies not only in prediction but also in linking empirical data to theoretical insight, making it a particularly suitable tool for advancing the study of fractal dynamics.
The fundamental analysis of BTC and ETH cryptocurrency markets is consistent with previous research. While most studies suggest that BTC is evolving toward a more mature and efficient market, the analysis presented here demonstrates that it remains far from fully efficient. Because no prior research has applied similar approaches for price and trend prediction in cryptocurrency markets, direct comparisons with other studies are not possible.
The analysis further indicates that cryptocurrency price data do not conform to a normal distribution. For Bitcoin, the Hurst exponent was calculated as , indicating anti-persistence and short-term dependence. The Lévy index was measured at , with a fractal dimension of . Long-term market memory was also evident from the autocorrelation functions. These results confirm that cryptocurrency markets are inefficient, validating the relevance of a fractal modelling approach. In this context, the self-affine properties of the markets were confirmed by observing similar probability density functions (PDFs) across scaled time series, a pattern that is also reflected in Ethereum markets.
8.1. Summary
Based on the principles of the FMH, two fundamental indicators were derived from distinct theoretical frameworks, each scaled by volatility to produce a pair of trend analysis metrics. The effectiveness of these metrics is reflected in the zero-crossings of their signals, which indicate potential changes in trend. Positive returns were observed across all analysed time series, even during bear-dominated periods.
Parameter sweep analysis provided a basis for selecting optimal parameter values. Filter window sizes must be minimised to reduce trading delays while still removing noise to ensure accurate system outputs. For a given range of filter widths (W), the window length (T) should satisfy . This parameter selection rule was shown to yield high accuracy and profitable returns. Short time steps produce time series with elevated noise levels, reducing trend prediction accuracy and increasing susceptibility to micro-trends, which can lead to positions that incur losses. Consequently, a combination of long- and short-term analyses is recommended, where short-term results identify the most opportune moments to enter positions as guided by the associated long-term trends.
The application of machine learning (ML) to predict short-term price fluctuations and thereby reduce trading delays has been shown to be unreliable unless the data window extends at least T steps beyond the . Longer windows containing more data points improve predictive accuracy. In general, the TuringBot did not provide an efficient implementation of SR. Future system improvements should therefore focus on fully integrating ML-based approaches.
The analysis demonstrates that this methodology consistently outperforms traditional ‘Buy & Hold’ strategies across all time series considered and surpasses benchmark returns set by conventional stock market indices. Given the overall performance of cryptocurrencies over the past five years, this outcome is unsurprising; however, profitable returns were still achieved during bear-dominated periods.
Future research should investigate scale-combination strategies to further enhance profitability. A natural progression would be the development of a fully integrated SR algorithm. Additionally, further studies could explore whether a broader set of cryptocurrencies exhibit fractal properties and examine potential price correlations among them.
8.2. Discussion
The aim of this publication is to provide readers with a comprehensive background on the algorithms developed in this study. Apart from the introductory material presented in
Section 1 and
Section 2, the results reported here are, to the best of the authors’ knowledge, novel and original, particularly in terms of their application, the numerical outcomes obtained, and, more specifically, the type of data analysed—namely, cryptocurrencies.
A key feature of this work is the provision of the Matlab code used in the investigation. This enables readers to reproduce the results and extend the methodology, which the authors consider an important contribution, particularly for those interested in applying these algorithms to commodities markets or other financial domains.
It is worth noting that TuringBot is only one of several emerging tools in the field of genetic programming capable of evolving nonlinear functions to simulate real-world noise. Several Python-based alternatives allow the development of fully integrated programs without reliance on external applications such as TuringBot. One example is
gplearn, a Python library implementing genetic programming with an API inspired by and compatible with scikit-learn [
73].
8.3. Future Directions
The approach considered in this work is algorithmic, in that both long-term trends and short-term price values are derived from a set of quantifiable algorithms and their optimisation. For long-term trends, these algorithms are based on functions that compute metrics associated with the Fractal Market Hypothesis (FMH). For short-term price predictions, the algorithms are derived from nonlinear functions generated iteratively using symbolic regression. In this sense, the methods presented here integrate conventional time series modelling with machine learning techniques.
This approach contrasts with deep learning models, which can capture complex relationships between features in time series data and account for long-term dependencies [
74]. While deep learning models have the potential to improve prediction accuracy, they require large volumes of training data to operate effectively.
A natural avenue for future research is to compare the algorithmic approach presented in this paper with deep time series forecasting models, using data specific to cryptocurrency trading and other commodities.