Next Article in Journal
Entropy-Based Approach for the Detection of Changes in Arabic Newspapers’ Content
Next Article in Special Issue
Upgrading Model Selection Criteria with Goodness of Fit Tests for Practical Applications
Previous Article in Journal
A Nonparametric Bayesian Approach to the Rare Type Match Problem
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Association Factor for Identifying Linear and Nonlinear Correlations in Noisy Conditions

by
Nezamoddin N. Kachouie
1,*,† and
Wejdan Deebani
2
1
Department of Mathematical Sciences, Florida Institute of Technology, Melbourne, FL 32901, USA
2
Deparments of Mathematics, College of Science and Arts, King Abdulaziz University, P.O. Box 344, Rabigh 21911, Saudi Arabia
*
Author to whom correspondence should be addressed.
Current address: 150 W. University Blvd., Melbourne, FL 32901, USA.
Entropy 2020, 22(4), 440; https://doi.org/10.3390/e22040440
Submission received: 26 January 2020 / Revised: 4 April 2020 / Accepted: 6 April 2020 / Published: 13 April 2020

Abstract

:
Background: In data analysis and machine learning, we often need to identify and quantify the correlation between variables. Although Pearson’s correlation coefficient has been widely used, its value is reliable only for linear relationships and Distance correlation was introduced to address this shortcoming. Methods: Distance correlation can identify linear and nonlinear correlations. However, its performance drops in noisy conditions. In this paper, we introduce the Association Factor (AF) as a robust method for identification and quantification of linear and nonlinear associations in noisy conditions. Results: To test the performance of the proposed Association Factor, we modeled several simulations of linear and nonlinear relationships in different noise conditions and computed Pearson’s correlation, Distance correlation, and the proposed Association Factor. Conclusion: Our results show that the proposed method is robust in two ways. First, it can identify both linear and nonlinear associations. Second, the proposed Association Factor is reliable in both noiseless and noisy conditions.

1. Introduction

Analyzing large datasets is becoming central in science, engineering, and technology. In data mining and statistical analysis, it is essential to detect relationships between different variables [1]. Different correlation factors have been introduced to identify and quantify the relationship between variables. Pearson’s correlation coefficient has been broadly used to identify and measure the strength and direction of a linear relationship between two variables.
Pearson’s correlation can effectively detect linear relationships; however it is not reliable to identify nonlinear relationships between two variables. To address this shortcoming of Pearson’s correlation, Distance correlation was introduced by Gábor J. Székely [2,3] to find both linear and nonlinear relationships between two variables. Regardless of the relationship type, Distance correlation quantifies the degree of correlation by a value between zero and one. Values close to one represent strong correlation, while values close to zero suggest weak correlation between two variables. It has been demonstrated that Distance correlation is superior to Pearson’s correlation for identifying nonlinear relationships.
Several extensions of Distance correlation have been introduced such as Invariant Distance correlation [4], Conditional Distance correlation [5], Distance Correlation of Lancaster distributions [6], Distance Standard Deviation [7], Distance Correlation for locally stationary processes [8], Distance correlation coefficient for multivariate functional data [9], Partial Distance correlation [10], and Distance correlation t-test [11]. Distance correlation has been broadly used for different applications such as time series [12,13], clinical data analysis [14], genomics [15], and biomedical data analysis [16].
Although Distance correlation can identify and quantify nonlinear correlations, it does not necessarily obtain the same or comparable values for different nonlinear relationships. For example, the Distance correlation of an exponential relationship could be higher than a quadratic relationship. Moreover, Distance correlation values drop in noisy conditions and may not robustly demonstrate the strength of the correlation, where low correlation values may contribute to the wrong conclusion about the strength of relationship between two variables.
To address these shortcomings and improve the performance of Distance correlation, in this paper we propose the Association Factor (AF). The proposed AF performs robustly with regard to identifying both linear and nonlinear relationships. Moreover, we show that AF performs robustly in noisy conditions and outperforms Distance correlation in identifying noisy linear and nonlinear relationships. An overview of Pearson’s correlation and Distance correlation is provided in the next section. The proposed Association Factor is presented in Section 2. Simulation models, Results, and Conclusions are presented in Section 3, Section 4 and Section 5 respectively.

2. Methods

2.1. Quick Review

2.1.1. Pearson’s Correlation

Pearson’s correlation is a measure of the strength and direction of the linear relationship between two variables. Its score ranges between −1 and one, and it describes the degree to which one variable is linearly related to another. Pearson’s correlation between two variables X and Y is defined by:
ρ = c o v ( X , Y ) σ X σ Y
where c o v ( X , Y ) is the covariance between X and Y, σ X is the standard deviation of X, and σ Y is the standard deviation of Y. Pearson’s correlation is essentially the covariance of X and Y normalized by the product of the standard deviations of X and Y [17].

2.1.2. Distance Correlation

Distance correlation is a measure of the correlation between two random vectors X and Y, and its value ranges from zero to one. Analogous to product-moment correlation (Pearson’s correlation), Distance correlation can identify linear and nonlinear correlations using Euclidean distance. The empirical Distance correlation [2] is computed by:
R 2 ( X , Y ) = ν n 2 ( X , Y ) ν n 2 ( X ) ν n 2 ( Y ) , ν n 2 ( X ) ν n 2 ( Y ) > 0 , 0 , ν n 2 ( X ) ν n 2 ( Y ) = 0
where R ( X , Y ) is empirical Distance correlation, ν n ( X , Y ) is empirical Distance covariance of X and Y, ν n ( X ) and ν n ( Y ) are empirical Distance variances of X and Y respectively, n is sample size, and ν n ( . , . ) is a scalar. R ( X , Y ) is zero if and only if X and Y are independent.

2.2. Proposed Association Factor

In this paper, we introduce the Association Factor (AF), the Distance correlation of Optimal Transformations of variables X and Y:
R A F ( X , Y ) = R ( h 1 ( X ) , h 2 ( Y ) )
where R A F ( X , Y ) is the proposed AF, and h 1 : d o m ( X ) B and h 2 : d o m ( Y ) C are measurable mean zero transformations where B , C R , and for ν n 2 ( h 1 ( X ) ) ν n 2 ( h 2 ( Y ) ) > 0 , we have:
R A F 2 ( X , Y ) = ν n 2 ( h 1 ( X ) , h 2 ( Y ) ) ν n 2 ( h 1 ( X ) ) ν n 2 ( h 2 ( Y ) )
where ν n ( h 1 ( X ) , h 2 ( Y ) ) is empirical Distance covariance of h 1 ( X ) and h 2 ( Y ) , ν n ( h 1 ( X ) ) and ν n ( h 2 ( Y ) ) are empirical Distance variances of h 1 ( X ) and h 2 ( Y ) respectively, n is sample size, and ν n ( . , . ) is a scalar:
ν n 2 ( h 1 ( X ) , h 2 ( Y ) ) = 1 n 2 k , l = 1 n A k l , h 1 B k l , h 2
ν n 2 ( h 1 ( X ) ) = ν n 2 ( h 1 ( X ) , h 1 ( X ) ) = 1 n 2 k , l = 1 n A k l , h 1 2
ν n 2 ( h 2 ( Y ) ) = ν n 2 ( h 2 ( Y ) , h 2 ( Y ) ) = 1 n 2 k , l = 1 n B k l , h 2 2
where:
A k l , h 1 = a k l , h 1 a ¯ k . , h 1 a ¯ . l , h 1 + a ¯ . . , h 1 ,
a k l , h 1 = h 1 ( X k ) h 1 ( X l ) p , a ¯ k . , h 1 = 1 n l = 1 n a k l , h 1 ,
a ¯ . l , h 1 = 1 n k = 1 n a k l , h 1 , a ¯ . . , h 1 = 1 n 2 k , l = 1 n a k l , h 1 ,
k , l = 1 , . . . , n
Similarly,
B k l , h 2 = b k l , h 2 b ¯ k . , h 2 b ¯ . l , h 2 + b ¯ . . , h 2 ,
b k l , h 2 = h 2 ( Y k ) h 2 ( Y l ) q , b ¯ k . , h 2 = 1 n l = 1 n b k l , h 2 ,
b ¯ . l , h 2 = 1 n k = 1 n b k l , h 2 , b ¯ . . , h 2 = 1 n 2 k , l = 1 n b k l , h 2 ,
k , l = 1 , . . . , n
To quantify the degree of association between X and Y, we discuss a bivariate case of a response variable Y and a predictor X. Regardless of the relationship type between X and Y, we assume there are transforming functions h 1 ( X ) and h 2 ( Y ) that can transform the relationship between X and Y to a linear relation between h 1 ( X ) and h 2 ( Y ) :
h 2 ( Y ) = β 0 + h 1 ( X ) + ϵ
where ϵ has a Gaussian distribution with zero mean and standard deviation σ . We can find h 1 ( X ) and h 2 ( Y ) by minimizing the Sum of Squared Errors (SSE):
i = 1 n ϵ ^ i 2 = i = 1 n ( h 2 ^ ( Y i ) h 1 ^ ( X i ) ) 2
To minimize i = 1 n ϵ ^ i 2 with regard to h 1 ( X ) and h 2 ( Y ) , we use a simplified optimal transformation [18] by an iterative estimation. Let E [ h 2 ( Y ) ] = 0 , V a r [ h 2 ( Y ) ] = E [ ( h 2 ( Y ) ) 2 ] E [ h 2 ( Y ) ] 2 = 1 , and as a result, E [ ( h 2 ( Y ) ) 2 ] = 1 . We start with h 2 ( Y ) = Y | | Y | | . For a given h 2 ( Y ) , to minimize i = 1 n ϵ ^ i 2 , we have:
h 1 ( X ) = E [ h 2 ( Y ) | X ]
and for a given h 1 ( X ) , in a similar way, we have:
h 2 ( Y ) = E [ h 1 ( X ) | Y ] | | E [ h 1 ( X ) | Y ] | |
In each iteration, h 1 ( X ) and h 2 ( Y ) will be estimated, and an iterative optimization continues until the estimate of error i = 1 n ϵ i 2 ( T ) = i = 1 n ( h 2 ( Y i ) ( T ) h 1 ( X i ) ( T ) ) 2 does not decrease in iteration T, where h 2 ( Y ) ( T ) and h 1 ( X ) ( T ) are optimal estimates with regard to unexplained variance.
Estimated transforming functions h 1 ( X ) and h 2 ( Y ) are optimal and linear for a joint normal distribution [19], where marginal distributions of X and Y are normal. If joint distribution of X and Y is not normal, estimated transforming functions h 1 ( X ) and h 2 ( Y ) are not optimal, but they are close to optimal linear transformations [18]. AF has the following properties:
  • Non-negativity, R A F ( X , Y ) 0 .
  • Disappears if and only if the two vectors are not associated, R A F ( X , Y ) = 0 for unassociated X and Y.
  • Symmetry, R A F ( X , Y ) = R A F ( Y , X ) for noiseless vectors X and Y.
  • Triangular inequality, R A F ( X , Y ) R A F ( X , Z ) + R A F ( Z , Y ) .

3. Simulation Models

To test the performance of the proposed AF, we modeled several simulations and computed Pearson’s correlation, Distance correlation, and the proposed AF. Because Distance correlation and the proposed AF take values between zero and one, we calculate the absolute value of Pearson’s correlation to provide a fair comparison between these methods. The aforementioned correlation coefficients are quantified for linear and nonlinear correlations. We have also obtained these correlation coefficients for random relation (no relationship) as follows.

3.1. Linear and Nonlinear Relationships in Noiseless Conditions

We simulated the following relationships:
  • Linear: Y = β 0 + β 1 X , where β 0 is intercept and β 1 is slope.
  • Fourth order polynomial: Y = β 0 + β 1 X 1 + β 2 X 2 + β 3 X 3 + β 4 X 4 where β ’s are coefficients.
  • Exponential: Y = e x p ( λ x ) , where λ is rate.
  • Parabolic: Y = β 0 ( X β 1 ) 2 , where β 0 and β 1 are coefficients.
The simulation steps are summarized below.
  • Let Ω D be a set of D relationship types l 1 to l D . Generate pairwise variables using relationships in the relationship set Ω D so that D different datasets Γ 1 , Γ 2 , . . . , Γ D representing the relationship types l 1 to l D are obtained.
  • For each generated dataset Γ d , compute Pearson’s correlation (absolute value) ρ d , distance correlation R d , and Association Factor R A F d .
    d = 1 , 2 , . . . , D .

3.2. Linear and Nonlinear Relationships in Noisy Conditions

To test the performance of the proposed AF in noisy conditions, we corrupt the true relationships with low, medium, and high noise:
  • Linear: Y = β 0 + β 1 X + ϵ σ .
  • Fourth order polynomial: Y = β 0 + β 1 X 1 + β 2 X 2 + β 3 X 3 + β 4 X 4 + ϵ σ .
  • Exponential: Y = e x p ( λ x ) + ϵ σ .
  • Parabolic: Y = β 0 ( X β 1 ) 2 + ϵ σ .
where ϵ σ is White (Gaussian) noise and noise level is specified by standard deviation of the Gaussian distribution ( σ ). We then quantify the linear and nonlinear correlations in noisy conditions using Pearson’s correlation, Distance correlation, and Association Factor. In the noisy conditions, we calculate the Monte Carlo average of each correlation coefficient over T = 100 instances (trials) of the same noise level.
The simulation steps are summarized below.
  • Let Ω D be a set of D relationship types l 1 to l D . Generate pairwise variables using relationships in the relationship set Ω D so that D different datasets Γ 1 , Γ 2 , . . . , Γ D representing the relationship types l 1 to l D are obtained.
  • Set t = 1 (trial one).
  • Generate noisy relationships Φ t 1 , Φ t 2 , . . . , Φ t D by adding Gaussian noise (with noise level ϵ σ ) to the datasets Γ d ’s generated using the true relationships ( l d ’s).
  • Compute and save Pearson’s correlation (absolute value) ρ t d , distance correlation R t d , and association factor R A F , t d for each noisy dataset Φ t d .
  • Increase t by one ( t = t + 1 ).
  • Repeat Steps 3 to 5 while t T .
  • Compute the Monte Carlo average of each correlation measure as follows: ρ d = t = 1 T ρ t d T , R d = t = 1 T R t d T , R A F d = t = 1 T R A F , t d T .

3.3. No Relationship

We also investigated whether functions h 1 and h 2 may introduce a spurious relationship into the relationship between X and Y. To address this, we obtained Pearson’s correlation, Distance correlation, and AF for no relationship (random noise).

3.4. Symmetry Regarding Sample Size, Missing Data, and Noise Level

Next, we study the symmetry of AF regarding the response and factor. The goal here is to investigate whether the Association Factor quantifies the relationship between X and Y regardless of their order. This means whether R A F 2 ( X , Y ) is equal to R A F 2 ( Y , X ) . For the true relationship without noise, we calculate R A F 2 ( X , Y ) assuming:
h 2 ( Y ) = β 0 , X + h 1 ( X )
and to compute R A F 2 ( Y , X ) , we have:
h 2 ( X ) = β 0 , Y + h 1 ( Y )
For the noisy relationship, to compute R A F 2 ( X , Y ) , we assume:
h 2 ( Y ) = β 0 , X + h 1 ( X ) + ϵ
and similarly for R A F 2 ( Y , X ) , we have:
h 2 ( X ) = β 0 , Y + h 1 ( Y ) + ϵ
We study the symmetry of AF with regard to the sample size and noise level for nonlinear relationships. We will show that with a small sample size, the underlying relationship cannot be visually identified even in the noiseless case due to the missing data.

3.5. Entropic Distance

We also compute Entropic Distance (ED) and compare it with AF. Entropic Distance, also called “relative entropy”, is the differences between entropies with and without a prior condition [20]. The conditional entropy of two variables X and Y taking values x and y, respectively, is defined by:
R E D ( X | Y ) = y p ( y ) x p ( x | y ) l o g b p ( x | y ) ,
where b is the logarithm base. ED has the following properties [21]:
  • ED is symmetric.
  • ED is zero for comparing a distribution with itself.
  • ED is positive for two different distributions.
AF values are bounded between zero and one, but ED does not have an upper bound and can take any positive value. Therefore, the interpretation of ED is subjective, while AF can objectively represent the strength of the underlying relationship. Therefore, rather than comparing the ED and AF values, we computed the AF ratio and the ED ratio for different noise conditions. Let R A F L 2 and R A F H 2 be the AF for a relationship corrupted with different noise levels. The AF ratio, I A F , is computed by:
I A F = R A F H 2 R A F L 2 * 100
Hence, the AF ratio can be interpreted as:
  • I A F < 100 % indicates a decrease in AF.
  • I A F > 100 % indicates an increase in AF.
  • I A F = 100 % indicates no change in AF.

3.6. Detrended Fluctuation Analysis (DFA)

Peng et al. introduced Detrended Fluctuation Analysis (DFA), which is commonly used in time series analysis and stochastic processes [22]. It is an alternative method in comparison with the auto-correlation function and is often used for determining the statistical self-similarity of a signal. It can detect long-range correlations in a patchy signal. The computation of DFA [22,23] is summarized below.
For a time series of total length N:
  • Integrate the time series:
    y ( k ) = i = 1 k [ B i B a v e ]
    where B i is the i th interval and B a v g is the average interval.
  • Divide the integrated time series into boxes of equal length n.
  • Fit a line to the data in each box of size n separately. The y coordinate of the straight line segment in a box is denoted by y n ( k ) .
  • Remove the trend (detrend) from the integrated time series y ( k ) by subtracting the local trend y n ( k ) in each box.
  • Calculate the root-mean-squared fluctuation, F ( n ) , of the obtained detrended time series by:
    F ( n ) = 1 N k = 1 N [ y ( k ) y n ( k ) ] 2
  • Repeat this computation over all time scales (box size n) to provide a relationship between F ( n ) and the box size (n).

4. Simulation Results and Discussion

We compared the performance of Pearson’s correlation, Distance correlation, and the Association Factor for the following relationships that were explained in detail in the previous section:
  • Linear: Y = 1 + X + ϵ σ .
  • Fourth order polynomial: Y = X 4 + ϵ σ .
  • Exponential: Y = e x p ( 0 . 05 x ) + ϵ σ .
  • Parabolic: Y = 4 ( X 0.5 ) 2 + ϵ σ .
Noiseless linear and nonlinear relationships are depicted in Figure 1. The noisy relationships with low, moderate, and high noise are shown in Figure 2, Figure 3 and Figure 4, respectively. The performance of Pearson’s correlation, Distance correlation, and Association Factor in identifying these linear and nonlinear relationships are depicted in Figure 5, Figure 6, Figure 7 and Figure 8. The performance of these correlation factors at different noise levels are discussed in the following section.

4.1. True Signal (No Noise)

Noiseless linear, exponential, parabolic, and fourth order polynomial are shown in Figure 1. Quantified correlations by Pearson’s correlation, Distance correlation, and Association Factor are summarized in Table 1. As we expect, Pearson’s correlation obtained a value of one for noiseless linear relationship, but its value was not reliable for nonlinear relationships such as exponential and polynomial. Distance correlation identified both linear and nonlinear relationships, but as we can see in Table 1, its performance was not robust with regard to the underlying relationship type between two variables. It scored one for a noiseless linear relationship, while it scored 0.47 for the fourth order polynomial, 0.91 for exponential, and 0.5 for parabolic. In contrast, the proposed AF could robustly identify the underlying relationship, and its value was one regardless of the relationship type (linear, exponential, or polynomial).

4.2. Noisy Relationships

Linear, exponential, parabolic, and fourth order polynomial relationships corrupted with low, moderate, and high noise are shown in Figure 2, Figure 3 and Figure 4 respectively. Pearson’s correlation, Distance correlation, and Association Factor are summarized for low, moderate, and high noise in Table 2, Table 3 and Table 4 respectively. The Pearson’s correlation absolute value dropped from one (noiseless) to 0.98, 0.82, and 0.58 for low, moderate, and high noise in identifying linear relationship between two variables. We can observe that its value is not reliable for nonlinear relationships. Its absolute value dropped from 0.86 (noiseless) to 0.58 (high noise) for exponential relationship. For fourth order polynomial, its absolute value increased from noiseless (0.06) to low noise (0.23) and then dropped from low noise to high noise (0.14). For parabolic relationship, the Pearson’s correlation absolute value increased from noiseless (0.05) to low noise (0.21) and then dropped from low noise to high noise (0.16).
Distance correlation had steady performance for linear relationship, and its value decreased from one (noiseless) to 0.56 (high noise). However, its performance for nonlinear relationships was not consistent. Its score in identifying exponential relationship was comparable with its score for linear relationship. Its value for exponential relationship was 0.98 (for noiseless) and decreased to 0.56 (for high noise). Its score for parabolic relationship was 0.50 (for noiseless) and dropped to 0.37 (for high noise). In identifying the fourth order polynomial, Distance correlation scored 0.47 for noiseless and decreased to 0.29 for high noise.
In contrast, as we can see, the proposed AF had robust performance regardless of the relationship type (Figure 5, Figure 6, Figure 7 and Figure 8). Moreover, it had robust performance in noiseless and noisy conditions. Its value for noiseless relationships (linear, exponential, parabolic, and fourth order polynomial) was steady and equal to one. Its value in low noise was still about one (0.99) regardless of the relationship type. In moderate noise condition, AF consistently identified the underlying relationship with scores from 0.82 (linear) to 0.95 (parabolic). Even in high noise condition where the underlying relationship was substantially corrupted with noise (Figure 4), AF was able to identify the underlying correlations with scores from 0.58 (linear) to 0.72 (parabolic).
AF had comparable performance with Pearson’s correlation in identifying linear relationship. It outperformed Distance correlation in identifying noiseless nonlinear relationships. Moreover, it outperformed Distance correlation in identifying nonlinear relationships in noisy conditions. Its score was up to twice (0.69) as high as Distance correlation (0.29) in identifying nonlinear correlations in high noise.

4.3. No Relationship

We computed Pearson’s correlation, Distance correlation, and AF for no relationship (random noise). The results are summarized in Table 5. As we can see, all correlation factors including Pearson’s correlation, Distance correlation, and AF obtained values close to zero, indicating there was no relationship between X and Y. This also clarifies that functions h 1 and h 2 did not introduce a spurious relationship into the relationship between X and Y.

4.4. Test of Symmetry, Sample Size, Missing Data, and Noise Level

To study the symmetry of AF quantifying the relationship between response Y and factor X regardless of their order, we computed R A F 2 ( X , Y ) and R A F 2 ( Y , X ) and compared them. We computed AF with regard to sample size and noise level for different relationships. Figure 9 from top to bottom shows randomly sampled true (noiseless) circular relationship of size 100, 50, and 30, respectively. The second and third columns of Figure 9 show h 1 and h 2 obtained by R A F 2 ( X , Y ) and R A F 2 ( Y , X ) respectively. As we can observe, the transform functions regardless of the order of the response and factor were symmetric and linear even for small sample size.
The first and second row of Figure 10 show two instances of the randomly sampled true (noiseless) circular relationship of size 10. The third and fourth row of Figure 10 show two instances of the randomly sampled true (noiseless) circular relationship of size 30. As we can observe in this figure, because of small sample size, the true underlying relationship is not visible due to the missing data points. The second and third columns of Figure 10 show h 1 and h 2 obtained by R A F 2 ( X , Y ) and R A F 2 ( Y , X ) respectively. As we can observe, regardless of the order of the response and factor, and despite missing data, transform functions h 1 and h 2 were symmetric and almost linear even for a dataset with very small sample size.
Distance correlation, Maximal Information Coefficient (MIC) [24], R A F 2 ( X , Y ) , and R A F 2 ( Y , X ) are obtained for the randomly sampled true circular relation of different sample sizes and are summarized in Table 6. As we can see, regardless of the sample size, AF could quantify the relationship even for a very small sample size of 10 (with missing data points). Moreover, AF was symmetric even for a small sample size of 30 and was almost symmetric for a very small sample size of 10. AF slightly decreased by reducing the sample size. AF outperformed MIC, and MIC performed better than Distance correlation. MIC also decreased by reducing the sample size. Distance correlation was in a range between 0.22 and 0.25 for sample sizes from 30 to 100. Its performance for sample size of 10 was sporadic. For example it scored 0.56 for a typical example of randomly sampled true circular relationship of size 10 depicted in Figure 10, second row. This could be potentially due to the arrangement of data points in this random sample of a circle that rather represents a linear relationship.
Figure 11 from top to bottom shows a randomly sampled circular relationship of size 10, 30, 50, and 100 respectively corrupted with Gaussian noise. The second and third columns of Figure 11 show h 1 and h 2 obtained by R A F 2 ( X , Y ) and R A F 2 ( Y , X ) respectively. As we can observe regardless of the order of response and factor, the transform functions were symmetric and almost linear even for a very small sample of size 10.
Distance correlation and the Maximal Information Coefficient (MIC), R A F 2 ( X , Y ) , and R A F 2 ( Y , X ) were obtained for the randomly sampled circular relation of different sample sizes corrupted with high noise and are summarized in Table 7. As we can see regardless of the sample size, AF could quantify the relationship even for a very small sample size of 10 (with missing data points). Moreover, AF was symmetric for moderate sample size (from 50 to 100) and was almost symmetric for small and very small sample size of 30 and 10 respectively. Similar to the noiseless scenario, AF slightly decreased by reducing the sample size. AF outperformed MIC, and MIC performed better than Distance correlation. MIC values were in a range from 0.28 to 0.31 for sample size from 30 to 100, but MIC was higher for the noisy relationship with a sample size of 10. The Distance correlation values were in a range from 0.19 to 0.25 for sample size from 30 to 100; however it was 0.36 for a typical random sample of size 10 from circular relationship corrupted with noise (depicted in Figure 11).

4.5. Empirical Distribution of Distance Correlation, Maximal Information Coefficient, R A F 2 ( X , Y ) and R A F 2 ( Y , X )

Next, we investigated the distribution of Distance correlation, Maximal Information Coefficient (MIC), R A F 2 ( X , Y ) , and R A F 2 ( Y , X ) . Figure 12 shows the Monte Carlo empirical distribution of these correlation measures for randomly sampled true circular relationship of size 30. Distributions were estimated by 1000 Monte Carlo samples. Distance correlation had a positively skewed distribution with a mode at about 0.3. MIC had a multimodal distribution with modes at about 0.3, 0.4, 0.5, and 0.6 with the highest mode at about 0.4. AF was negatively skewed with a mode at about 0.9975. We can also observe that R A F 2 ( X , Y ) and R A F 2 ( Y , X ) have similar distributions.
Figure 13 shows the Monte Carlo empirical distribution of these correlation measures for randomly sampled circular relationship of size 30 corrupted with high Gaussian noise. Similar to the previous simulation, distributions were estimated by 1000 Monte Carlo samples. Distance correlation had a positively skewed distribution with a mode at about 0.27. MIC had a multimodal distribution with the highest mode at about 0.25. AF was negatively skewed with a mode at about 0.7. Again here, we can see that R A F 2 ( X , Y ) and R A F 2 ( Y , X ) have similar distributions.
To study AF in a different noise condition, we corrupted the randomly sampled circular distribution with exponential noise and obtained the values of Distance Correlation, Maximal Information Coefficient, R A F 2 ( X , Y ) , and R A F 2 ( Y , X ) . Figure 14 shows the Monte Carlo empirical distribution of these correlation measures for randomly sampled circular relationship of size 30 corrupted with high exponential noise. Distributions were estimated by 1000 Monte Carlo samples. We see again that Distance correlation had a positively skewed distribution with a mode at about 0.27. MIC had a multimodal distribution with the highest mode at about 0.3. AF was negatively skewed with a mode at about 0.8. As we can see, R A F 2 ( X , Y ) and R A F 2 ( Y , X ) have similar distributions.

4.6. Entropic Distance

We compared the performance of Entropic Distance (ED) and the Association Factor (AF) for linear, polynomial, exponential, and parabolic relationships. Their values for noiseless, low, moderate, and high noise are summarized in Table 8. AF performed consistently with a value of one for true relationships regardless of the relationship type. ED ranged from 1.1 to about two for different true relationships. The highest value obtained by ED was for the true linear relationship. In low, moderate, and high noise, the lowest value obtained by ED was for the parabolic relationship (0.848, 0.825, and 0.812, respectively). ED did not have an upper bound and could take any positive value, while the AF values were bounded between zero and one. Therefore, rather than comparing AF and ED, we computed the AF ratio and the ED ratio in different noise conditions.
Table 9 shows the ratios as a percentage for both metrics where the noise level was increased from (a) noiseless to low noise, (b) low noise to moderate noise, and (c) moderate noise to high noise. ED ratios indicated that ED decreased by increasing the noise level. Similarly, the AF ratio decreased by increasing the noise level (Table 9). For polynomial and parabolic relationships, ED had a substantial decrease from noiseless to low noise, and it almost stabilized and had slight changes afterward by increasing the noise level. In contrast, AF had a consistent response to noise, and it decreased gradually, while its values for low noise were almost the same as the noiseless case. In comparison with ED: 1. AF is bounded; 2. AF obtains the same value regardless of the relationship type in noiseless condition; 3. AF can better quantify the correlation in noisy conditions.

4.7. Detrended Fluctuation Analysis

Pearson’s correlation, AF, and Detrended Fluctuation Analysis (DFA) [22] are obtained for different relationships and are summarized in Table 10. The Pearson’s correlation coefficient was computed before and after detrending the data. As we can see in the Table 10, Pearson’s correlation could identify a strong correlation even for nonlinear relationships after detrending the data. Interestingly, the DFA values were almost identical to the Pearson’s correlation values obtained for detrended data. This could be explained by visualizing the detrended data for a nonlinear relationship. As we can see in Figure 15, a polynomial relationship (Figure 15, left) was transformed to an approximately linear relationship after detrending the data (Figure 15, right). Hence, after detrending the data, Pearson’s correlation could detect the nonlinear relation. We can conclude that AF and DFA are both hybrid methods. Both methods, transform the data first, and then quantify the relationship of the transformed data.

5. Conclusions and Future Work

We introduced a new method to identify and quantify correlation between two variables. The proposed coefficient, Association Factor (AF), is a robust method for the identification and quantification of both linear and nonlinear associations. We applied the proposed method to several different relationships including linear, exponential, parabolic, polynomial, and circular. The results demonstrated that AF could identify both linear and nonlinear relationships. Its value was equal to one in noiseless conditions regardless of the relationship type. Moreover, we tested AF in noisy conditions where the true relationships were corrupted with noise. AF could successfully identify the correlations in low, moderate, and high noise conditions. We also tested AF under different noise distributions, Gaussian and exponential. Regardless of the noise distribution, AF could successfully quantify the correlation.
We studied the distribution of AF and compared it with the distributions of Distance correlation and MIC. We also investigated the AF values for a very small sample size where the relationship was severely under-sampled. Despite the fact that a substantial amount of data was missing due to very small sample size, AF still could quantify the underlying correlation. We compared AF with ED and discussed its advantages over ED. AF had similar performance to Pearson’s correlation in identifying linear relationship in noiseless and noisy conditions, and its value was equal to one for the noiseless linear relationship. AF outperformed Distance correlation and MIC in noiseless linear and nonlinear relationships. It also outperformed Distance correlation and MIC in noisy linear and nonlinear relationships. The results demonstrated that AF was robust with regard to the relationship type, as well as the noise condition. Although, we studied the bivariate case in this work, AF could be extended to quantify the relationship between several factors and a response, and our future work is focused on implementing the Multivariate Association Factor (MAF). The potential iterative model for a k x 1 vector of factors X k and response Y can be defined by:
h 1 ^ ( X k ) ( t ) = E ( h 2 ^ ( Y ) ( t 1 ) h 1 ^ ( X k ) ( t 1 ) | X k )
and:
h 2 ^ ( Y ) ( t ) = E ( h 1 ^ ( X k ) ( t 1 ) | Y )
where h 1 ^ ( X k ) ( t ) and h 2 ^ ( Y ) ( t ) are estimates of h 1 ( X k ) and h 2 ( Y ) at iteration t, respectively.

Author Contributions

Formal analysis, N.N.K. and W.D.; Investigation, N.N.K. and W.D.; Methodology, N.N.K.; Project administration, N.N.K.; Visualization, N.N.K. and W.D.; Writing–original draft, N.N.K. and W.D.; Writing–review and editing, N.N.K. and W.D. All authors have read and agreed to the published version of the manuscript.

Funding

The publication of this article was funded by the Deanship of Scientific Research (DSR), King Abdulaziz University, Jeddah, under grant No. (D-466-662-1441). The authors would like to gratefully acknowledge DSR support.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hastie, T.; Tibshirani, R.; Friedman, J. The Elements of Statistical Learning: Data Mining, Inference, and Prediction; Springer Science & Business Media: Berlin/Heidelberger, Germany, 2009. [Google Scholar]
  2. Székely, G.J.; Rizzo, M.L.; Bakirov, N.K. Measuring and testing dependence by correlation of distances. Ann. Stat. 2007, 35, 2769–2794. [Google Scholar] [CrossRef]
  3. Székely, G.J.; Rizzo, M.L. Brownian distance covariance. Ann. Appl. Stat. 2009, 3, 1236–1265. [Google Scholar] [CrossRef] [Green Version]
  4. Dueck, J.; Edelmann, D.; Gneiting, T.; Richards, D. The affinely invariant distance correlation. Bernoulli 2014, 20, 2305–2330. [Google Scholar] [CrossRef] [Green Version]
  5. Póczos, B.; Schneider, J. Conditional Distance Variance and Correlation. Figshare 2012. Available online: https://www.cs.cmu.edu/~bapoczos/articles/poczos12distancecorr.pdf (accessed on 5 April 2020).
  6. Dueck, J.; Edelmann, D.; Richards, D. Distance correlation coefficients for Lancaster distributions. J. Multivar. Anal. 2017, 154, 19–39. [Google Scholar] [CrossRef] [Green Version]
  7. Edelmann, D.; Richards, D.; Vogel, D. The distance standard deviation. arXiv 2017, arXiv:1705.05777. [Google Scholar]
  8. Jentsch, C.; Leucht, A.; Meyer, M.; Beering, C. Empirical Characteristic Functions-Based Estimation and Distance Correlation for Locally Stationary Processes; Technical Report; University of Mannheim: Mannheim, Germany, 2016. [Google Scholar]
  9. Górecki, T.; Krzyśko, M.; Ratajczak, W.; Wołyński, W. An Extension of the Classical Distance Correlation Coefficient for Multivariate Functional Data with Applications. Stat. Transit. New Ser. 2016, 17, 449–466. [Google Scholar] [CrossRef] [Green Version]
  10. Szekely, G.J.; Rizzo, M.L. Partial distance correlation with methods for dissimilarities. Ann. Stat. 2014, 42, 2382–2412. [Google Scholar] [CrossRef]
  11. Székely, G.J.; Rizzo, M.L. The distance correlation t-test of independence in high dimension. J. Multivar. Anal. 2013, 117, 193–213. [Google Scholar] [CrossRef]
  12. Davis, R.A.; Matsui, M.; Mikosch, T.; Wan, P. Applications of distance correlation to time series. Bernoulli 2018, 24, 3087–3116. [Google Scholar] [CrossRef] [Green Version]
  13. Zhou, Z. Measuring nonlinear dependence in time-series, a distance correlation approach. J. Time Ser. Anal. 2012, 33, 438–457. [Google Scholar] [CrossRef]
  14. Bhattacharjee, A. Distance correlation coefficient: An application with bayesian approach in clinical data analysis. J. Mod. Appl. Stat. Methods 2014, 13, 23. [Google Scholar] [CrossRef]
  15. Jaskowiak, P.A.; Campello, R.J.; Costa, I.G. On the selection of appropriate distances for gene expression data clustering. BMC Bioinform. 2014, 15, S2. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Kong, J.; Wang, S.; Wahba, G. Using distance covariance for improved variable selection with application to learning genetic risk models. Stat. Med. 2015, 34, 1708–1720. [Google Scholar] [CrossRef] [Green Version]
  17. Pearson, K. Note on regression and inheritance in the case of two parents. Proc. R. Soc. Lond. 1895, 58, 240–242. [Google Scholar]
  18. Breiman, L.; Friedman, J.H. Estimating optimal transformations for multiple regression and correlation. J. Am. Stat. Assoc. 1985, 80, 580–598. [Google Scholar] [CrossRef]
  19. Lancaster, H.O. Rankings and Preferences: New Results in Weighted Correlation and Weighted Principal Component Analysis With Applications; John Wiley & Sons: Chichester, UK, 1969. [Google Scholar]
  20. Biró, T.S.; Telcs, A.; Néda, Z. Entropic Distance for Nonlinear Master Equation. Universe 2018, 4, 10. [Google Scholar] [CrossRef] [Green Version]
  21. Biró, T.S.; Schram, Z. Non-Extensive Entropic Distance Based on Diffusion: Restrictions on Parameters in Entropy Formulae. Entropy 2016, 18, 42. [Google Scholar] [CrossRef]
  22. Peng, C.K.; Buldyrev, S.V.; Havlin, S.; Simons, M.; Stanley, H.E.; Goldberger, A.L. Mosaic organization of DNA nucleotides. Phys. Rev. E 1994, 49, 1685. [Google Scholar] [CrossRef] [Green Version]
  23. Peng, C.K.; Havlin, S.; Stanley, H.E.; Goldberger, A.L. Quantification of scaling exponents and crossover phenomena in nonstationary heartbeat time series. Chaos Interdiscip. J. Nonlinear Sci. 1995, 5, 82–87. [Google Scholar] [CrossRef]
  24. Reshef, D.N.; Reshef, Y.A.; Finucane, H.K.; Grossman, S.R.; McVean, G.; Turnbaugh, P.J.; Lander, E.S.; Mitzenmacher, M.; Sabeti, P.C. Detecting novel associations in large data sets. Science 2011, 334, 1518–1524. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. True (without noise) linear, polynomial, exponential, and parabolic relationship types.
Figure 1. True (without noise) linear, polynomial, exponential, and parabolic relationship types.
Entropy 22 00440 g001
Figure 2. Linear, polynomial, exponential, and parabolic relationship types corrupted with low noise.
Figure 2. Linear, polynomial, exponential, and parabolic relationship types corrupted with low noise.
Entropy 22 00440 g002
Figure 3. Linear, polynomial, exponential, and parabolic relationship types corrupted with medium noise.
Figure 3. Linear, polynomial, exponential, and parabolic relationship types corrupted with medium noise.
Entropy 22 00440 g003
Figure 4. Linear, polynomial, exponential, and parabolic relationship types corrupted with high noise.
Figure 4. Linear, polynomial, exponential, and parabolic relationship types corrupted with high noise.
Entropy 22 00440 g004
Figure 5. Pearson’s correlation (blue), Distance correlation (orange), and Association Factor (gray) scores for true linear relationship (with no noise), and linear relationship corrupted with low, moderate, and high noise.
Figure 5. Pearson’s correlation (blue), Distance correlation (orange), and Association Factor (gray) scores for true linear relationship (with no noise), and linear relationship corrupted with low, moderate, and high noise.
Entropy 22 00440 g005
Figure 6. Pearson’s correlation (blue), Distance correlation (orange), and Association Factor (gray) scores for true polynomial relationship (with no noise), and polynomial relationship corrupted with low, moderate, and high noise.
Figure 6. Pearson’s correlation (blue), Distance correlation (orange), and Association Factor (gray) scores for true polynomial relationship (with no noise), and polynomial relationship corrupted with low, moderate, and high noise.
Entropy 22 00440 g006
Figure 7. Pearson’s correlation (blue), Distance correlation (orange), and Association Factor (gray) scores for true exponential relationship (with no noise), and exponential relationship corrupted with low, moderate, and high noise.
Figure 7. Pearson’s correlation (blue), Distance correlation (orange), and Association Factor (gray) scores for true exponential relationship (with no noise), and exponential relationship corrupted with low, moderate, and high noise.
Entropy 22 00440 g007
Figure 8. Pearson’s correlation (blue), Distance correlation (orange), and Association Factor (gray) scores for true parabolic relationship (with no noise), and parabolic relationship corrupted with low, moderate, and high noise.
Figure 8. Pearson’s correlation (blue), Distance correlation (orange), and Association Factor (gray) scores for true parabolic relationship (with no noise), and parabolic relationship corrupted with low, moderate, and high noise.
Entropy 22 00440 g008
Figure 9. Randomly sampled true (noiseless) circular relationship. (From top to bottom): sample size of 100, 50, and 30 respectively; (From left to right): sampled true (noiseless) circular relationship; h 1 and h 2 obtained by R A F 2 ( X , Y ) ; h 1 and h 2 obtained by R A F 2 ( Y , X ) .
Figure 9. Randomly sampled true (noiseless) circular relationship. (From top to bottom): sample size of 100, 50, and 30 respectively; (From left to right): sampled true (noiseless) circular relationship; h 1 and h 2 obtained by R A F 2 ( X , Y ) ; h 1 and h 2 obtained by R A F 2 ( Y , X ) .
Entropy 22 00440 g009
Figure 10. Randomly sampled true (noiseless) circular relationship with missing data (small sample size). (First and second row): two instances of the randomly sampled true (noiseless) circular relationship of size 10. (Third and fourth row): two instances of the randomly sampled true (noiseless) circular relationship of size 30. (From left to right): sampled true (noiseless) circular relationship; h 1 and h 2 obtained by R A F 2 ( X , Y ) ; h 1 and h 2 obtained by R A F 2 ( Y , X ) .
Figure 10. Randomly sampled true (noiseless) circular relationship with missing data (small sample size). (First and second row): two instances of the randomly sampled true (noiseless) circular relationship of size 10. (Third and fourth row): two instances of the randomly sampled true (noiseless) circular relationship of size 30. (From left to right): sampled true (noiseless) circular relationship; h 1 and h 2 obtained by R A F 2 ( X , Y ) ; h 1 and h 2 obtained by R A F 2 ( Y , X ) .
Entropy 22 00440 g010
Figure 11. Randomly sampled circular relationship corrupted with high Gaussian noise. (From top to bottom): sample size of 10, 30, 50, and 100, respectively; (From left to right): sampled noisy circular relationship; h 1 and h 2 obtained by R A F 2 ( X , Y ) ; h 1 and h 2 obtained by R A F 2 ( Y , X ) .
Figure 11. Randomly sampled circular relationship corrupted with high Gaussian noise. (From top to bottom): sample size of 10, 30, 50, and 100, respectively; (From left to right): sampled noisy circular relationship; h 1 and h 2 obtained by R A F 2 ( X , Y ) ; h 1 and h 2 obtained by R A F 2 ( Y , X ) .
Entropy 22 00440 g011
Figure 12. Monte Carlo empirical PDF for the true circular relationship obtained using sample size of 30. (Top row): Distance correlation (left); MIC (right). (Bottom row): Association Factor (AF) for response Y (left); AF for response X (right).
Figure 12. Monte Carlo empirical PDF for the true circular relationship obtained using sample size of 30. (Top row): Distance correlation (left); MIC (right). (Bottom row): Association Factor (AF) for response Y (left); AF for response X (right).
Entropy 22 00440 g012
Figure 13. Monte Carlo empirical PDF for the circular relationship obtained using sample size of 30 corrupted with high level of Gaussian noise. (Top row): Distance correlation (left); MIC (right). (Bottom row): AF for response Y (left); AF for response X (right).
Figure 13. Monte Carlo empirical PDF for the circular relationship obtained using sample size of 30 corrupted with high level of Gaussian noise. (Top row): Distance correlation (left); MIC (right). (Bottom row): AF for response Y (left); AF for response X (right).
Entropy 22 00440 g013
Figure 14. Monte Carlo empirical PDF for the circular relationship obtained using sample size of 30 corrupted with high exponential noise. (Top row): Distance correlation (left); MIC (right). (Bottom row): AF for response Y (left); AF for response X (right).
Figure 14. Monte Carlo empirical PDF for the circular relationship obtained using sample size of 30 corrupted with high exponential noise. (Top row): Distance correlation (left); MIC (right). (Bottom row): AF for response Y (left); AF for response X (right).
Entropy 22 00440 g014
Figure 15. Polynomial relationship (left) and corresponding detrended data (right).
Figure 15. Polynomial relationship (left) and corresponding detrended data (right).
Entropy 22 00440 g015
Table 1. Pearson’s correlation, Distance correlation, and Association Factor for true relationships (without noise).
Table 1. Pearson’s correlation, Distance correlation, and Association Factor for true relationships (without noise).
Relationship
Type
Pearson’s
Correlation
Distance
Correlation
Association
Factor
Linear1.001.001.00
Polynomial0.060.471.00
Exponential0.860.911.00
Parabolic0.050.501.00
Table 2. Pearson’s correlation, Distance correlation, and Association Factor for relationships with low noise.
Table 2. Pearson’s correlation, Distance correlation, and Association Factor for relationships with low noise.
Relationship
Type
Pearson’s
Correlation
Distance
Correlation
Association
Factor
Linear0.980.980.98
Polynomial−0.230.480.99
Exponential0.850.890.99
Parabolic−0.210.510.99
Table 3. Pearson’s correlation, Distance correlation, and Association Factor for relationships with moderate noise.
Table 3. Pearson’s correlation, Distance correlation, and Association Factor for relationships with moderate noise.
Relationship
Type
Pearson’s
Correlation
Distance
Correlation
Association
Factor
Linear0.820.810.82
Polynomial−0.200.390.83
Exponential0.750.760.90
Parabolic−0.210.490.95
Table 4. Pearson’s correlation, Distance correlation, and Association Factor for relationships with high noise.
Table 4. Pearson’s correlation, Distance correlation, and Association Factor for relationships with high noise.
Relationship
Type
Pearson’s
Correlation
Distance
Correlation
Association
Factor
Linear0.580.560.58
Polynomial−0.140.290.59
Exponential0.580.560.69
Parabolic−0.160.370.72
Table 5. Pearson’s correlation, Distance correlation, and Association Factor for no relationship.
Table 5. Pearson’s correlation, Distance correlation, and Association Factor for no relationship.
Relationship
Type
Pearson’s
Correlation
Distance
Correlation
Association
Factor
No Relationship0.030.060.07
Table 6. Distance correlation, Maximal Information Coefficient (MIC), Association Factor: R A F 2 ( X , Y ) and R A F 2 ( Y , X ) for the true circular relationship.
Table 6. Distance correlation, Maximal Information Coefficient (MIC), Association Factor: R A F 2 ( X , Y ) and R A F 2 ( Y , X ) for the true circular relationship.
Sample SizeDistance
Correlation
MICAF(X,Y)AF(Y,X)
1000.2210.5630.9990.999
500.2190.4840.9930.999
300.2460.4900.9950.994
100.5590.3960.9380.968
Table 7. Distance correlation, MIC, Association Factor: R A F 2 ( X , Y ) and R A F 2 ( Y , X ) for the noisy circular relationship.
Table 7. Distance correlation, MIC, Association Factor: R A F 2 ( X , Y ) and R A F 2 ( Y , X ) for the noisy circular relationship.
Sample SizeDistance
Correlation
MICAF(X,Y)AF(Y,X)
1000.2050.2820.5790.579
500.1860.3140.6660.667
300.2490.3040.5730.536
100.3590.3960.5420.534
Table 8. Entropic Distance and Association Factor for true relationships with no noise and relationships with low, moderate, and high noise.
Table 8. Entropic Distance and Association Factor for true relationships with no noise and relationships with low, moderate, and high noise.
No NoiseLow NoiseModerate NoiseHigh Noise
Relationship
Type
Entropic
Distance
Association
Factor
Entropic
Distance
Association
Factor
Entropic
Distance
Association
Factor
Entropic
Distance
Association
Factor
Linear1.9891.0001.9930.9801.7960.8201.5390.580
Polynomial1.2031.0001.0000.9900.9530.8300.9310.590
Exponential1.8451.0001.8470.9901.7490.9001.5120.690
Parabolic1.1061.0000.8480.9900.8250.9500.8120.720
Table 9. Changes of Entropic Distance and Association Factor in response to change of noise level.
Table 9. Changes of Entropic Distance and Association Factor in response to change of noise level.
Noiseless to
Low Noise
Low Noise to
Moderate Noise
Moderate Noise to
High Noise
Relationship
Types
Entropic
Distance
Association
Factor
Entropic
Distance
Association
Entropic
Distance
Association
Factor
Linear100%98%90%84%86%71%
Polynomial83%99%95%84%98%71%
Exponential100%99%95%91%86%77%
Parabolic77%99%97%96%98%76%
Table 10. Pearson’s correlation obtained for the original and detrended data along with DFA.
Table 10. Pearson’s correlation obtained for the original and detrended data along with DFA.
Relationship
Type
Pearson’s Correlation
Original Data
Pearson’s Correlation
Detrended Data
DFA
Linear111
Polynomial0.060.770.76
Exponential0.860.990.99
Parabolic0.050.880.88

Share and Cite

MDPI and ACS Style

Kachouie, N.N.; Deebani, W. Association Factor for Identifying Linear and Nonlinear Correlations in Noisy Conditions. Entropy 2020, 22, 440. https://doi.org/10.3390/e22040440

AMA Style

Kachouie NN, Deebani W. Association Factor for Identifying Linear and Nonlinear Correlations in Noisy Conditions. Entropy. 2020; 22(4):440. https://doi.org/10.3390/e22040440

Chicago/Turabian Style

Kachouie, Nezamoddin N., and Wejdan Deebani. 2020. "Association Factor for Identifying Linear and Nonlinear Correlations in Noisy Conditions" Entropy 22, no. 4: 440. https://doi.org/10.3390/e22040440

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop