Next Article in Journal
Use of Entropy in Developing SDG-based Indices for Assessing Regional Sustainable Development: A Provincial Case Study of China
Next Article in Special Issue
Statistical Characteristics of Stationary Flow of Substance in a Network Channel Containing Arbitrary Number of Arms
Previous Article in Journal
The Fisher–Rao Distance between Multivariate Normal Distributions: Special Cases, Bounds and Applications
Previous Article in Special Issue
Applications of Information Theory Methods for Evolutionary Optimization of Chemical Computers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Changes of Conformation in Albumin with Temperature by Molecular Dynamics Simulations

by
Piotr Weber
1,†,
Piotr Bełdowski
2,*,†,
Krzysztof Domino
3,†,
Damian Ledziński
4,† and
Adam Gadomski
2,†
1
Atomic and Optical Physics Division, Department of Atomic, Molecular and Optical Physics, Faculty of Applied Physics and Mathematics, Gdańsk University of Technology, ul. G. Narutowicza 11/12, 80-233 Gdańsk, Poland
2
Institute of Mathematics and Physics, UTP University of Science and Technology, Kaliskiego 7, 85-796 Bydgoszcz, Poland
3
Institute of Theoretical and Applied Informatics, Polish Academy of Sciences, Bałtycka 5, 44-100 Gliwice, Poland
4
Faculty of Telecommunications, Computer Science and Technology, UTP University of Science and Technology, 85-796 Bydgoszcz, Poland
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Entropy 2020, 22(4), 405; https://doi.org/10.3390/e22040405
Submission received: 16 January 2020 / Revised: 28 March 2020 / Accepted: 29 March 2020 / Published: 1 April 2020

Abstract

:
This work presents the analysis of the conformation of albumin in the temperature range of 300 K 312 K , i.e., in the physiological range. Using molecular dynamics simulations, we calculate values of the backbone and dihedral angles for this molecule. We analyze the global dynamic properties of albumin treated as a chain. In this range of temperature, we study parameters of the molecule and the conformational entropy derived from two angles that reflect global dynamics in the conformational space. A thorough rationalization, based on the scaling theory, for the subdiffusion Flory–De Gennes type exponent of 0.4 unfolds in conjunction with picking up the most appreciable fluctuations of the corresponding statistical-test parameter. These fluctuations coincide adequately with entropy fluctuations, namely the oscillations out of thermodynamic equilibrium. Using Fisher’s test, we investigate the conformational entropy over time and suggest its oscillatory properties in the corresponding time domain. Using the Kruscal–Wallis test, we also analyze differences between the root mean square displacement of a molecule at various temperatures. Here we show that its values in the range of 306 K 309 K are different than in another temperature. Using the Kullback–Leibler theory, we investigate differences between the distribution of the root mean square displacement for each temperature and time window.

1. Introduction

The dynamics of proteins in solution can be modeled by a complex physical system [1], yielding many interesting effects such as the competition and cooperation between elasticity and phenomena caused by swelling. Biopolymers, such as proteins, consist of monomers (amino acids) that are connected linearly. Thus, a chain seems to be the most natural model of such polymers. However, depending on the properties of the polymer molecule, modeling it as a chain may give more or less accurate outcomes [2]. The most straightforward approach to the polymer is the freely jointed chain, where each monomer moves independently. The excluded volume chain effect, which prevents a polymer’s segments from overlapping, is one of the reasons why the real polymer chain differs from an ideal chain [3]. The diffusive dynamics of biopolymers are discussed further in [4] and the references therein, and the subdiffusive dynamics are discussed in [5,6]. For a given combination of polymer and solvent, the excluded volume varies with the temperature and can be zero at a specific temperature, named the Flory temperature [7]. In other words, at the Flory temperature, the polymer chain becomes nearly ideal [8]. However, this parameter is global, and a more detailed description of the dynamics is required to perform additional detailed analyses. Our work concentrates on an analysis of albumin (shown in Figure 1) [9]. Albumin plays several crucial roles as the main blood plasma protein (60% of all proteins): controlling pH, inducing oncotic pressure, and transporting fluid. This protein (of nonlinear viscoelastic, with characteristics of rheopexy [10]) plays an essential role in the process of articular cartilage lubrication [11] by synovial fluid (SF). SF is a complex fluid of crucial relevance for lowering the coefficients of friction in articulating joints, thus facilitating their lubrication [12]. From a biological point of view, this fluid is a mixture of water, proteins, lipids, and other biopolymers [13]. As an active component of lubricating mechanisms, albumin is affected by friction-induced increases of temperature in synovial fluid. The temperature inside the articular cartilage is in the range of 306 K 312 K [14], and changes occurring inside the fluid can affect the outcome of albumin’s impact on lubrication [10]. Albumin can help to regulate temperature during fever. Therefore, a similar mechanism could take place during joint friction, increasing SF lubricating properties as an efficient heat removal. We analyze this range of temperature: from 300 K 312 K .
Let us move to the technical introduction of albumin dynamics. Consider a chain that was formed by N + 1 monomers, assuming the following positions, designated by { r 0 , r 1 , , r N } . We denote by τ j = r j r j 1 a vector of a distance between two neighboring monomers. While considering the albumin, we have N + 1 = 579 amino acids. For the length of the distance vector τ j , we take the distance between two α -carbons in two consecutive amino acids, hence we have | τ j | = τ = 3.8 [Å] [15]. Using such a notation, we can define the end-to-end vector R by the following formula:
R = r N r 0 = j = 1 N τ j .
In the case of an ideal chain, the distance vectors (for all Ns) are completely independent of each other. This property is expressed by a lack of correlation between any two different bonds, that is [16]:
τ i · τ j = τ 2 δ i j ,
where δ i j is the Kronecker delta. Equations (1) and (2) imply that a dot self-product of vector R reads:
R · R = R 2 = i = 1 N j = 1 N τ i · τ j = N τ 2 ,
where R is the length of the vector R . The value of R is measured by calculating the root mean square end-to-end distance of a polymer. The R’th value is assumed to depend on the polymer’s length N in the following manner [17]:
R τ N μ ,
where μ is called the size exponent. If a chain is freely joined, it appears that μ = 1 / 2 . Relation (4) suggests we consider the dynamics of the albumin protein in some range of temperature and look for dependence between the temperature and exponent μ . Applying the natural-logarithm operation to (4) yields:
μ log R log τ log N .
In Equation (4) we use a ∼ symbol to emphasize that we use R 2 to obtain R. In the case discussed here, both parameters τ and N (the number of amino acids) are constant. Importantly, using molecular dynamics simulations, one can obtain the root mean square end-to-end distance for various temperatures. The idea of the chain points to its description by backbone angles and dihedral angles. Backbone angles are angles between three α -carbons occurring after each other and signed by ϕ . The second angle, designated by ψ , is a dihedral angle [18]. Using this parametrization, we can obtain an experimental distribution of probabilities. Such distributions have features that may display some time-dependent changes.
These time-dependent changes may have an utterly random course, but sometimes periodic behavior can be seen. This property can be analyzed using a particular statistical test [19]. There is an assumption that the time series sampling is even, and the test treats it as a realization of the stochastic model evolution, which consists of specific terms of the form:
y n = β cos n ω + α + ϵ n ,
where β 0 , n = 0 , 1 , 2 , , N , 0 < ω < π , π < α π is uniformly distributed in ( π ; π ] phase factor and ϵ n is an independent, identically distributed random noise sequence. The null hypothesis is that amplitude β is equal zero: ( H 0 : β = 0 ) against the alternative hypothesis that H 1 : β 0 .
A periodogram of a time series ( y 0 , y 1 , , y N ) is used to test the above mentioned hypothesis and it is defined by the following formula:
I ( ω ) = 1 N | j = 0 N exp ( i ω j ) y j | 2 .
where in general, ω [ 0 , π ] and i is the imaginary unit. For a discrete time series, a discrete value of ω from [ 0 , π ] is being used:
ω k = 2 π k N , k = 0 , 1 , , N ,
which are known as Fourier frequencies.
The periodogram (for a single time series) can contain a peak. The statistical test should give us information about the significance of this selected peak. It is measured by the p-value parameter. The statistic which is used in this test has the form:
g = m a x 1 i q I ( ω i ) i = 1 q I ( ω i ) ,
where: I ( ω i ) is the spectral estimate (the periodogram) evaluated at Fourier frequencies and q = ( N 1 ) / 2 is the entire of the number ( N 1 ) / 2 . The p-value can be calculated according to the formula [20]:
P ( g x ) = j = 1 p 1 j 1 j n 1 j x n 1
where p is the largest integer less than 1 / x . In our analysis, we use this framework to evaluate a p-value for a given time series of entropy. If we get g * from the calculation of g, then Equation (10) gives a p-value for probability that the statistic g is larger than g * .
Proteins can be characterized by various parameters that can be obtained from simulations, such as the earlier mentioned R or μ . Another meaningful parameter is the RMSD (root mean squared displacement), which can be calculated in every moment of the simulation and is defined as:
R M S D = i = 1 n R i · R i n
where: R i is a position of the i-th atom. In this way, one can obtain a series of this parameter as a function of time for the further processing. The drawing procedure is realized at the beginning of the simulation from the area of initial values. The initial condition for the albumin molecule drawing is simulated for a given temperature.
We can then group molecules according to mentioned parameters ( R , μ , RMSD) and obtain five sets for 300 K , 303 K , 306 K , 309 K , and 312 K , and compare these parameters’ values. We decided to use the Kruscal–Wallis statistical test because it does not assume a normal distribution of the data. This nonparametric test indicates the statistical significance of differences in the median between sets. A statistic of this test has the form [21]:
H = 12 N ( N + 1 ) i k r s i 2 n i 3 ( N + 1 ) ,
where r s i is a sum of all rank of the given sample, N represents the number of all data in all samples, and k is the number of the groups (set into temperature). We compare the H value of this statistic to the critical value of the χ 2 distribution. If the statistic crosses this critical value, we use a nonparametric multi-comparison test—the Conover–Iman. The formula which we employ has the form [22]:
| r s i n i r s j n j | > t 1 α / 2 ( S r C ) ( N 1 H ) n i n j ( N k ) ( N 1 )
where S r is the sum of all square of ranks, t 1 α / 2 is a quartile from the Student t distribution on N k degrees of freedom, and C has the value:
C = 0.25 N ( N + 1 ) 2 .
The multi-comparison test gives the possibility of clustering sets of molecules into those between which there is a statistically significant difference and those between which there is no such difference.
Statistical tests describe differences between medians of the considered sets of molecules. However, they are not informative about differences between distributions of probabilities. In our work, we use the Kullback–Leibler divergence, which is used in information theory, to compare two distributions. The Kullback–Leibler divergence between two distributions p and the reference distribution q is defined as:
K L ( p , q ) = i p i log p i q i .
The Kullback–Leibler divergence is also called the relative entropy of p with respect to q. Divergence defined by (15) is not symmetric measure. Therefore, sometimes one can use a symmetric form and obtain the symmetric distance measure [23]:
K L d i s t ( p , q ) = K L ( p , q ) + K L ( q , p ) 2 .
The measure defined by (15) is called relative entropy. It is widely used, especially for classification purposes. In our work, we also use this measure to recognize how a given probability distribution is different from a given reference probability distribution. We can look at this method as a complement to the Kruscall–Wallis test. The Kruscall–Wallis test gives information about the differences between medians, while the Kullback–Leibler divergence gives us information about the differences in distributions.

2. Results

We begin the analysis by examining the global dynamics of molecules at different temperatures using the Flory theory, briefly described in the introduction. To estimate the exponent μ from definition in Formula (5), we collect outcomes from a molecular dynamics simulation in the temperature range of 300 K 312 K . We use a temperature step of 3 K . Such a range of temperature was chosen because we operated within the physiological range of temperatures. During the simulation, we can follow the structural properties of molecules in the considered range of temperatures. In the albumin structure, we can distinguish 22 α -helices, which are presented in Figure 1. Molecules in the given range of parameters display roughly similar structures. Therefore, we chose to study their dynamics in several temperatures. One important property, which we can follow during simulations, is a polymer’s end-to-end distance vector as a function of time. We perform the analysis in two time windows, first, between 0–30 ns and, second, between 70–100 ns. For each time window, we obtained the root mean square end-to-end distance of this quantity, whose values are presented in Table 1:
We can see that the root mean square length of the end-to-end vector for albumin varies with temperature. To obtain statistics, we can perform the Kruscal–Wallis test of this variation. In our work, values of the statistics in the test are treated as continuous parameters. We do not specify a statistical significance threshold. In Table 2, in the second column, we can see the values of statistics calculated according to Formula (12). The H parameters from the Kruscal–Wallis test have small values for this test, so the results tend to support the hypothesis about equal medians.
The Kruscal–Wallis test only gives information about differences between medians of the sets, where each set represents one temperature and elements of the set are calculated as root mean square end-to-end vectors over time. In the analysis, for one temperature, we had the set of root mean square end-to-end vectors, and in the next step, we calculated a numerical normalized histogram. Each element of this set corresponds to the single realization of the experiment. The values of the Kullback–Leibler divergence for mean end-to-end signals in the time window 0–30 ns are presented in Table 3. In the first column is the reference distribution for the temperature range. We can see that the biggest value is for the distribution of the temperature 312 K in the reference distribution 309 K .
We can calculate the Kullback–Leibler distance according to Formula (16). The results are presented in Table 4. This measure indicates that, between distributions, the biggest differences are between the distribution for 309 K and the distribution for 312 K .
We can perform the same analyses for the window 70–100 ns. The obtained values of the Kullback–Leibler divergence, which are presented in Table 5. One can see that the biggest difference is between the distribution for 312 K and the distribution for 303 K . One can see that the maximum has changed and its value is bigger than the maximum for the 0–30 ns window.
For the symmetric case, results are presented in Table 6. We can see that a symmetrical form of the measure is also between the distribution for 312 K and the distribution for 303 K .
Table 7, Table 8, Table 9 and Table 10 give the different statistical approaches, without the averaging over the time step). We present there all end-to-end signals for one temperature, and for each moment of time, we calculate the average only over all simulated molecules. In the next step, we calculate normalized histograms. After obtaining numerical distributions, we calculate the Kullback–Leibler divergence between distributions of mean end-to-end signals. The results for the time window 0–30 ns are presented in Table 7.
We can see that the large value of the Kullback–Leibler divergence appears for the distribution for temperature 306 K , where the distribution for temperature 300 K is a reference distribution. Calculations for the Kullback–Leibler distance for mean end-to-end signals in the time window 0–30 ns are presented in Table 8. The results coincide with the measure of Kullback–Leibler divergence.
Following the copula approach presented in [24], we performed the bivariate histograms of some signals to analyse the type of cross-correlation between signals. For the selected example of the mean end-to-end signals in the 0–30 ns time window see Figure 2, while for the 70–100 ns time window, see Figure 3. Observe that for the 0–30 ns case, we have simultaneous “high” events, which refer to the “upper tail dependency”. After the simulation time has passed, this dependency is diminished. The copula with the “upper tail dependency”, such as the Gumbel one, can provide the proper model of the mean end-to-end signal for the initial simulation time window. For the later simulation time window, we should look for the copula with no “tail dependencies”, such as the Frank or Gaussian one.
A similar calculation has been performed for the next time window 70–100 ns. In Table 9, we collect results for Kullback–Leibler divergence for mean end-to-end signals in the time window 70–100 ns.
Its maximal value is between the distribution for 306 K and the distribution for 303 K as a reference distribution.
In Table 10 we can see that the biggest value of the Kullback–Leibler distance appears between the distributions for 306 K and 309 K . Comparing this result with the previous one, we see that the maximal Kullback–Leibler distance for the window 0–30 ns changes its place.
In Table 7, Table 8, Table 9 and Table 10 the diagonal terms are equal 0, since we are comparing a distribution of end-to-end distributions of raw data for the same time window and temperature. In Table 11 we present selected Kullback–Leibler distances for distributions of raw data end-to-end signals in the time window 70–100 ns and 306 K temperature.
As two different statistical approaches can not unambiguously distinguish between temperatures (or their subsets), we can conclude that dynamics of the system at this temperatures range is roughly similar. This will be approved by the analysis of the the Flory–De Gennes exponent.
Using Equation (5), we estimate values of the Flory–De Gennes exponent for the given protein chain. Results are presented in Table 12. We can see that the values of the Flory–De Gennes exponents do not differ too much within the examined range of temperature, and they all seem to attain a value of ca. 0.4 . The power-law exponents for oligomers span a narrow range of 0.38–0.41, which is close to the value of 0.40 obtained for monomers [25,26,27].
Detailed values for two time windows are presented in Table 13. We can see that from a global point of view and Flory theory, the situation is quite uniform for all initial conditions and is within the considered range of temperature. Thus, in our modeling, we see an effect of elasticity vs. swelling. This competition is preserved for the examined temperature range.
The H parameters from the Kruscal–Wallis test have small values for this test, so the results tend to support the hypothesis about equal medians.
The distribution of backbone dihedral angles carries information of the molecule’s dynamics, as they are tightly connected to its elasticity. Using simulations, we can determine the angle ϕ and the angle ψ (see Introduction). Next, we can determine energies that depend on these angles by employing Formula (17). In Figure 4 and Figure 5, we present the logarithm of the angle energy component for two temperatures. Figure 4 presents values in a period of time 0–10 ns, and Figure 5 presents values in a period of time 90–100 ns. We can see that these values fluctuate around the mean value. Such dynamics are similar for all temperatures, but the mean energy rises slightly with increasing temperature. In Figure 5, the blue line presents a logarithm of the mean value of angle energy and the gray line presents values of the standard deviation. We can see that changes in energy are much more evident.
Different outcomes come from an analysis of the energy component connected with the dihedral angles ψ . See Figure 6 (blue lines) in double logarithmic scale, and note the linear regression lines. We present a period of time from 1 ns to 10 ns.
Simulations also show that for further times in the log-log scale, these values exhibit almost no change with time.
In Figure 7, we present regression parameters from Figure 6.
The dynamics of proteins are governed mainly by non-covalent interactions [28]. Therefore, it would be useful to study other components of energy, such as the Coulomb component and the van der Waals component (see Equation (17) and the discussion following it). The Coulomb component oscillates around the almost constant trend present for every temperature. However, the mean value of these oscillations increases with the temperature. The opposite situation appears for the van der Waals component. Similar to the Coulomb component, for each selected temperature, it follows an almost constant trend. However, when the temperature increases, the mean value of such oscillations decreases. The Coulomb component and the van der Waals component obey simple dynamics. Hence, we can move to a further discussion of more interesting dihedral and angular parts of the energy.
Simulations of albumin dynamics can be used to obtain frequency distributions of angles ϕ and ψ . An exemplary 2 D histogram of such a distribution is presented in Figure 8. We can see that most data are concentrated in a small area of space of angles. Similar behavior can be observed for other temperatures and simulation times. This is because the structure of the investigated protein is quite rigid, and there appears to be only a little angular movement. Nevertheless, we can observe variations of the conformational entropy, estimated with such a histogram (see Figure 9).
According to formulas connecting the probability distribution of angles and entropy [17], we can follow changes of entropy in time and temperature. Values of conformational entropy for various temperatures and with simulation time 0–30 ns are presented in Figure 9, where the blue line represents the course of the mean entropy value over time. The calculations take into account nine simulations, and one dot presents one arithmetic mean of entropy. Here, we can see that during simulations, some oscillations can appear. Therefore, for each temperature, we perform Fisher’s test. Results of this test are presented in Table 14.
A simple comparison of the p-values in Table 14 shows that the lowest value is located around 306 K degrees. According to Formula (6), mean entropy is supposed to be treated as stochastic, i.e., partially deterministic and partially random. In our work, we treat the p-value as a parameter, which gives us information about the tendency on maintaining the truth of the hypothesis: β = 0 . If the p-value is lower than in another case, then it presents a stronger tendency in favor of the alternative hypothesis: β 0 .
Much more detailed calculations of the p-values for this case are presented in Table 15. We can see that at 306 K , there is a local minimum of the mean p-value. It indicates that at 306 K , the tendency towards the alternative hypothesis: β 0 is stronger than it is at 303 K and 309 K . However, for 300 K , the mean p-value parameter is the smallest. The mean p-value parameter is almost the same for 306 K and 312 K . This means that both temperatures exhibit nearly the same tendency for the hypothesis: β 0 . Such regularities can be seen in Figure 9.
The same analyzes can be done on the data presented in the image.
The same analyzes can be done on the data presented in Figure 10. The results of these analyzes, in relation to individual time series of entropy, are presented in the Table 16. For these data, we also observe the local minimum of the p-value parameter for 306 K . We have a maximum at 303 K . Thus, the tendency towards the hypothesis β = 0 is strongest here in comparison to the rest of the analyzed temperatures. This maximum value of the p-value parameter for 303 K is also present for the previous window.
Formula (11) can be calculated for each moment of simulation, so we obtain a time series. To get one parameter for one molecule, we calculate a mean value of this parameter for each molecule. In the next step, we can perform the same statistical considerations to derive values of the H statistic. For the time window 0–30 ns, the value of the Kruscal–Wallis test gives H = 9.41 ; however, for the time window 70–100 ns, the value of this statistic is H = 12.52 . Because we treat the statistic as an ordinary parameter, we do not specify whether results are significantly different. In our approach, we conclude that the tendency for the alternative hypothesis about not equal medians is much higher for the time window 70–100 ns. Since this tendency is more significant in the second case than in the previous one, we decided to perform a multi-comparison test. We chose the Conover–Iman multi-comparison test. We treated t α 1 / 2 as a parameter, which depends on an assumed statistical significance. Therefore in Table 17, we only give the left side of Equation (13).
Here, we have a comparison between medians for different sets. Using values from Table 17 and calculating the right sight of Equation (13) for p = 0.05 , we can state that there are sets that differ from another cluster of sets. The cluster of sets for temperatures 306 K , 309 K , and 312 K differ from the set for temperature 303 K . Unfortunately, the set for temperature 300 K is between these two.
We can compare the above statistical results to the other methods such as the Kullback–Leibler distance, which is presented in the Introduction. In Table 18, we have values of this quantity. We can see that the biggest value is between the distribution for 300 K and the distribution for 306 K as a reference distribution.
The calculation of the Kullback–Leibler distance is presented in Table 19. We can conclude that the 306 K case differs most from all other cases, so the dynamics of the system in the temperature range of 306K–309K appears to differ most from the dynamics of the system in other temperature ranges.
The calculations of Kullback–Leibler divergence and distance for the time window 70–100 ns are presented in Table 20 and Table 21. We can conclude that the biggest value is between the distributions for 306 K and 303 K .
In a further analysis, we compare distributions from signals of the RMSD obtained in the same way as a mean root square of end-to-end signals. Results for the Kullback–Leibler distance are presented in Table 22 and Table 23. We can conclude that for the time window 0–30 ns, the most significant value of differences is for the distribution for 309 K and the distribution for 300 K . For the time window 70–100 ns, the most significant value is for 306 K and 303 K .
In all tables of the Kullback–Leibler divergence measures we can see 0 value. For all these cases we compare two raw data from the same time window in the same temperature. Results for 306 K we present in Table 24. We can see that the distribution of Kullback–Leibler distances are widely dispersed. We can observe a similar dispersion in other temperatures.
Referring to Figure 11, we observe simultaneous small valued events for the 0–30 ns simulation time window. There is no such events for the 70–100 ns simulation time window, see Figure 12. Here, the copula with the “lower tail dependency”, such as the Clayton, can provide the proper model of the RMSD for the 0–30 ns simulation time window.

3. Discussion

Albumin plays an essential role in many biological processes, such as the lubrication of articular cartilage. Hence, knowledge of its dynamics in various physiological conditions can help us to understand better its role in reducing friction. From a general point of view, albumin behaves stable regardless of the temperature. The angle energy of the protein fluctuates similarly regardless of the temperature (see Figure 4 and Figure 5). However, its mean value rises monotonically with temperature. On the other hand, the size exponent μ fluctuates with temperature. We calculate, using the Kruscal–Wallis test, that these fluctuations generate small values of H parameter and, as a result, indicate a hypothesis about the equality of medians.
Further, referring to Figure 6 and Figure 7, we can see that the dihedral energy seems to obey (at least at some range of simulation times) the power law-like relation, although the scaling exponent is relatively small. In Figure 7, we can see this relation in the range of temperature from 306K–309K. This exponent does not change in time t too much. It suggests stable behavior. We consider the conformational entropy of the system. For its plot versus simulation time see Figure 9 and Figure 10. On each graph, we can observe a pattern which seems to attain an oscillatory behavior. Fisher’s test allows describing this property by providing the p-value parameter. If this parameter, for some time series of entropy, is relatively small in contrast to p-value parameters for another time series of entropy, then the statistical hypothesis H 0 : β = 0 is less probable than in another case. In Table 14, we can observe that in the range of temperature 303K–309K, there is a local minimum of p-values. A similar effect can be observed in Table 15 when we calculate the mean p-value for each column. We can also observe that there is a local minimum of around 306 K . A smaller mean p-value is seen for 300 K and 312 K . This means that both temperatures exhibit the analogical tendency to favor the hypothesis β 0 . We can also see this in Figure 9, where for temperatures 300 K , 306 K , and 312 K , the regularity of periodicity increases more than at other temperatures. Other tests, including the Kruscal–Wallis test followed by the Conover–Iman multi-comparison test, Kullback–Leibler distance, and Kullback–Leibler divergence give consistent results, while referring to the mean RMSD of the protein of interest.
This globular protein is soluble in water and can bind both cations and anions. By analyzing Figure 1 and Figure 8, it is clear that albumin’s secondary structure is mainly standard α -helix. Charged amino acids (AA) have a large share in albumin’s structure, especially Aspartic and Glutamic acid (25% of all AA) and hydrophobic Leucine and Isoleucine (16%). This composition enables it to preserve its conformation, due to intramolecular interactions as well as interactions with the solvent. Two main factors are at play here, namely hydrophobic contacts and hydrogen bonds (which also occur between protein and water). A large number of charged AA result in a considerable number of inter- and intramolecular hydrogen bonds. On the other hand, hydrophobic AA result in a more substantial impact when the hydrophobic effect stabilizes a protein’s core. As shown by Rezaei-Tavirani et al. [29], the increase of temperature in the physiological range of temperatures results in conformational changes in albumin, which cause more positively charged molecules to be exposed. This, in turn, results in a lower concentration of cations near a molecule. Albumin is known to largely contribute to the osmotic pressure of plasma, where the presence of ions largely influences this property. An increase of temperature results in a reduction of the osmotic pressure of blood, which in feverish conditions can lead to a higher concentration of urine. Because water is the main component of blood and it has a high heat capacity, the increase of urine concentration results in better removal of heat from the body [29]. The difference in dynamics between temperatures shown in the present study could indicate albumin’s binding affinity with other SF components and thus changes its role in biolubrication toward anti-inflamatory. However, due to the high complexity of the fluid, more research has to be performed.
These chemical properties result in an effect of elasticity vs. swelling competition for albumin chains immersed in water. If the elasticity of the chain fully dominates over albumin’s swelling-induced counterpart (interactions of polymer beads and water with network/bond creation), the exponent would have a value of 1 / 2 . If, in turn, a reverse effect applied, the value of the exponent should approach the value of 1 / 4 , [30] (note the two first Eqs only). As a consequence, and what has not been discovered by any other study, the exponent obtained from the simulation in the examined range of temperature looks as if it is near an arithmetic mean of the two exponents mentioned, namely: μ = 1 / 2 ( 2 / 4 + 1 / 4 ) = 3 / 8 , thus pointing to 0.375 0.4 . The difference of about 0.025 very likely comes from the fact that the elastic effect on albumin shows up to win over the swelling-assisted counterterm(s), cf. the Hamiltonian used for the simulations. The exponent μ = 3 / 8 is for itself called the De Gennes exponent, and it is reminiscent of the De Gennes dimension-dependent gelation exponent 2 / ( d + 2 ) , see [30] (eq 15 therein; d = 3 -case). It suggests that albumin’s elastic effect, centering at about 0.4 , is fairly well supported by the internal network and thus supports the creation of bonds, which has been revealed by the present study. Of course, the overall framework is well-substantiated in terms of the scaling concept [3,7]. Regular oscillations of entropy suggest that the system oscillates around the equilibrium state. Furthermore, the Flory–De Gennes exponent appears to be unchanged for both observation windows. Therefore, we expect the system to be near the equilibrium state. We expect the dynamic of the system, for longer simulation times, to be similar to the presented one. The AMBER force field has been reported on overstabilizing helices in proteins. Therefore our next work will report on the effect of force field and water model used.

4. Materials and Methods

The structure of human albumin serum (code 1e78) has been downloaded from a protein data bank (https://www.rcsb.org/structure/1E78) as a starting point to simulations. We use the YASARA Structure Software (Vienna, Austria) [31] to perform MD simulations. Besides this, a three-site model (TIP3P) of water was used [32]. All-atom simulations were performed under the following conditions: temperature 310 K , p H = 7.0 and in 0.9 % NaCl aqueous solution, with a time step of 2 f s . Berendsen barostat and thermostat with a relaxation time of 1 f s were used to maintain constant temperature and pressure. To minimize the inuence of rescaling, YASARA does not use the strongly fluctuating instantaneous microscopic temperature to rescale velocities at each simulation step (classic Berendsen thermostat). Instead, the scaling factor is calculated according to Berendsen’s formula from the time average temperature. The charge was –15. Simulations were carried for 100 ns, due to the fact that throughout this time very small fluctuations of entropy measured could be seen. All-atom molecular dynamics simulations were performed using the AMBER03 force field [33]. The AMBER03 potential function describing interactions among particles takes into account electrostatic, van der Waals, bond, bond angle, and dihedral terms:
E t o t a l = b o n d s k b r r e q 2 + a n g l e k ϕ ϕ ϕ e q 2               + d i h e d r a l s V n 2 [ 1 + cos n ψ γ ] + i < j A i j r i j 12 B i j r i j 6 + q i q j ϵ r i j
where, k b and k ϕ are the force constants for the bond and bond angles, respectively; r and ϕ are bond length and bond angle; r e q and ϕ e q are the equilibrium bond length and bond angle; ψ is the dihedral angle and V n is the corresponding force constant; the phase angle γ takes values of either 0 or 180 . The non-bonded part of the potential is represented by van der Waals ( A i j ) and London dispersion terms ( B i j ) and interactions between partial atomic charges ( q i and q j ). ϵ is the dielectric constant.
For each spherical angle distribution, we performed a normalized histogram consisting of 50 bins. Such a bin number is a compromise between a resolution and smoothness of histograms. Taking the empirical probability p i of data being in i-th bin, we estimate the entropy [18,34]:
S = R 0 i p i log ( p i ) ,
where we use the gas constant R 0 = 8.314 J K · m o l . The sum goes over the discretized Ramachandran space. Obviously, since entropy is an information measure, it depends on a bin size, hence we have here only an estimation applicable for entropy comparisons for different temperatures and simulation times, since bin sizes are constant for all simulations and temperatures. Numerical data analyzes were performed in the Scilab development environment [35].

5. Conclusions

We present several scenarios of an analysis of simulations of the albumin dynamics. The albumin is the linear protein, which makes simulations and analysis straightforward. In particular, we get the end-to-end vector values, and we present that changes for exponent do not vary significantly. The more detailed analysis can be performed, however, using the entropy that appears to oscillate in the function of the simulation time. In general, these oscillations are regular, which is approved by employing the statistical test.
However, our findings show that near the 306 K temperature, we can observe a local minimum of the p-value of this test. Interestingly, this minimum changes its position near the 312 K temperature for the 70–100 ns time window. We also consider the RMSD parameter, which for the second time window, exhibits significant differences between the median of the sets defined for temperatures—we observe clustering.
When summarizing, let us note that the elastic energy terms included in the Hamiltonian (Equation (17)) provide a particularly robust bond vs. angle contribution to the elastic energy of the biopolymer, especially when arguing its role in the angles’ domain. The remaining non-elastic terms in Equation (17), in turn, contribute, in general, to the swelling assisted co-effect. Thus, the already mentioned Van der Waals and electrostatic terms are responsible for the overall attraction–repulsion behavior of the network-like swollen albumin, whereas the last “liquid–crystalline” type of energetic inclusion to Equation (13) samples the dihedral angles space, an observation already discussed above.

Author Contributions

Conceptualization, P.W., P.B., K.D., and A.G.; investigation, P.W., P.B., K.D., and D.L.; writing—original draft preparation, P.W. and P.B.; writing—review and editing, P.B., K.D., D.L. and A.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by UTP University of Science and Technology, Poland, grant BN-10/19 (P.B. and A.G.).

Acknowledgments

We wish to thank Maryellen Zbrozek, E.T.A. at UTP (Fulbright fellowship) for her kind help in assessing the final form of our paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Grimaldo, M.; Roosen-Runge, F.; Zhang, F.; Schreiber, F.; Seydel, T. Dynamics of proteins in solution. Quarter. Rev. Biophys. 2019, 52, e7. [Google Scholar] [CrossRef] [Green Version]
  2. Rubinstein, M.; Colby, R.H. Polymer Physics; Oxford University Press: Oxford, UK, 2003. [Google Scholar]
  3. De Gennes, P.G. Scaling Concepts in Polymer Physics; Cornell University Press: Ithaca, NY, USA, 1979. [Google Scholar]
  4. Gadomski, A. On (sub) mesoscopic scale peculiarities of diffusion driven growth in an active matter confined space, and related (bio) material realizations. Biosystems 2019, 176, 56–58. [Google Scholar] [CrossRef] [PubMed]
  5. Metzler, R.; Jeon, J.H.; Cherstvy, A.G.; Barkai, E. Anomalous diffusion models and their properties: Non-stationarity, non-ergodicity, and ageing at the centenary of single particle tracking. Phys. Chem. Chem. Phys. 2014, 16, 24128. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Weber, P.; Bełdowski, P.; Bier, M.; Gadomski, A. Entropy Production Associated with Aggregation into Granules in a Subdiffusive Environment. Entropy 2018, 20, 651. [Google Scholar] [CrossRef] [Green Version]
  7. Flory, P. Principles of Polymer Chemistry; Cornell University Press: Ithaca, NY, USA, 1953. [Google Scholar]
  8. Doi, M.; Edwards, S.F. The Theory of The Polymer Dynamics. In The International Series of Monographs on Physics; Oxford University Press: New York, NY, USA, 1986; pp. 24–32. [Google Scholar]
  9. Bhattacharya, A.A.; Curry, S.; Franks, N.P. Binding of the General Anesthetics Propofol and Halothane to Human Serum Albumin. J. Biol. Chem. 2000, 275, 38731–38738. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Oates, K.M.N.; Krause, W.E.; Jones, R.L.; Colby, R.H. Rheopexy of synovial fluid and protein aggregation. J. R. Soc. Interface 2006, 3, 167–174. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Rebenda, D.; Čípek, P.; Nečas, D.; Vrbka, M.; Hartl, M. Effect of hyaluronic acid on friction of articular cartilage. Eng. Mech. 2018, 24, 709–712. [Google Scholar]
  12. Seror, J.; Zhu, L.; Goldberg, R.; Day, A.J.; Klein, J. Supramolecular synergy in the boundary lubrication of synovial joints. Nat. Commun. 2015, 6, 6497. [Google Scholar] [CrossRef]
  13. Katta, J.; Jin, Z.; Ingham, E.; Fisher, J. Biotribology of articular cartilage—A review of the recent advances. Med. Eng. Phys. 2000, 30, 1349–1363. [Google Scholar] [CrossRef]
  14. Moghadam, M.N.; Abdel-Sayed, P.; Camine, V.M.; Pioletti, D.P. Impact of synovial fluid flow on temperature regulation in knee cartilage. J. Biomech. 2015, 48, 370–374. [Google Scholar] [CrossRef] [Green Version]
  15. Liwo, A.; Ołdziej, S.; Kaźmierkiewicz, R.; Groth, M.; Czaplewski, C. Design of a knowledge-based force field for off-latice simulations of protein structure. Acta Biochim. Pol. 1997, 44, 527–548. [Google Scholar] [CrossRef]
  16. Havlin, S.; Ben-Avraham, D. Diffusion in disordered media. Adv. Phys. 2002, 51, 187–292. [Google Scholar] [CrossRef]
  17. Bhattacharjee, S.M.; Giacometti, A.; Maritan, A. Flory theory for polymers. J. Phys. Condens. Matter 2013, 25, 503101. [Google Scholar] [CrossRef] [PubMed]
  18. Baxa, M.C.; Haddadian, E.J.; Jumper, J.M.; Freed, K.F.; Sosnick, T.R. Loss of conformational entropy in protein folding calculated using realistic ensembles and its implications for NMR-based calculations. Proc. Natl. Acad. Sci. USA 2014, 111, 15396–15401. [Google Scholar] [CrossRef] [Green Version]
  19. Ahdesmaki, M.; Lahdesmaki, H.; Yli-Harja, O. Robust Fisher’s test for periodicity detection in noisy biological time series. In Proceedings of the IEEE International Workshop on Genomic Signal Processing and Statistics, Tuusula, Finland, 10–12 June 2007. [Google Scholar]
  20. Wichert, S.; Fokianos, K.; Strimmer, K. Identifying periodically expressed transcripts in microarray time series data. Bioinformatics 2004, 20, 5–20. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Kanji, G.K. 100 Statiatical Tests; SAGE Publications Ltd.: Thousand Oaks, CA, USA, 2006. [Google Scholar]
  22. Conover, W.J. Practical Nonparametric Statistics, 2nd ed.; John Wiley & Sons: New York, NY, USA, 1980. [Google Scholar]
  23. Gupta, A.; Parameswaran, S.; Lee, C.-H. Classification of electroencephalography signals for different mental activities using Kullback-Leibler divergence. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Taipei, Taiwan, 19–24 April 2009. [Google Scholar]
  24. Domino, K.; Błachowicz, T.; Ciupak, M. The use of copula functions for predictive analysis of correlations between extreme storm tides. Physica A 2014, 413, 489–497. [Google Scholar] [CrossRef]
  25. Tanner, J.J. Empirical power laws for the radii of gyration of protein oligomers. Acta Crystallogr. D Struct. Biol. 2016, 72, 1119–1129. [Google Scholar] [CrossRef] [Green Version]
  26. Korasick, D.A.; Tanner, J.J. Determination of protein oligomeric structure from small-angle X-ray scattering. Protein Sci. 2018, 27, 814–824. [Google Scholar] [CrossRef] [Green Version]
  27. De Bruyn, P.; Hadži, S.; Vandervelde, A.; Konijnenberg, A.; Prolič-Kalinšek, M.; Sterckx, Y.G.J.; Sobott, F.; Lah, J.; Van Melderen, L.; Loris, R. Thermodynamic Stability of the Transcription Regulator PaaR2 from Escherichia coli O157:H7. Biophys. J. 2019, 116, 1420–1431. [Google Scholar] [CrossRef]
  28. Durell, S.R.; Ben-Naim, A. Hydrophobic-Hydrophilic Forces in Protein Folding. Biopolymers 2017, 107, e23020. [Google Scholar] [CrossRef]
  29. Rezaei-Tavirani, M.; Moghaddamnia, S.H.; Ranjbar, B.; Amani, M.; Marashi, S.A. Conformational Study of Human Serum Albumin in Pre-denaturation Temperatures by Differential Scanning Calorimetry, Circular Dichroism and UV Spectroscopy. J. Biochem. Mol. Biol. 2006, 39, 530–536. [Google Scholar] [CrossRef] [Green Version]
  30. Isaacson, J.; Lubensky, T.C. Flory exponents for generalized polymer problems. J. Phys. Lett. 1980, 41, L-469–L-471. [Google Scholar] [CrossRef] [Green Version]
  31. Krieger, E.; Vriend, G. New ways to boost molecular dynamics simulations. J. Comput. Chem. 2015, 36, 996–1007. [Google Scholar] [CrossRef] [PubMed]
  32. Mark, P.; Nilsson, L. Structure and Dynamics of the TIP3P, SPC, and SPC/E Water Models at 298 K. J. Phys. Chem. A 2001, 105, 9954–9960. [Google Scholar] [CrossRef]
  33. Duan, Y.; Wu, C.; Chowdhury, S.; Lee, M.C.; Xiong, G.; Zhang, W.; Yang, R.; Cieplak, P.; Luo, R.; Lee, T.; et al. A point-charge force field for molecular mechanics simulations of proteins based on condensed-phase quantum mechanical calculations. J. Comput. Chem. 2003, 24, 1999–2012. [Google Scholar] [CrossRef] [PubMed]
  34. Baruah, A.; Rani, P.; Biswas, P. Conformational Entropy of Intrinsically Disordered Proteins from Amino Acid Triads. Sci. Rep. 2015, 5, 11740. [Google Scholar] [CrossRef] [Green Version]
  35. SciLab. Available online: https://www.scilab.org (accessed on 30 April 2017).
Figure 1. Structure of the albumin protein in a ribbon-like form. The colors on the molecular surface indicate the secondary structure: blue— α -helix, green—turn, yellow—3–10 helix, cyan—coil.
Figure 1. Structure of the albumin protein in a ribbon-like form. The colors on the molecular surface indicate the secondary structure: blue— α -helix, green—turn, yellow—3–10 helix, cyan—coil.
Entropy 22 00405 g001
Figure 2. The bivariate histogram of chosen mean end-to-end signals in the 0–30 ns simulation time window.
Figure 2. The bivariate histogram of chosen mean end-to-end signals in the 0–30 ns simulation time window.
Entropy 22 00405 g002
Figure 3. The bivariate histogram of chosen mean end-to-end signals in the 70–100 ns simulation time window.
Figure 3. The bivariate histogram of chosen mean end-to-end signals in the 70–100 ns simulation time window.
Entropy 22 00405 g003
Figure 4. Logarithms of angle energies vs. the simulation time (0–10 ns), exemplary outcome of T = 300 K and T = 309 K .
Figure 4. Logarithms of angle energies vs. the simulation time (0–10 ns), exemplary outcome of T = 300 K and T = 309 K .
Entropy 22 00405 g004
Figure 5. Logarithms of angle energies vs. the simulation time (90–100 ns), exemplary outcome of T = 300 K and T = 309 K .
Figure 5. Logarithms of angle energies vs. the simulation time (90–100 ns), exemplary outcome of T = 300 K and T = 309 K .
Entropy 22 00405 g005
Figure 6. Dihedral energies for each temperature vs. simulation time. We use the double logarithmic scale. Green lines represent linear regression.
Figure 6. Dihedral energies for each temperature vs. simulation time. We use the double logarithmic scale. Green lines represent linear regression.
Entropy 22 00405 g006
Figure 7. Slope of dihedral angle from regression.
Figure 7. Slope of dihedral angle from regression.
Entropy 22 00405 g007
Figure 8. Numerically obtained distribution of angles ϕ and ψ for T = 300 K and 30 ns of simulation.
Figure 8. Numerically obtained distribution of angles ϕ and ψ for T = 300 K and 30 ns of simulation.
Entropy 22 00405 g008
Figure 9. Value of conformational entropy for various temperatures and 0–30 ns of simulation time.
Figure 9. Value of conformational entropy for various temperatures and 0–30 ns of simulation time.
Entropy 22 00405 g009
Figure 10. Value of conformational entropy for various temperatures and 70–100 ns of simulation time.
Figure 10. Value of conformational entropy for various temperatures and 70–100 ns of simulation time.
Entropy 22 00405 g010
Figure 11. The bivariate histogram of the chosen mean RMSD signals in the 0–30 ns simulation time window.
Figure 11. The bivariate histogram of the chosen mean RMSD signals in the 0–30 ns simulation time window.
Entropy 22 00405 g011
Figure 12. The bivariate histogram of the chosen mean RMSD signals in the 70–100 ns time window.
Figure 12. The bivariate histogram of the chosen mean RMSD signals in the 70–100 ns time window.
Entropy 22 00405 g012
Table 1. Root mean square length of the end-to-end vector for 300 K 312 K .
Table 1. Root mean square length of the end-to-end vector for 300 K 312 K .
Parameter 300 K 303 K 306 K 309 K 312 K
R 2 1 / 2 [Å] for 0 ns–30 ns47.8846.6946.7247.0947.81
R 2 1 / 2 [Å] for 70 ns–100 ns47.1246.6747.7247.1347.64
Table 2. Values of H statistic in the Kruscal–Wallis test for two time windows for a root mean square end-to-end vector.
Table 2. Values of H statistic in the Kruscal–Wallis test for two time windows for a root mean square end-to-end vector.
WindowValue of H
0–30 ns H = 0.62
70–100 ns H = 1.22
Table 3. Kullback–Leibler divergence for root mean square length of mean end-to-end signals in the time window 0–30 ns. The initial data is the same as for the Kruscal–Wallis test.
Table 3. Kullback–Leibler divergence for root mean square length of mean end-to-end signals in the time window 0–30 ns. The initial data is the same as for the Kruscal–Wallis test.
Temperature of the Reference Probability 300 K 303 K 306 K 309 K 312 K
300 K 00.93241.38580.89460.7604
303 K 0.932401.46280.81761.6165
306 K 0.28020.370300.40960.1532
309 K 0.90620.81611.481602.4139
312 K 0.20400.37100.70231.18870
Table 4. Symmetric Kullback–Leibler distance for a root mean square length of end-to-end signals in the time window 0–30 ns. The initial data is the same as for the Kruscal–Wallis test.
Table 4. Symmetric Kullback–Leibler distance for a root mean square length of end-to-end signals in the time window 0–30 ns. The initial data is the same as for the Kruscal–Wallis test.
300 K 303 K 306 K 309 K 312 K
300 K 00.93240.83300.90040.4822
303 K 0.932400.91650.81690.9938
306 K 0.83300.916500.94560.4278
309 K 0.90040.81690.945601.8013
312 K 0.48220.99380.42781.80130
Table 5. Kullback–Leibler divergence for a root mean square length of end-to-end signals in the time window 70–100 ns. The initial data is the same as for the Kruscal–Wallis test.
Table 5. Kullback–Leibler divergence for a root mean square length of end-to-end signals in the time window 70–100 ns. The initial data is the same as for the Kruscal–Wallis test.
Temperature of the Reference Probability 300 K 303 K 306 K 309 K 312 K
300 K 00.19830.19830.14090.2883
303 K 0.747400.85551.66752.6390
306 K 0.74740.855501.66751.1576
309 K 0.13510.44230.442300.1925
312 K 0.81131.20841.86560.75320
Table 6. Symmetric Kullback–Leibler distance for a root mean square length of end-to-end signals in the time window 70–100 ns. The data is the same for the Kruscal–Wallis test.
Table 6. Symmetric Kullback–Leibler distance for a root mean square length of end-to-end signals in the time window 70–100 ns. The data is the same for the Kruscal–Wallis test.
300 K 303 K 306 K 309 K 312 K
300 K 00.47280.47280.13800.5498
303 K 0.472800.85551.05491.9237
306 K 0.47280.855501.05491.5116
309 K 0.13801.05491.054900.4728
312 K 0.54981.92371.51160.47280
Table 7. Kullback–Leibler divergence between distributions of mean end-to-end signals in the time window 0–30 ns.
Table 7. Kullback–Leibler divergence between distributions of mean end-to-end signals in the time window 0–30 ns.
Temperature of the Reference Probability 300 K 303 K 306 K 309 K 312 K
300 K 00.10560.31940.13670.1689
303 K 0.093900.23700.12610.1425
306 K 0.20620.145000.15280.1841
309 K 0.12760.11440.187700.2319
312 K 0.12880.12750.21260.18480
Table 8. Kullback–Leibler distance for mean end-to-end signals in the time window 0–30 ns.
Table 8. Kullback–Leibler distance for mean end-to-end signals in the time window 0–30 ns.
300 K 303 K 306 K 309 K 312 K
300 K 00.09970.26280.13210.1489
303 K 0.099700.19350.12020.1350
306 K 0.26280.193500.17030.1984
309 K 0.13210.12020.170300.2084
312 K 0.14890.13500.19840.20840
Table 9. Kullback–Leibler divergence for mean end-to-end signals in the time window 70–100 ns.
Table 9. Kullback–Leibler divergence for mean end-to-end signals in the time window 70–100 ns.
Temperature of the Reference Probability 300 K 303 K 306 K 309 K 312 K
300 K 00.09200.16050.18240.1183
303 K 0.115500.29690.16420.1474
306 K 0.11080.132400.16210.1206
309 K 0.16280.12660.271900.1206
312 K 0.08860.09790.13560.11220
Table 10. Kullback–Leibler distance for mean end-to-end signals in the time window 70–100 ns.
Table 10. Kullback–Leibler distance for mean end-to-end signals in the time window 70–100 ns.
300 K 303 K 306 K 309 K 312 K
300 K 00.10380.13560.17260.1034
303 K 0.103800.21470.14540.1226
306 K 0.13560.214700.21700.1281
309 K 0.17260.14540.217000.1164
312 K 0.10340.12260.12810.11640
Table 11. Kullback–Leibler distance for raw data end-to-end signals in the time window 70–100 ns and 306 K .
Table 11. Kullback–Leibler distance for raw data end-to-end signals in the time window 70–100 ns and 306 K .
Number of Molecule123456789
100.29270.20730.21560.24500.30160.17510.20720.4298
20.292700.16610.09190.09310.15890.23670.13100.2765
30.20730.166100.13470.21350.30590.12820.10130.1843
40.21560.09190.134700.12030.13530.14990.13380.2585
50.24500.09310.21350.120300.14780.26350.14340.3083
60.30160.15890.30590.13530.147800.28120.19780.3122
70.17510.23660.12820.14990.26350.281200.13180.2281
80.20720.13100.10130.13380.14340.19780.131800.1614
90.42980.27650.18430.25850.30830.31220.22810.16150
Table 12. Size exponent obtained according to Formula (5).
Table 12. Size exponent obtained according to Formula (5).
Parameter 300 K 303 K 306 K 309 K 312 K
Size exponent ( μ ) for 0 ns–30 ns0.39860.39470.39480.39600.3984
Size exponent ( μ ) for 70 ns–100 ns0.39610.39460.39810.39610.3978
Table 13. Values of H statistic in the Kruscal–Wallis test for μ parameter for two time windows.
Table 13. Values of H statistic in the Kruscal–Wallis test for μ parameter for two time windows.
WindowValue of Parameter H
0–30 ns H = 0.48
70–100 ns H = 1.47
Table 14. Probability (p-value) for mean entropy over time in the time interval 0–30 ns.
Table 14. Probability (p-value) for mean entropy over time in the time interval 0–30 ns.
300 K 303 K 306 K 309 K 312 K
p-value 0.849 0.601 0.009 0.536 0.143
Table 15. Probability (p-value) for entropy of single trajectory over time in the time interval 0–30 ns.
Table 15. Probability (p-value) for entropy of single trajectory over time in the time interval 0–30 ns.
Number of Signal 300 K 303 K 306 K 309 K 312 K
10.4390.7370.3670.7540.643
20.1530.6600.9610.9220.835
30.3200.4830.0120.5580.0551
40.2630.3720.7000.2230.323
50.5440.6570.0570.1300.066
60.1140.3850.2320.4160.360
70.4990.5290.0080.0060.007
80.0130.8710.6100.6920.456
90.2450.8910.8710.8320.694
mean0.2880.6210.4240.5040.382
Table 16. Probability (p-value) for entropy of a single trajectory over time in the time interval 70–100 ns.
Table 16. Probability (p-value) for entropy of a single trajectory over time in the time interval 70–100 ns.
Number of Signal 300 K 303 K 306 K 309 K 312 K
10.3280.7200.4600.0780.041
20.1830.8900.1460.7130.898
30.0920.3780.2190.4640.264
40.7640.8000.4400.6290.297
50.0300.8020.2750.3230.313
60.2200.1080.3300.4400.540
70.2790.1210.3820.2530.046
80.5820.2990.9760.3240.457
90.4380.5500.4380.9510.642
mean0.3240.5190.4070.4640.389
Table 17. Results of the Conover–Iman multi-comparison test in the time interval 70–100 ns.
Table 17. Results of the Conover–Iman multi-comparison test in the time interval 70–100 ns.
Set of Molecule 300 K 303 K 306 K 309 K 312 K
300 K -1.222216.111112.555610.3333
303 K 1.2222-17.333313.777811.5556
306 K 16.111117.3333-3.55565.7778
309 K 12.555613.77783.5556-2.2222
312 K 10.333311.55565.77782.2222-
Table 18. Kullback–Leibler divergence for the mean RMSD in the time window 0–30 ns. The data are the same as for the Kruscal–Wallis test.
Table 18. Kullback–Leibler divergence for the mean RMSD in the time window 0–30 ns. The data are the same as for the Kruscal–Wallis test.
Temperature of Reference Distribution 300 K 303 K 306 K 309 K 312 K
300 K 00.40452.07672.9170.1932
303 K 1.628001.68532.49130.6891
306 K 3.39672.605701.85972.3939
309 K 2.20530.50031.150201.1444
312 K 0.15980.13431.68452.58060
Table 19. Symmetric Kullback–Leibler distance for the mean RMSD in the time window 0–30 ns. The data are the same as for the Kruscal–Wallis test.
Table 19. Symmetric Kullback–Leibler distance for the mean RMSD in the time window 0–30 ns. The data are the same as for the Kruscal–Wallis test.
300 K 303 K 306 K 309 K 312 K
300 K 01.01622.73672.56000.1765
303 K 1.101602.14551.49580.4117
306 K 2.73672.145501.50492.0392
309 K 2.56001.49581.504901.8625
312 K 0.17650.41172.03921.86250
Table 20. Kullback–Leibler divergence for a mean RMSD in the time window 70–100 ns. The data are the same as for the Kruscal–Wallis test.
Table 20. Kullback–Leibler divergence for a mean RMSD in the time window 70–100 ns. The data are the same as for the Kruscal–Wallis test.
Temperature of Reference Distribution 300 K 303 K 306 K 309 K 312 K
300 K 00.29412.79302.15273.9428
303 K 0.837405.17562.23043.7127
306 K 1.33052.480400.89022.8295
309 K 4.01733.39753.342102.7354
312 K 3.66802.23723.38620.82710
Table 21. Symmetric Kullback–Leibler distance for the mean RMSD in the time window 70–100 ns. The data are the same as for the Kruscal–Wallis test.
Table 21. Symmetric Kullback–Leibler distance for the mean RMSD in the time window 70–100 ns. The data are the same as for the Kruscal–Wallis test.
300 K 303 K 306 K 309 K 312 K
300 K 00.56572.06173.08503.8054
303 K 0.565703.82802.81392.9749
306 K 2.06173.828002.11613.1078
309 K 3.08502.81392.116101.7812
312 K 3.80542.97493.10781.78120
Table 22. Symmetric Kullback–Leibler distance for signals of the RMSD in the time window 0–30 ns.
Table 22. Symmetric Kullback–Leibler distance for signals of the RMSD in the time window 0–30 ns.
300 K 303 K 306 K 309 K 312 K
300 K 00.36860.70901.71090.1870
303 K 0.368600.30511.29090.1838
306 K 0.70900.305100.74680.5706
309 K 1.71091.29090.746801.4513
312 K 0.18700.18380.57061.45130
Table 23. Symmetric Kullback–Leibler distance for signals of the RMSD in the time window 70–100 ns.
Table 23. Symmetric Kullback–Leibler distance for signals of the RMSD in the time window 70–100 ns.
300 K 303 K 306 K 309 K 312 K
300 K 00.10566.34875.52323.9800
303 K 0.105606.40655.77374.3824
306 K 6.34876.406500.96481.7829
309 K 5.52325.77370.964800.4260
312 K 3.98004.38241.78290.42600
Table 24. Symmetric Kullback–Leibler distance for raw signals of the RMSD in the time window 70–100 ns for 306 K .
Table 24. Symmetric Kullback–Leibler distance for raw signals of the RMSD in the time window 70–100 ns for 306 K .
Number of Molecule123456789
101.6690.3880.2770.1370.5003.2600.5600.249
21.66901.9742.5652.0741.6150.6601.4981.410
30.3881.97400.2410.6170.2463.5890.2990.352
40.2772.5650.24100.2500.4943.8670.6280.406
50.1372.0740.6170.25000.5403.5930.5970.275
60.5001.6150.2460.4940.54003.3380.0800.140
73.2570.6603.5893.8673.5933.33703.1182.795
80.5601.4980.3000.6280.5970.0803.11800.165
90.2491.4100.3520.4060.2750.1402.7950.1650

Share and Cite

MDPI and ACS Style

Weber, P.; Bełdowski, P.; Domino, K.; Ledziński, D.; Gadomski, A. Changes of Conformation in Albumin with Temperature by Molecular Dynamics Simulations. Entropy 2020, 22, 405. https://doi.org/10.3390/e22040405

AMA Style

Weber P, Bełdowski P, Domino K, Ledziński D, Gadomski A. Changes of Conformation in Albumin with Temperature by Molecular Dynamics Simulations. Entropy. 2020; 22(4):405. https://doi.org/10.3390/e22040405

Chicago/Turabian Style

Weber, Piotr, Piotr Bełdowski, Krzysztof Domino, Damian Ledziński, and Adam Gadomski. 2020. "Changes of Conformation in Albumin with Temperature by Molecular Dynamics Simulations" Entropy 22, no. 4: 405. https://doi.org/10.3390/e22040405

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop