Display options:
Normal
Show Abstracts
Compact
Select/unselect all
Displaying article 133
Research
p. 33123324
Received: 8 July 2013; in revised form: 30 July 2013 / Accepted: 19 August 2013 / Published: 23 August 2013
Show/Hide Abstract
 Cited by 1  PDF Fulltext (244 KB)
Abstract: We consider a random generalization of the classical Fibonacci substitution. The substitution we consider is defined as the rule mapping, a → baa and b → ab, with probability , and → ba, with probability 1 – p for 0 < p < 1, and where the random rule is applied each time it acts on a . We show that the topological entropy of this object is given by the growth rate of the set of inflated random Fibonacci words, and we exactly calculate its value.
p. 33253339
Received: 21 May 2013; in revised form: 4 August 2013 / Accepted: 16 August 2013 / Published: 23 August 2013
Show/Hide Abstract
 Cited by 1  PDF Fulltext (437 KB)
Abstract: EEG (Electroencephalography) signals can express the human awareness activities and consequently it can indicate the depth of anesthesia. On the other hand, Bispectralindex (BIS) is often used as an indicator to assess the depth of anesthesia. This study is aimed at using an advanced signal processing method to analyze EEG signals and compare them with existing BIS indexes from a commercial product (i.e ., IntelliVue MP60 BIS module). Multivariate empirical mode decomposition (MEMD) algorithm is utilized to filter the EEG signals. A combination of two MEMD components (IMF2 + IMF3) is used to express the raw EEG. Then, sample entropy algorithm is used to calculate the complexity of the patients’ EEG signal. Furthermore, linear regression and artificial neural network (ANN) methods were used to model the sample entropy using BIS index as the gold standard. ANN can produce better target value than linear regression. The correlation coefficient is 0.790 ± 0.069 and MAE is 8.448 ± 1.887. In conclusion, the area under the receiver operating characteristic (ROC) curve (AUC) of sample entropy value using ANN and MEMD is 0.969 ± 0.028 while the AUC of sample entropy value without filter is 0.733 ± 0.123. It means the MEMD method can filter out noise of the brain waves, so that the sample entropy of EEG can be closely related to the depth of anesthesia. Therefore, the resulting index can be adopted as the reference for the physician, in order to reduce the risk of surgery.
p. 33403360
Received: 4 July 2013; in revised form: 3 August 2013 / Accepted: 20 August 2013 / Published: 23 August 2013
Show/Hide Abstract
 Cited by 2  PDF Fulltext (4076 KB)
Abstract: Geographical location and landforms of various types have strong effects on the developments of many cities and associated street networks. This study presents new results of landform effects, in particular the geometry of shorelines, on the grid street networks (a total of 10,442 streets) of three Brazilian coastal cities. The results are compared with the street networks of three coastal cities in Britain (a total of 22,002 streets) that have evolved through a more natural “bottomup” process. Gibbs/Shannon entropy (a measure of dispersion) generally has a positive linear correlation with length ranges and the average lengths of the street, and for the powerlaw tails in particular. The geometry of the adjacent shorelines has great effect on the spatial orientation of streets in the Brazilian cities but less so for the networks of the British cities. More specifically, the more curved the shoreline, the greater is the dispersion in the street orientation and the greater the associated entropy. The results also show that the lengthentropies of the outer parts of the Brazilian cities are generally lower than those of the inner parts, whereas the entropies of outer parts of the British cities are higher than those of the inner parts, indicating dispersion during streetnetwork growth in the British cities.
p. 33613378
Received: 16 July 2013; in revised form: 12 August 2013 / Accepted: 16 August 2013 / Published: 23 August 2013
Show/Hide Abstract
 Cited by 3  PDF Fulltext (293 KB)
Abstract: Information geometry provides a tool to systematically investigate the parameter sensitivity of the state of a system. If a physical system is described by a linear combination of eigenstates of a complex (that is, nonHermitian) Hamiltonian, then there can be phase transitions where dynamical properties of the system change abruptly. In the vicinities of the transition points, the state of the system becomes highly sensitive to the changes of the parameters in the Hamiltonian. The parameter sensitivity can then be measured in terms of the FisherRao metric and the associated curvature of the parameterspace manifold. A general scheme for the geometric study of parameterspace manifolds of eigenstates of complex Hamiltonians is outlined here, leading to generic expressions for the metric.
p. 33793395
Received: 19 June 2013; in revised form: 24 July 2013 / Accepted: 13 August 2013 / Published: 29 August 2013
Show/Hide Abstract
 PDF Fulltext (284 KB)
Abstract: The idealgas barometric and pressure laws are derived from the Democritian concept of independent corpuscles moving in vacuum, plus a principle of simplicity, namely that these laws are independent of the kinetic part of the Hamiltonian. A single corpuscle in contact with a heat bath in a cylinder and submitted to a constant force (weight) is considered. The paper importantly supplements a previously published paper: First, the stability of ideal gases is established. Second, we show that when walls separate the cylinder into parts and are later removed, the entropy is unaffected. We obtain full agreement with Landsberg’s and others’ (1994) classical thermodynamic result for the entropy of a column of gas submitted to gravity.
p. 33963418
Received: 27 June 2013; in revised form: 12 August 2013 / Accepted: 12 August 2013 / Published: 29 August 2013
Show/Hide Abstract
 Cited by 1  PDF Fulltext (913 KB)
Abstract: What is information? What role does information entropy play in this information exploding age, especially in understanding emergent behaviors of complex systems? To answer these questions, we discuss the origin of information entropy, the difference between information entropy and thermodynamic entropy, the role of information entropy in complexity theories, including chaos theory and fractal theory, and speculate new fields in which information entropy may play important roles.
p. 34193434
Received: 21 June 2013; in revised form: 19 August 2013 / Accepted: 27 August 2013 / Published: 29 August 2013
Show/Hide Abstract
 PDF Fulltext (648 KB)
Abstract: To monitor water quality continuously over the entire sewer network is important for efficient management of the system. However, it is practically impossible to implement continuous water quality monitoring of all junctions of a sewer system due to budget constraints. Therefore, water quality monitoring locations must be selected as those points which are the most representative of the dataset throughout a system. However, the optimal selection of water quality monitoring locations in urban sewer networks has rarely been studied. This study proposes a method for the optimal selection of water quality monitoring points in sewer systems based on entropy theory. The proposed model searches for a quantitative assessment of data collected from monitoring points. The points that maximize the total information among the collected data at multiple locations are selected using genetic algorithm (GA) for water quality monitoring. The proposed model is demonstrated for a small urban sewer system.
p. 34353448
Received: 22 May 2013; in revised form: 2 August 2013 / Accepted: 27 August 2013 / Published: 30 August 2013
Show/Hide Abstract
 PDF Fulltext (526 KB)
Abstract: It is widely accepted that the advances in DNA sequencing techniques have contributed to an unprecedented growth of genomic data. This fact has increased the interest in DNA compression, not only from the information theory and biology points of view, but also from a practical perspective, since such sequences require storage resources. Several compression methods exist, and particularly, those using finitecontext models (FCMs) have received increasing attention, as they have been proven to effectively compress DNA sequences with low bitsperbase, as well as low encoding/decoding timeperbase. However, the amount of runtime memory required to store highorder finitecontext models may become impractical, since a contextorder as low as 16 requires a maximum of 17.2 x 10^{9} memory entries. This paper presents a method to reduce such a memory requirement by using a novel application of artificial neural networks (ANN) to build such probabilistic models in a compact way and shows how to use them to estimate the probabilities. Such a system was implemented, and its performance compared against stateofthe art compressors, such as XMDNA (expert model) and FCMMx (mixture of finitecontext models) , as well as with generalpurpose compressors. Using a combination of order10 FCM and ANN, similar encoding results to those of FCM, up to order16, are obtained using only 17 megabytes of memory, whereas the latter, even employing hashtables, uses several hundreds of megabytes.
p. 34493457
Received: 24 May 2013; in revised form: 15 August 2013 / Accepted: 19 August 2013 / Published: 30 August 2013
Show/Hide Abstract
 PDF Fulltext (1102 KB)
Abstract: The relation between the atmospheric entropy flow field and local rainfall is examined in terms of the theory of dissipative structures. In this paper, the entropy balance equation in a form suitable for describing the entropy budget of the Earth’s atmosphere is derived starting from the Gibbs relation, and, as examples, the entropy flows of the two severe weather events associated with the development of an extratropical cyclone and a tropical storm are calculated, respectively. The results show that negative entropy flow (NEF) has a significant effect on the precipitation intensity and scope with an apparent matching of the NEF’s pattern with the rainfall distribution revealed and, that the diagnosis of NEF is able to provide a good indicator for precipitation forecasting.
p. 34583470
Received: 15 July 2013; in revised form: 6 August 2013 / Accepted: 27 August 2013 / Published: 30 August 2013
Show/Hide Abstract
 Cited by 2  PDF Fulltext (1070 KB)
Abstract: In monitoring the depth of anesthesia (DOA), the electroencephalography (EEG) signals of patients have been utilized during surgeries to diagnose their level of consciousness. Different entropy methods were applied to analyze the EEG signal and measure its complexity, such as spectral entropy, approximate entropy (ApEn) and sample entropy (SampEn). However, as a weak physiological signal, EEG is easily subject to interference from external sources such as the electric power, electric knives and other electrophysiological signal sources, which lead to a reduction in the accuracy of DOA determination. In this study, we adopt the multivariate empirical mode decomposition (MEMD) to decompose and reconstruct the EEG recorded from clinical surgeries according to its best performance among the empirical mode decomposition (EMD), the ensemble EMD (EEMD), and the complementary EEMD (CEEMD) and the MEMD. Moreover, according to the comparison between SampEn and ApEn in measuring DOA, the SampEn is a practical and efficient method to monitor the DOA during surgeries at real time.
p. 34713489
Received: 29 July 2013; in revised form: 22 August 2013 / Accepted: 26 August 2013 / Published: 2 September 2013
Show/Hide Abstract
 Cited by 1  PDF Fulltext (134 KB)
Abstract: We illustrate some financial applications of the Tsallis and Kaniadakis deformed exponential. The minimization of the corresponding deformed divergence is discussed as a criterion to select a pricing measure in the valuation problems of incomplete markets. Moreover, heavytailed models for price processes are proposed, which generalized the wellknown Black and Scholes model.
p. 34903506
Received: 9 May 2013; in revised form: 17 August 2013 / Accepted: 28 August 2013 / Published: 2 September 2013
Show/Hide Abstract
 PDF Fulltext (476 KB)
Abstract: In recent years, changes in land use resulting from rapid urbanization or urban sprawl have brought about many negative effects to land ecosystems, and have led to entropy increases. This study introduces the novel ideas of a planning regulation coefficient for sustainable landuse planning in order to decrease entropy, combined with the CLUES model to predict landuse change. Three scenarios were designed as the basis for landuse projections for Guangzhou, China, in 2015, and the changes in the land ecological service function for each scenario were predicted. The results show that, although the current landuse plan is quite reasonable, it will be necessary to further strengthen the protection of farmland and important ecological service function areas.
p. 35073527
Received: 10 May 2013; in revised form: 19 August 2013 / Accepted: 27 August 2013 / Published: 3 September 2013
Show/Hide Abstract
 PDF Fulltext (1625 KB)
Abstract: All brain functions require the coordinated activity of many neurons, and therefore there is considerable interest in estimating the amount of information that the discharge of a neural population transmits to its targets. In the past, such estimates had presented a significant challenge for populations of more than a few neurons, but we have recently described a novel method for providing such estimates for populations of essentially arbitrary size. Here, we explore the influence of some important aspects of the neuronal population discharge on such estimates. In particular, we investigate the roles of mean firing rate and of the degree and nature of correlations among neurons. The results provide constraints on the applicability of our new method and should help neuroscientists determine whether such an application is appropriate for their data.
p. 35283591
Received: 28 June 2013; in revised form: 21 August 2013 / Accepted: 21 August 2013 / Published: 4 September 2013
Show/Hide Abstract
 PDF Fulltext (732 KB)
Abstract: Objective Bayesian epistemology invokes three norms: the strengths of our beliefs should be probabilities; they should be calibrated to our evidence of physical probabilities; and they should otherwise equivocate sufficiently between the basic propositions that we can express. The three norms are sometimes explicated by appealing to the maximum entropy principle, which says that a belief function should be a probability function, from all those that are calibrated to evidence, that has maximum entropy. However, the three norms of objective Bayesianism are usually justified in different ways. In this paper, we show that the three norms can all be subsumed under a single justification in terms of minimising worstcase expected loss. This, in turn, is equivalent to maximising a generalised notion of entropy. We suggest that requiring language invariance, in addition to minimising worstcase expected loss, motivates maximisation of standard entropy as opposed to maximisation of other instances of generalised entropy. Our argument also provides a qualified justification for updating degrees of belief by Bayesian conditionalisation. However, conditional probabilities play a less central part in the objective Bayesian account than they do under the subjective view of Bayesianism, leading to a reduced role for Bayes’ Theorem.
p. 35923601
Received: 15 June 2013; in revised form: 30 August 2013 / Accepted: 30 August 2013 / Published: 4 September 2013
Show/Hide Abstract
 PDF Fulltext (231 KB)
Abstract: Largescale binary integer programs occur frequently in many realworld applications. For some binary integer problems, finding an optimal solution or even a feasible solution is computationally expensive. In this paper, we develop a discrete metacontrol procedure to approximately solve largescale binary integer programs efficiently. The key idea is to map the vector of n binary decision variables into a scalar function defined over a time interval [0; n] and construct a linear quadratic tracking (LQT) problem that can be solved efficiently. We prove that an LQT formulation has an optimal binary solution, analogous to a classical bangbang control in continuous time. Our LQT approach can provide advantages in reducing computation while generating a good approximate solution. Numerical examples are presented to demonstrate the usefulness of the proposed method.
p. 36023619
Received: 4 May 2013; in revised form: 17 June 2013 / Accepted: 30 August 2013 / Published: 4 September 2013
Show/Hide Abstract
 PDF Fulltext (228 KB)
Abstract: A quantum entropy space is suggested as the fundamental arena describing the quantum effects. In the quantum regime the entropy is expressed as the superposition of many different Boltzmann entropies that span the space of the entropies before any measure. When a measure is performed the quantum entropy collapses to one component. A suggestive reading of the relational interpretation of quantum mechanics and of Bohm’s quantum potential in terms of the quantum entropy are provided. The space associated with the quantum entropy determines a distortion in the classical space of position, which appears as a Weyllike gauge potential connected with Fisher information. This Weyllike gauge potential produces a deformation of the moments which changes the classical action in such a way that Bohm’s quantum potential emerges as consequence of the non classical definition of entropy, in a nonEuclidean information space under the constraint of a minimum condition of Fisher information (Fisher Bohm entropy). Finally, the possible quantum relativistic extensions of the theory and the connections with the problem of quantum gravity are investigated. The non classical thermodynamic approach to quantum phenomena changes the geometry of the particle phase space. In the light of the representation of gravity in ordinary phase space by torsion in the flat space (Teleparallel gravity), the change of geometry in the phase space introduces quantum phenomena in a natural way. This gives a new force to F. Shojai’s and A. Shojai’s theory where the geometry of spacetime is highly coupled with a quantum potential whose origin is not the Schrödinger equation but the non classical entropy of a system of many particles that together change the geometry of the phase space of the positions (entanglement). In this way the non classical thermodynamic changes the classical geodetic as a consequence of the quantum phenomena and quantum and gravity are unified. Quantum affects geometry of multidimensional phase space and gravity changes in any point the torsion in the ordinary fourdimensional Lorenz spacetime metric.
p. 36203639
Received: 28 June 2013; in revised form: 16 July 2013 / Accepted: 30 August 2013 / Published: 4 September 2013
Show/Hide Abstract
 PDF Fulltext (427 KB)
Abstract: The main topic of this paper is a description of the generation of entropy at the end of the inflationary era. As a generalization of the present standard model of the Universe dominated by pressureless dust and a Lorentz invariant vacuum energy (LIVE), we first present a flat Friedmann universe model, where the dust is replaced with an ideal gas. It is shown that the pressure of the gas is inversely proportional to the fifth power of the scale factor and that the entropy in a comoving volume does not change during the expansion. We then review different measures of gravitational entropy related to the Weyl curvature conjecture and calculate the time evolution of two proposed measures of gravitational entropy in a LIVEdominated Bianchi type I universe, and a LemaitreBondiTolman universe with LIVE. Finally, we elaborate upon a model of energy transition from vacuum energy to radiation energy, that of Bonanno and Reuter, and calculate the time evolution of the entropies of vacuum energy and radiation energy. We also calculate the evolution of the maximal entropy according to some recipes and demonstrate how a gap between the maximal entropy and the actual entropy opens up at the end of the inflationary era.
p. 36403687
Received: 16 February 2013; in revised form: 15 July 2013 / Accepted: 28 August 2013 / Published: 6 September 2013
Show/Hide Abstract
 PDF Fulltext (1468 KB)
Abstract: Molecular modeling is an important subdomain in the field of computational modeling, regarding both scientific and industrial applications. This is because computer simulations on a molecular level are a virtuous instrument to study the impact of microscopic on macroscopic phenomena. Accurate molecular models are indispensable for such simulations in order to predict physical target observables, like density, pressure, diffusion coefficients or energetic properties, quantitatively over a wide range of temperatures. Thereby, molecular interactions are described mathematically by force fields. The mathematical description includes parameters for both intramolecular and intermolecular interactions. While intramolecular force field parameters can be determined by quantum mechanics, the parameterization of the intermolecular part is often tedious. Recently, an empirical procedure, based on the minimization of a loss function between simulated and experimental physical properties, was published by the authors. Thereby, efficient gradientbased numerical optimization algorithms were used. However, empirical force field optimization is inhibited by the two following central issues appearing in molecular simulations: firstly, they are extremely timeconsuming, even on modern and highperformance computer clusters, and secondly, simulation data is affected by statistical noise. The latter provokes the fact that an accurate computation of gradients or Hessians is nearly impossible close to a local or global minimum, mainly because the loss function is flat. Therefore, the question arises of whether to apply a derivativefree method approximating the loss function by an appropriate model function. In this paper, a new Sparse Gridbased Optimization Workflow (SpaGrOW) is presented, which accomplishes this task robustly and, at the same time, keeps the number of timeconsuming simulations relatively small. This is achieved by an efficient sampling procedure for the approximation based on sparse grids, which is described in full detail: in order to counteract the fact that sparse grids are fully occupied on their boundaries, a mathematical transformation is applied to generate homogeneous Dirichlet boundary conditions. As the main drawback of sparse grids methods is the assumption that the function to be modeled exhibits certain smoothness properties, it has to be approximated by smooth functions first. Radial basis functions turned out to be very suitable to solve this task. The smoothing procedure and the subsequent interpolation on sparse grids are performed within sufficiently large compact trust regions of the parameter space. It is shown and explained how the combination of the three ingredients leads to a new efficient derivativefree algorithm, which has the additional advantage that it is capable of reducing the overall number of simulations by a factor of about two in comparison to gradientbased optimization methods. At the same time, the robustness with respect to statistical noise is maintained. This assertion is proven by both theoretical considerations and practical evaluations for molecular simulations on chemical example substances.
p. 36883697
Received: 18 June 2013; in revised form: 20 August 2013 / Accepted: 2 September 2013 / Published: 6 September 2013
Show/Hide Abstract
 Cited by 1  PDF Fulltext (247 KB)
Abstract: Distance measures are used to quantify the extent to which information is preserved or altered by quantum processes, and thus are indispensable tools in quantum information and quantum computing. In this paper we propose a new distance measure for mixed quantum states, which we call the dynamic distance measure, and we show that it is a proper distance measure. The dynamic distance measure is defined in terms of a measurable quantity, which makes it suitable for applications. In a final section we compare the dynamic distance measure with the wellknown Bures distance measure.
p. 36983713
Received: 21 June 2013; in revised form: 21 August 2013 / Accepted: 3 September 2013 / Published: 6 September 2013
Show/Hide Abstract
 PDF Fulltext (321 KB)
Abstract: The correlation distance quantifies the statistical independence of two classical or quantum systems, via the distance from their joint state to the product of the marginal states. Tight lower bounds are given for the mutual information between pairs of twovalued classical variables and quantum qubits, in terms of the corresponding classical and quantum correlation distances. These bounds are stronger than the Pinsker inequality (and refinements thereof) for relative entropy. The classical lower bound may be used to quantify properties of statistical models that violate Bell inequalities. Partially entangled qubits can have lower mutual information than can any twovalued classical variables having the same correlation distance. The qubit correlation distance also provides a direct entanglement criterion, related to the spin covariance matrix. Connections of results with classicallycorrelated quantum states are briefly discussed.
p. 37143733
Received: 21 June 2013; in revised form: 29 August 2013 / Accepted: 30 August 2013 / Published: 6 September 2013
Show/Hide Abstract
 PDF Fulltext (210 KB)
Abstract: We consider the learning coefficients in learning theory and give two new methods for obtaining these coefficients in a homogeneous case: a method for finding a deepest singular point and a method to add variables. In application to Vandermonde matrixtype singularities, we show that these methods are effective. The learning coefficient of the generalization error in Bayesian estimation serves to measure the learning efficiency in singular learning models. Mathematically, the learning coefficient corresponds to a real log canonical threshold of singularities for the Kullback functions (relative entropy) in learning theory.
p. 37343745
Received: 21 July 2013; in revised form: 2 September 2013 / Accepted: 3 September 2013 / Published: 6 September 2013
Show/Hide Abstract
 Cited by 1  PDF Fulltext (2804 KB)

Supplementary Files
Abstract: A methodology for the determination of the solidfluid contact angle, to be employed within molecular dynamics (MD) simulations, is developed and systematically applied. The calculation of the contact angle of a fluid drop on a given surface, averaged over an equilibrated MD trajectory, is divided in three main steps: (i) the determination of the fluid molecules that constitute the interface, (ii) the treatment of the interfacial molecules as a point cloud data set to define a geometric surface, using surface meshing techniques to compute the surface normals from the mesh, (iii) the collection and averaging of the interface normals collected from the postprocessing of the MD trajectory. The average vector thus found is used to calculate the Cassie contact angle (i.e. , the arccosine of the averaged normal z component). As an example we explore the effect of the size of a drop of water on the observed solidfluid contact angle. A single coarsegrained bead representing two water molecules and parameterized using the SAFTγ Mie equation of state (EoS) is employed, meanwhile the solid surfaces are mimicked using integrated potentials. The contact angle is seen to be a strong function of the system size for small nanodroplets. The thermodynamic limit, corresponding to the infinite size (macroscopic) drop is only truly recovered when using an excess of half a million water coarsegrained beads and/or a drop radius of over 26 nm.
p. 37463761
Received: 5 August 2013; in revised form: 3 September 2013 / Accepted: 4 September 2013 / Published: 10 September 2013
Show/Hide Abstract
 PDF Fulltext (280 KB)
Abstract: In this paper, we investigate the combination synchronization of three nonlinear complex hyperchaotic systems: the complex hyperchaotic Lorenz system, the complex hyperchaotic Chen system and the complex hyperchaotic L¨u system. Based on the Lyapunov stability theory, corresponding controllers to achieve combination synchronization among three identical or different nonlinear complex hyperchaotic systems are derived, respectively. Numerical simulations are presented to demonstrate the validity and feasibility of the theoretical analysis.
p. 37623795
Received: 27 May 2013; in revised form: 2 September 2013 / Accepted: 3 September 2013 / Published: 11 September 2013
Show/Hide Abstract
 PDF Fulltext (1256 KB)
Abstract: Existing algorithms for learning Boolean networks (BNs) have time complexities of at least O(N · n^{0:7(k+1))} , where n is the number of variables, N is the number of samples and k is the number of inputs in Boolean functions. Some recent studies propose more efficient methods with O(N · n^{2} ) time complexities. However, these methods can only be used to learn monotonic BNs, and their performances are not satisfactory when the sample size is small. In this paper, we mathematically prove that OR/AND BNs, where the variables are related with logical OR/AND operations, can be found with the time complexity of O(k·(N+ logn)·n^{2} ) , if there are enough noiseless training samples randomly generated from a uniform distribution. We also demonstrate that our method can successfully learn most BNs, whose variables are not related with exclusive OR and Boolean equality operations, with the same order of time complexity for learning OR/AND BNs, indicating our method has good efficiency for learning general BNs other than monotonic BNs. When the datasets are noisy, our method can still successfully identify most BNs with the same efficiency. When compared with two existing methods with the same settings, our method achieves a better comprehensive performance than both of them, especially for small training sample sizes. More importantly, our method can be used to learn all BNs. However, of the two methods that are compared, one can only be used to learn monotonic BNs, and the other one has a much worse time complexity than our method. In conclusion, our results demonstrate that Boolean networks can be learned with improved time complexities.
p. 37963809
Received: 7 August 2013; in revised form: 2 September 2013 / Accepted: 9 September 2013 / Published: 12 September 2013
Show/Hide Abstract
 PDF Fulltext (1782 KB)
Abstract: Microstructure and phase composition of a CrMo_{0.5} NbTa_{0.5} TiZr high entropy alloy were studied in the assolidified and heat treated conditions. In the assolidified condition, the alloy consisted of two disordered BCC phases and an ordered cubic Laves phase. The BCC1 phase solidified in the form of dendrites enriched with Mo, Ta and Nb, and its volume fraction was 42%. The BCC2 and Laves phases solidified by the eutectictype reaction, and their volume fractions were 27% and 31%, respectively. The BCC2 phase was enriched with Ti and Zr and the Laves phase was heavily enriched with Cr. After hot isostatic pressing at 1450 °C for 3 h, the BCC1 dendrites coagulated into roundshaped particles and their volume fraction increased to 67%. The volume fractions of the BCC2 and Laves phases decreased to 16% and 17%, respectively. After subsequent annealing at 1000 °C for 100 h, submicronsized Laves particles precipitated inside the BCC1 phase, and the alloy consisted of 52% BCC1, 16% BCC2 and 32% Laves phases. Solidification and phase equilibrium simulations were conducted for the CrMo_{0.5} NbTa_{0.5} TiZr alloy using a thermodynamic database developed by CompuTherm LLC. Some discrepancies were found between the calculated and experimental results and the reasons for these discrepancies were discussed.
p. 38103821
Received: 19 July 2013; in revised form: 10 September 2013 / Accepted: 10 September 2013 / Published: 12 September 2013
Show/Hide Abstract
 Cited by 3  PDF Fulltext (1339 KB)

Supplementary Files
Abstract: Highentropy (HE) alloys, bulk metallic glasses (BMGs) and highentropy BMGs (HEBMGs) were statistically analyzed with the help of a database of ternary amorphous alloys. Thermodynamic quantities corresponding to heat of mixing and atomic size differences were calculated as a function of composition of the multicomponent alloys. Actual calculations were performed for configurational entropy (S _{config.} ) in defining the HE alloys and mismatch entropy (S _{s} ) normalized with Boltzmann constant (k _{B} ), together with mixing enthalpy (DH _{mix} ) based on Miedema’s empirical model and Delta parameter (d ) as a corresponding parameter to S _{s} /k _{B} . The comparison between DH _{mix} –d and DH _{mix} – diagrams for the ternary amorphous alloys revealed S _{s} /k _{B} ~ (d /22)^{2} . The zones S, S′ and B’s where HE alloys with disordered solid solutions, ordered alloys and BMGs are plotted in the DH _{mix} –d diagram are correlated with the areas in the DH _{mix} – S _{s } /k _{B} diagram. The results provide mutual understandings among HE alloys, BMGs and HEBMGs.
p. 38773891
Received: 9 July 2013; in revised form: 22 August 2013 / Accepted: 10 September 2013 / Published: 13 September 2013
Show/Hide Abstract
 PDF Fulltext (584 KB)
Abstract: Applying the particle filter (PF) technique, this paper proposes a PFbased algorithm to blindly demodulate the chaotic direct sequence spread spectrum (CDSSS) signals under the colored or nonGaussian noises condition. To implement this algorithm, the PFs are modified by (i) the colored or nonGaussian noises are formulated by autoregressive moving average (ARMA) models, and then the parameters that model the noises are included in the state vector; (ii) the rangedifferentiating factor is imported into the intruder’s chaotic system equation. Since the rangedifferentiating factor is able to make the inevitable chaos fitting error advantageous based on the chaos fitting method, thus the CDSSS signals can be demodulated according to the range of the estimated message. Simulations show that the proposed PFbased algorithm can obtain a good biterror rate performance when extracting the original binary message from the CDSSS signals without any knowledge of the transmitter’s chaotic map, or initial value, even when colored or nonGaussian noises exist.
p. 38923909
Received: 26 July 2013; in revised form: 10 September 2013 / Accepted: 12 September 2013 / Published: 16 September 2013
Show/Hide Abstract
 Cited by 1  PDF Fulltext (843 KB)
Abstract: Seismic data is difficult to analyze and classical mathematical tools reveal strong limitations in exposing hidden relationships between earthquakes. In this paper, we study earthquake phenomena in the perspective of complex systems. Global seismic data, covering the period from 1962 up to 2011 is analyzed. The events, characterized by their magnitude, geographic location and time of occurrence, are divided into groups, either according to the FlinnEngdahl (FE) seismic regions of Earth or using a rectangular grid based in latitude and longitude coordinates. Two methods of analysis are considered and compared in this study. In a first method, the distributions of magnitudes are approximated by GutenbergRichter (GR) distributions and the parameters used to reveal the relationships among regions. In the second method, the mutual information is calculated and adopted as a measure of similarity between regions. In both cases, using clustering analysis, visualization maps are generated, providing an intuitive and useful representation of the complex relationships that are present among seismic data. Such relationships might not be perceived on classical geographic maps. Therefore, the generated charts are a valid alternative to other visualization tools, for understanding the global behavior of earthquakes.
p. 39103930
Received: 24 July 2013; in revised form: 29 August 2013 / Accepted: 11 September 2013 / Published: 16 September 2013
Show/Hide Abstract
 PDF Fulltext (306 KB)
Abstract: Recently, the duality between values (words) and orderings (permutations) has been proposed by the authors as a basis to discuss the relationship between information theoretic measures for finitealphabet stationary stochastic processes and their permutatio nanalogues. It has been used to give a simple proof of the equality between the entropy rate and the permutation entropy rate for any finitealphabet stationary stochastic process and to show some results on the excess entropy and the transfer entropy for finitealphabet stationary ergodic Markov processes. In this paper, we extend our previous results to hidden Markov models and show the equalities between various information theoretic complexity and coupling measures and their permutation analogues. In particular, we show the following two results within the realm of hidden Markov models with ergodic internal processes: the two permutation analogues of the transfer entropy, the symbolic transfer entropy and the transfer entropy on rank vectors, are both equivalent to the transfer entropy if they are considered as the rates, and the directed information theory can be captured by the permutation entropy approach.
p. 39313940
Received: 1 August 2013; in revised form: 26 August 2013 / Accepted: 11 September 2013 / Published: 18 September 2013
Show/Hide Abstract
 PDF Fulltext (163 KB)
Abstract: We investigate solutions of a generalized diffusion equation that contains nonlinear terms in the presence of external forces and reaction terms. The solutions found here can have a compact or long tail behavior and can be expressed in terms of the q exponential functions present in the Tsallis framework. In the case of the longtailed behavior, in the asymptotic limit, these solutions can also be connected with the L´evy distributions. In addition, from the results presented here, a rich class of diffusive processes, including normal and anomalous ones, can be obtained.
p. 39703982
Received: 13 May 2013; in revised form: 6 August 2013 / Accepted: 27 August 2013 / Published: 23 September 2013
Show/Hide Abstract
 PDF Fulltext (722 KB)
Abstract: The paper discusses methods of evaluating thermodynamic properties of landscape cover based on multispectral measurements by the Landsat satellites. Authors demonstrate how these methods could be used for studying functionality of landscapes and for spatial interpolation of Flux NET system measurements.
Review
p. 38223876
Received: 3 June 2013; in revised form: 26 August 2013 / Accepted: 30 August 2013 / Published: 13 September 2013
Show/Hide Abstract
 Cited by 3  PDF Fulltext (708 KB)
Abstract: This paper postulates that water structure is altered by biomolecules as well as by diseaseenabling entities such as certain solvated ions, and in turn water dynamics and structure affect the function of biomolecular interactions. Although the structural and dynamical alterations are subtle, they perturb a wellbalanced system sufficiently to facilitate disease. We propose that the disruption of water dynamics between and within cells underlies many disease conditions. We survey recent advances in magnetobiology, nanobiology, and colloid and interface science that point compellingly to the crucial role played by the unique physical properties of quantum coherent nanomolecular clusters of magnetized water in enabling life at the cellular level by solving the “problems” of thermal diffusion, intracellular crowding, and molecular selfassembly. Interphase water and cellular surface tension, normally maintained by biological sulfates at membrane surfaces, are compromised by exogenous interfacial water stressors such as cationic aluminum, with consequences that include greater local water hydrophobicity, increased water tension, and interphase stretching. The ultimate result is greater “stiffness” in the extracellular matrix and either the “soft” cancerous state or the “soft” neurodegenerative state within cells. Our hypothesis provides a basis for understanding why so many idiopathic diseases of today are highly stereotyped and pluricausal.
p. 39413969
Received: 29 July 2013; in revised form: 6 September 2013 / Accepted: 16 September 2013 / Published: 23 September 2013
Show/Hide Abstract
 PDF Fulltext (1039 KB)
Abstract: Since most experimental observations are performed at constant temperature and pressure, the isothermalisobaric (NPT ) ensemble has been widely used in molecular simulations. Nevertheless, the NPT ensemble has only recently been placed on a rigorous foundation. The proper formulation of the NPT ensemble requires a “shell” particle to uniquely identify the volume of the system, thereby avoiding the redundant counting of configurations. Here, we review our recent work in incorporating a shell particle into molecular dynamics simulation algorithms to generate the correct NPT ensemble averages. Unlike previous methods, a piston of unknown mass is no longer needed to control the response time of the volume fluctuations. As the volume of the system is attached to the shell particle, the system itself now sets the time scales for volume and pressure fluctuations. Finally, we discuss a number of tests that ensure the equations of motion sample phase space correctly and consider the response time of the system to pressure changes with and without the shell particle. Overall, the shell particle algorithm is an effective simulation method for studying systems exposed to a constant external pressure and may provide an advantage over other existing constant pressure approaches when developing nonequilibrium molecular dynamics methods.
Select/unselect all
Displaying article 133
Export citation of selected articles as:
Plain Text
BibTeX
BibTeX (without abstracts)
Endnote
Endnote (without abstracts)
Tabdelimited
PubMed XML
DOAJ XML
AGRIS XML