- freely available
cells 2013, 2(2), 393-413; doi:10.3390/cells2020393
Published: 31 May 2013
Abstract: One of the principle tasks of systems biology has been the reverse engineering of signaling networks. Because of the striking similarities to engineering systems, a number of analysis and design tools from engineering disciplines have been used in this process. This review looks at several examples including the analysis of homeostasis using control theory, the attenuation of noise using signal processing, statistical inference and the use of information theory to understand both binary decision systems and the response of eukaryotic chemotactic cells.
Though no two researchers are likely to agree on a definition of systems biology, there is no doubt that an important aspect is the way that it integrates research from an array of disciplines by combining novel experimental and computational tools. The various engineering disciplines have contributed many tools that facilitate experimentation, computation and data processing. Equally important is the role that our understanding of the engineering design process can have in deciphering biological systems. Issues such as modularity, robustness, optimality, tradeoffs, and physical constraints are dealt with every day in engineering design, and a large number of theoretical tools have been developed to facilitate this process. In studying biology, it is important to understand that these tradeoffs and constraints also arise and hence must be dealt with [1,2].
This review highlights how some of this understanding has played a role in the reverse engineering of biological systems. The emphasis is on a number of theoretical tools from research areas in engineering that in some sense deal with abstract representations of signals and systems and hence are readily amenable for use in system biology, particularly in the study of signal transduction pathways. An example is the set of design and analysis tools developed by control engineers. Though this field grew out of different application areas in chemical, electrical and aerospace engineering , the tools used by control engineers are usually not specific to any particular application, but instead focus on signals that need to be controlled and mathematical descriptions, rather than hardware implementation, of systems that perform that regulation. Thus, control engineering tools have found applications in the systems biology research [4–9]. Similarly, in signal processing and information theory, the emphasis is placed on the statistical properties of signals and how to alter these to achieve the desired objectives while emphasizing constraints that are placed by noise.
The rest of this review is divided into four parts. The first three look at different engineering disciplines, though it is important to stress that even within engineering, the divisions between these areas is not always so clear-cut. Finally, some concluding remarks are presented.
2. Homeostasis and Control Engineering
Since the work of Claude Bernard in the 19th century, the process by which an organism's internal state is kept constant—le milieu intérieur—has long fascinated biologists . In the 20th century, Walter Cannon termed this process homeostasis and argued that diseases arose because of a homeostatic imbalance . This imbalance could be due to an inability for the organism to cope with changing environments, either because of a loss of function of its regulatory mechanisms, or because the level of the stimulus is too large for these mechanisms to cope. Cannon proposed six principles that allowed a system to achieve homeostasis. One of these:
If the state remains steady, there is an automatic arrangement whereby any tendency toward change is effectively met by increased action of the factor or factors which resist the change. (Cited in .)
clearly expresses the need for negative feedback. The necessity of Cannon's principles, although a result of intuition based on his understanding of human physiology, can be stated in a mathematically formal way, as we will see below.
2.1. Perfect Adaptation and the Internal Model Principle
To illustrate how feedback arises in the context of a homeostatic response, we use a simple three-node network where an external stimulus (input; denoted U) gives rise to a response (output; denoted Z) by activating a signaling cascade ; see Figure 1(a). Each node in the network represents a signaling element that exists in two states: active and inactive, and the conversion between the two is based on enzymatic reactions, represented by Michaelis–Menten kinetics. Positive interactions increase the rate of the enzymatic reaction activating the product; negative interactions increase the rate of inactivation. The network includes a negative feedback loop achieved through node Y. The strength of this interaction can be changed by altering the parameter kZY.
Based on these assumptions, the following set of differential equations describes the system, where x, y, z and u denote suitably normalized concentrations of the active state of the various componentsThe specific parameter values are not important, but all coefficients are assumed to be positive.
Suppose that we are interested in making the system insensitive to changes in the stimulus concentration. Because this is a dynamical system, we need to be more precise as to the class of inputs to which we would like to be insensitive, and we need a means for measuring sensitivity. For the former, it is customary to consider changes in the set point of the system—that is, we assume that the stimulus is at a steady-state value, ui, and that at a given time, t = 0, it is changed to a new value uf > ui. This step change in concentration gives rise to a transient change in the response from its pre-stimulus value (zi), before settling to a new steady-state (zf); see Figure 1(b). The precision, or steady-state sensitivity, is the fractional change in the concentration of the response, relative to the fractional change in the stimulus: P = ((uf − ui)/ui)/((zf − zi)/zi) . In general, this is a nonlinear function, so that the initial stimulus level makes a difference. The analysis can be simplified if we assume that the system is initially at rest; that is, ui = 0, from which xi = yi = z = 0 follows. In this case, the definition above is not suitable; instead, we let P = uf/zf. With either definition, if the system output returns to the initial value, that is, it is completely insensitive to the external stimulus, then the precision is infinite and the system is said to achieve perfect adaptation.
Figure 1(c) illustrates the effect that varying the strength of the negative feedback has on the system. Note that, for a given feedback strength, the response increases after the change in the stimulus level. However, as the strength of the negative feedback is increased, the sensitivity decreases. To make the system adapt perfectly requires an infinitely large negative feedback gain. This is obviously not possible in general, but can be achieved in certain cases. If the negative feedback pathway is operating at saturation (1 − y ≫ kY and ) then the differential equation for the feedback is approximatelywhere z* = k−Y/kZY is the steady-state set-point, which is independent of the stimulus level. Evaluating this equation at the equilibrium (where the derivative is zero) shows that the output returns to the set-point that is independent of the level of stimulus. Integrating Equation 1 yields where e(t) = z* − z(t) is the deviation of the output from its set point. It is clear from this expression that the feedback connection is integrating changes from the set-point. In engineering, this form of feedback is referred to as integral control, and it is well known that for a system to be completely insensitive to external step changes in the set point, like those considered here, the system must have integral control feedback . Moreover, this is a special case of a more general set of results known as the Internal Model Principle (IMP), which states that for a system to be robustly insensitive to an external change in its environment, it must have both a negative feedback connection (echoing Cannon's postulate) and an internal dynamical system that can recreate the class of disturbances [14,15]. Specifically, integrators recreate step disturbances, but other classes of stimuli are possible and require other forms of negative feedback.
To illustrate how integral feedback is providing infinite gain, it is useful to write the integrator in terms of a Fourier transform. In this case,The ratio between the Fourier transforms of the error e(t) and feedback signal y(t) is known as the transfer function and, in this case, equals −kZY/(iω). In particular, we see that at steady-state, which corresponds to frequency ω = 0, there is an infinite gain between the deviation from steady-state and the control signal. Thus, though the controller gain is not infinite for all frequencies, it is at the particular frequency where the desired adaptation precision is to be measured (ω = 0).
The frequency domain analysis presented above forms one of the strongest tools of control engineering and signal processing. It does require that the particular system be linear, or at least linearized so that only small deviations from the steady-state are considered. In linearization, a signal (e.g., x(t)) is considered as the sum of two parts: the constant steady-state (x*) and deviations from this steady-state (δx(t) = x(t) − x*). If we perform this linearization on the system above and write the corresponding signals in the frequency domain using the Fourier transform, we obtain the closed-loop transfer functionwhich describes the effect of a sinusoidal deviation from steady-state on the output; see Figure 1(d). The coefficients in the denominator are all positive (shown in the Appendix), guaranteeing that the system is stable. It is worth noting that the term (iω) that appeared in the denominator when describing the controller (Equation 2) is now in the numerator of the closed-loop system. This term says that the gain at steady-state is zero—which is equivalent to saying that the system is attenuating constant signals; in other words, the system is performing temporal differentiation. The IMP states that integral feedback is required for perfect adaptation to step changes in the environmental stimuli. Thus, we can expect to see integral feedback in action in biological systems that exhibit perfect adaptation, and this has been the case. The first, and best known, was in the study of the signaling system controlling chemotaxis in E. coli. These cells move through the action of a number of flagella. When rotating in a counterclockwise fashion, the flagella form a bundle and propel the cell forward in a run. If the rotation is clockwise, the flagella unbundle, and the cell comes to a sudden stop. In this state, the cell undergoes a tumble, a random reorientation in space. Through a series of runs and tumbles, the cell performs a random walk through space. In response to chemoattractants, E. coli modulates the time between tumbles, lengthening it if moving up a gradient, and shortening when moving away from chemoattractants. This biases the direction of the random walk so that, on average, cells migrate up a chemoattractant gradient. Experimentally, the effect of chemoattractant on the rotational bias can be observed in the lab by monitoring the probability that flagella rotate in clockwise fashion before and after stimulation. After the addition of a persistent stimulus, the probability of a clockwise rotation drops for approximately 3–4 seconds before adapting to its pre-stimulus level. Alon et al. showed that the chemotactic system displayed perfect adaptation in spite of genetic mutations that greatly changed the expression levels of various components of the signaling pathway . Analysis of the equations that govern adaptation in bacteria  demonstrated that this robustness could be attributed to the presence of an integral control feedback element in the loop .
A second example is the network regulating the response to osmotic shock in S. cerevisiae yeast cells. When faced by changes in turgor pressure, yeast cells respond by activating the High Osmolarity Glycerol (Hog1) pathway, a mitogen-activated protein kinase (MAPK) cascade . This activation, which involves the dual phosphorylation of the MAPK Hog1, can be detected by tracking fluorescently-tagged Hog1. Inactive Hog1 proteins are usually cytoplasmic but, after phosphorylation and activation, translocate to the nucleus . The osmotic-shock system was known to have multiple feedback loops operating at different time scales, but there had previously been no understanding of their specific biological functions. Mettetal et al. applied periodically varying stimuli at different frequencies to both wild-type and mutant cells  as well as other stimuli like steps and ramps . By measuring the response to these stimuli, they were able to recreate transfer functions experimentally for the various cells strains, which were used to conclude that the osmo-adaptation response is dominated by the feedback path through the kinase Hog1.
While the IMP states that integral control must be present in order to achieve perfect adaptation to step changes in stimuli, the nature of the integral control may not always be readily apparent. For example, though adaptation in the network of Figure 1 is achieved using saturating negative feedback, there is a second way that adaptation can be achieved in a three node network , as illustrated in Figure 2. The external signal U increases the production rate of X, which increases the production of both Y and Z. Signal Y acts negatively on the response Z. Because the external signal regulates the production of Z in two complementary ways (positive through X; negative through Y) the system is said to have an incoherent feedforward loop (IFL) . If these regulations are such that forward reactions are saturated (e.g., 1 − x ≫ kX, etc.) and the reverse reactions linear (e.g., , etc.) then:(In these equations we have incorporated into k−X, etc.) At steady-state, X is proportional to the stimulus level (xf = (kUX/k−X)uf), Y is proportional to X (yf = (kXY/k−Y)xf) and the response is proportional to the ratio between X and Y: and hence is independent of the external stimulus. This mechanism of perfect adaptation was first proposed by Koshland to explain adaptation in E. coli , but is now more widely accepted as a model for sensory adaptation in eukaryotic cells [24–26]. Note this adaptation is perfect and is also extremely robust to parameters , though it is not apparent that it includes feedback, much less integral control. However, using a change of coordinates, it can be shown that this system can be recast in terms of an integral feedback [27,28]. This shows that though the IMP holds, its application is not always straightforward.
2.2. Fundamental Constraints on Sensitivity Minimization
In the frequency domain, the transfer function H(iω) exhibits a biphasic response (Figure 1(d)). The (infinitely) small gains at low frequency arise because of the presence of integral control. However, at mid-frequencies, the gain increases, signifying that there are periodic stimuli whose response is actually enhanced. In the time domain, this is manifested as the oscillatory response seen in Figure 1(c). This behavior is a result of a fundamental result from control theory usually referred to as Bode's Integral Formula, but more colloquially known as the waterbed effect, which places a constraint on the total amount of sensitivity reduction that a system can provide [29,30]; see Figure 3.
To express the sensitivity we consider the transfer function and write this as the product of two functions, H(iω) = H0(iω)S(iω) (see Appendix). The first function describes the system without any feedback. The second, referred to as the sensitivity function, specifies how feedback reduces the gain of the closed-loop system. In terms of the sensitivity transfer function at various frequencies, Bode's integral states thatThough written for systems with transfer functions, this result is considerably more general [31,32]. Moreover, for certain classes of systems, the right-hand side can be made more precise . Nevertheless, as written, this result states that the set of frequencies for which the sensitivity is attenuated (log |S(iω)| < 0) must be countered by other frequencies for which the signal is amplified (log |S(iω)| > 0). Moreover, the greater the attenuation is imposed at some frequencies, the larger the amplification must be at others (Figure 3(b)). The regions of frequency over which the sensitivity is attenuated can be said to correspond to regions where the feedback is negative; similarly, sensitivity amplification corresponds to positive feedback. Thus, Bode's integral can be interpreted as implying that in any given system, there must be regions of both negative and positive feedback and that these must be balanced. This echoes another of Cannon's dictums on homeostasis:
Factors which may be antagonistic in one region, where they effect a balance, may be cooperative in another region. (Cited in .)
Chandra and coworkers have shown how this constraint manifests itself during glycolysis, the process by which cells convert glucose into adenosine triphosphate . Under certain conditions, metabolites involved in glycolysis display oscillations. A large number of models ranging in complexity have been used to recreate these oscillations and aspects such as synchronization. However, the precise reason for these oscillations has not been clear. Chandra et al. considered a minimal model of glycolytic oscillations that focused on allosteric activation of phosphofructokinase (PFK) by adenosine monophosphate (AMP) and used this model to demonstrate that, as the control gain increases, the steady-state response decreases, but this is accompanied by an amplification in the gain at mid-frequencies. For sufficiently high feedback gain, this increase can even manifest itself as sustained oscillations. Thus, glycolytic oscillations arise as the inevitable tradeoff of desiring low sensitivity at low frequencies and which can be quantified by Bode's integral.
3. Extracting Information in the Presence of Noise
The proper function of cells and organisms depends on their ability to sense their environment and to make correct decisions based on these observations. Both these steps are hampered by stochastic fluctuations inherent in biochemical systems, which have their origin in the random collisions that take place at the molecular level. These fluctuations are usually referred to as noise, as they tend to mask the mean level, or signal. Engineering systems are no different, and complex and elaborate theory has evolved for studying and handling this stochasticity. These tools have found their way into the study of biological signaling pathways, and are increasingly making an impact. Here we do not go into the details of how the fluctuations arise or how they are best incorporated into mathematical models, as there are excellent reviews on both these subjects [34,35]. Rather, we focus on how engineering principles arise in studying these systems.
3.1. Statistical Inference and Bayes's Rule
To make well-informed decisions, cells must be able to sense their environments accurately. This process is usually done with cell-surface receptors in a reversible binding process we describe bywhere the receptor (R) binds to a molecule of the ligand (L) and becomes occupied (C). The deterministic differential equation recreates changes in the expected number of occupied receptors but does not account for fluctuations (Figure 4(a)). If a cell is to make an accurate decision based on receptor occupancy, it must do so not based on the actual external signal (we assume that the concentration of the ligand captures this accurately, though even in this case, there will be fluctuations that can mask its true nature ), but on the sensed signal, which is receptor occupancy In mathematics and statistics, this is usually the realm of statistical inference . In this field, we make a distinction between the observed or measured signal (which in our case is the steady-state receptor occupancy, C) and the environmental signal L. We seek to infer the state of the environment given measurements of receptor occupancy. Because of stochasticity, there is no single measure for either; instead, we work in a probabilistic sense and ask, what is the conditional probability that the system is in state (or has concentration) L if level C is being observed? The optimal solution to this problem is based on Bayes's rule: Note that, in our example, receptor occupancy is related to the conditional probability P(C∣L), which specifies the likelihood that for a given ligand concentration a certain level of receptor occupancy is obtained. This expression can be rewritten as where the summation in the denominator is taken over all possible states of the environment. It is also worth noting that to solve this problem, one must have a probabilistic description of the environment (that is, P(Li)) as well as a suitable model of how these environmental signals give rise to the observed signal (P(C∣Li)).
As an example of this optimal statistical inference, Libby et al. considered a simple model of the lac operon in E. coli . The lac operon is responsible for up-regulating enzymes needed to metabolize lactose. These enzymes are not found constitutively in cells, but are only produced when lactose is present and glucose is absent in the environment. Libby et al. assumed that the environment can be described by two states, corresponding to either an environment that is rich in lactose and poor in glucose, in which cells respond by expressing the lac operon, or one that is poor in lactose and rich in glucose, where expression should be repressed. They considered a number of probability distributions for the various terms in the expression for P(L∣C) (note that there are only two terms in the summation) and identified a unique posterior probability distribution that matched the observed lac operon transcription data. In doing so they demonstrated that the lac operon regulatory network solves an optimal statistical inference problem.
3.2. Noise Suppression through Temporal Filtering
The inference problem considered above relies on a probabilistic description of the environment and the observations that can be expected. An alternate view of how the non-deterministic nature of the signaling system gives rise to fluctuations is possible. Here we assume that the observed signal, now denoted by C(t), is a time-varying observation of the environment, which can be represented by a constant ligand concentration, L. The expected level of receptor occupancy (assuming that the concentration of the ligand is constant) iswhere KD = koff/kon is the dissociation constant for the receptor. Over time, C(t) fluctuates around C*, but this fluctuation has zero mean (Figure 4(a)). Estimates of C* based on C(t) will be improved if we average our observations over time, which leads to a low pass filter [39–42]. This low-pass filtering can be carried out by the following reaction scheme whose expected response can be described by the differential equation: where C̃(t) denotes the estimate of C(t) and the parameter k specifies the filter's bandwidth—this is the highest frequency for which C(iω) is not attenuated by the filter. The latter can best be observed in the frequency domain by computing the transfer function of the filter: This transfer function's gain is . For frequencies smaller than k, the gain is approximately one. However, if ω ≫ k, then the gain decreases as k/ω. Thus, lower bandwidths (smaller k) filter out more noise, giving rise to better estimates as measured by the variance of the distribution (Figure 4(c)). However, this improved estimate is obtained by including a longer period of observation, and this can cause a significant delay before changes in the environment are detected.
As we see that the bandwidth has an important effect on the sensed signal, it is worth asking whether there is an optimal bandwidth. A number of reports have considered this question with respect to the signaling pathway controlling chemotaxis in E. coli [39,41,43,44]. As mentioned above, a step change in chemoattractant gives rise to a transient response that is followed by perfect adaptation to the pre-stimulus level—a process that can be described as temporal differentiation. In the wild, bacteria are not likely to experience step changes in chemoattractant concentrations, but instead experience gradients of chemoattractant concentration. At least to first degree, these can be approximated as temporal ramps in receptor occupancy that, when differentiated, give rise to a signal that is proportional to the steepness in the gradient and that cells can use to guide their motion. However, binding noise, which is amplified by the temporal differentiation process, perturbs this signal. Thus, cells must filter out the effect of noise to make well-informed decisions. This is done, by low-pass filtering. However, because of biophysical limitations imposed by bacterial swimming—cells experience rotational diffusion that does not allow them to maintain a straight course for more than a few seconds, there is a limit to how far back into the past the low pass filter can go. Estimates based on biophysical constraints have been made [36,41,45], but Andrews et al. used an engineering approach to estimate the optimal low-pass filter. Using a model of receptor binding and incorporating known physical limits, they posed the problem as an optimal estimation problem and used the Kalman filter theory  to obtain the optimal filter. A comparison of its transfer function to that observed in cells showed strong agreement . Interestingly, the Kalman filter is a special case of an optimal Bayes inference under the assumptions that the noise is Gaussian (a reasonable assumption for binding) and that the system is linear.
4. Information-Theoretic Analyses of Signaling Pathways
To a great extent, cell survival requires the ability to make informed decisions based on imprecise measurements of the external environment. Though filtering can help reduce the effect of random perturbations, even the best filtering may not sufficiently remove noise to allow adequate information about the environment to be transmitted to the cell. Prompted by the problem of transmitting signals reliably, Claude Shannon developed a framework that is now known as information theory [47,48]. This theory, which is sufficiently general that can be used to understand signal transduction in cell signaling, is only now beginning to have an impact in this biological setting [49–53].
4.1. Quantifying the Amount of Information in a Signal
Before seeing how the theory can be used to gain insight into biological systems, it is useful first to understand how information is quantified. The framework is probabilistic. A random signal X is assumed to be in one of a number of states, say xi, according to a probability density function p(x). The entropy of the signal is given by H(X) = − Σ p(xi) log2 p(xi), and is measured in bits. The entropy is a measure of the uncertainty in the variable. For example, if X can be in one of two states (say “high” or “low”) with equal probability, then , and the entropy is equal to one bit. If the signal is equally likely to be in one of four states, then the entropy is two bits.
A related notion is that of mutual information. For two variables X and Y the mutual informationis the reduction in uncertainty in one variable assuming that the other is known; it is symmetric (I(Y; X) = I(X; Y)), so that we can talk about the mutual information between two variables.
One of Shannon's remarkable results was to provide an upper bound on the information transmission capability of a communication channel. The set-up is as in Figure 5(a). Here, the signal X is the signal to be transmitted, and Y is the received signal. The nature of the channel dictates the joint probability density function between X and Y. For example, if data transmission is over a channel that introduces additive white Gaussian noise (AWGN), then for any input x ∈ X, the received signal is y = x + n, where n is a zero-mean Gaussian random variable with variance σ2. Shannon proved that the channel capacity—the maximum rate at which information can be transmitted without any loss in fidelity—is given by the greatest mutual information between X and Y, where the maximum is computed over all possible input probability density functions. In the case of the AWGN channel, this is infinite unless further constraints are placed on the signal X. Typically, this is in the form of a finite power: . In this case, the channel capacity can be shown to be , and this is achieved when the input signal is also a zero-mean Gaussian with variance P.
Cheong et al. sought to measure the mutual information in a biological signaling network . They stimulated mouse fibroblasts with tumor necrosis factor (TNF) at various concentrations and measured the concentrations of the transcription factor NF-κB. They found that the system could only produce slightly less than one bit; that is, it can only really discriminate between the presence and absence of TNF. Their results also suggest that the relatively low information processing capabilities of the system can be attributed to an information bottleneck around the receptor.
This study also showed that negative feedback has a limited role in improving the information capability of the network. Lestas et al. considered this question  for the problem of reducing stochastic fluctuations in the number of molecules of a system involving feedback control. The system involves two chemical species. Molecule X controls the rate of production of molecule Y. The latter provides feedback, controlling the rate of production of X. The precise form of this control, denoted U, is not specified. Instead, arbitrarily complex signaling systems are considered with the only restriction being that the system be causal—that is, the control signal can only use past values of Y. Information-theoretic methods provided lower bounds on the standard deviation in the number of molecules. These bounds decrease only with the quartic root of the number of signaling events, showing that increasing accuracy is prohibitively expensive.
4.2. Information Transmission in Binary Decision Processes
The measurements of the NF-κB system suggest that the system is processing binary information (the presence or absence of TNF) to make a binary decision (whether to initiate transcription or not). Binary decisions such as this one are quite common in biology and include the lac operon system described above. A branch of information theory known as rate distortion theory has been used to study the effectiveness of this class of decisions. Unlike Shannon's channel capacity theory, which deals with lossless information transmission, rate distortion theory assumes that some error (the distortion) is allowed. The amount of distortion is measured using the rate distortion function. In a binary system, one possible distortion function would be the Hamming distance , which is one if the wrong decision is made and zero otherwise. For example, in the lac operon, if lactose is present and the operon is not activated, then the distortion equals one. Rate distortion theory allows one to pose questions such as: what stimulus–response map (expressed as the conditional probability p(Y∣X)) achieves a pre-specified level of performance (expressed as a distortion level) while minimizing the required transmission cost (the number of bits of mutual information between X and Y)? Using this framework, it is possible to show that a number of responses observed in biological signaling pathways arise as the optimal response of systems . For example, by replacing the Hamming distance with a distortion function that penalizes incorrect decisions gradually, depending on where the stimulus concentration is in relation to a threshold, the optimal stimulus–response map is sigmoidal . Moreover, requiring higher fidelity (lower allowed distortion) leads to steeper sigmoidal functions as measured by the Hill coefficient. In this framework, responses that include hysteresis or irreversibility appear as optimal strategies.
4.3. Information Processing During Eukaryotic Chemotaxis
Not all the relevant information in cell signaling is binary. For example, chemotactic cells must interpret spatial differences (as small as 1%) in the concentration of diffusible chemical in order to guide their locomotion. The information-processing capabilities of these cells has led to a number of interesting studies in both amoebae [55,60,61] and axons [62,63].
To estimate mutual information between the chemoattractant gradient and the spatial response, the cell membrane can be divided into n segments, each with N/n receptors (N is the total number of receptors.) Assuming a simple model of ligand binding (as in Section 3.1), Fuller et al. calculated that the mutual information between the direction of the gradient (θs) and the concentration of occupied receptors (C) is given bywhere p is the gradient in chemoattractant concentration, expressed as a percentage of the mid-point of receptor occupancy . These calculations show that the mutual information is a biphasic curve of the concentration, with a peak at L = KD, but that it increases with the gradient steepness. Because this expression represents the mutual information between chemoattractant gradient and receptor occupancy, they termed this the external mutual information. They went on to compute the mutual information between receptor occupancy and the cell's directional response, which is the angle of cell motility (θr). Using experimentally observed tracks of cells moving in response to a gradient, they measured the direction of response and found that the mutual information between chemoattractant gradient and directional response peaks at about 0.5 bits and is relatively constant for gradients steeper than about 5 %. These results are similar to studies of axonal chemotaxis using a Bayesian model [62,63], which show that for very shallow gradients the response increases with gradient steepness but eventually saturates.
Rate distortion theory has also been used to compute the optimal stimulus response maps for chemotaxis . This optimal system was in strong agreement to that of a model used to explain the gradient sensing pathway of Dictyostelium cells (the so-called, local-excitation, global-inhibition model ). This study also demonstrated that the presence of prior information—as can be expected when cells are persistently moving in a stable chemoattractant gradient—improves chemotactic efficiency However, if the prior information is in conflict with the information from receptor occupancy (as might be the case if the direction of the gradient changes) then the cell trades off between the two sources of information. These results help to explain why polarized Dictyostelium cells perform gradual turns in response to changing gradient directions.
This review has highlighted some successes in understanding how biological systems can be studied using analysis and design tools from engineering. Our coverage is by no means exhaustive. Among the areas that we have not touched upon include tools developed to study biological reactions when the number of molecules is sufficiently small that continuum descriptions are not possible (see, for example, [64,65]), the development of algorithms for identifying parameters in biological networks (e.g., [66–68]), and the use of computational tools to guide synthetic biology (e.g., [69–71]).
The analogy between biological and engineering systems can sometimes be carried too far, as the general setting between the two is quite different. As an example, we turn to the standard paradigm of control systems that delineates between two subsystems: the plant, which is specific a priori and cannot be altered, and the controller, which is the system that is to be designed. This distinction is not necessarily possible in biological signaling. While we may be able to view certain aspects of biological systems as “controllers” (we do, after all, talk about regulation), most of the time we need to consider the system as a whole. Thus, straightforward application of these design tools may not be possible.
Where they are most suitable is when they highlight fundamental limitations and can thus be used to consider the performance of biological systems in relation to some optimal level of performance, or in light of tradeoffs that need to be made. Interestingly, this normative view of biology that specifies what a system “should do” is quite common in studying neural systems [72,73]. It is possible that the complexity of the neural system necessitates that systems be understood in the context of these optimality principles. In contrast, the development of mathematical models of cell signaling from known biochemistry is possible, and it has not been necessary to seek these underlying principles. The studies reviewed here show, however, that cell signaling does operate in this utility-maximizing regime. As the complexity of the signaling systems that we seek to understand increases, we can expect this point of view to gain greater prominence.
This work was supported in part by the National Institute of General Medical Sciences of the National Institutes of Health under award number R01GM86704. It was written while the author was a visiting scientist at the Max Planck Institute for the Physics of Complex Systems, in Dresden, Germany, whose support is greatly acknowledged. The author also thanks Geoff Goodhill who pointed him to some references on normative biology in neural systems.
This appendix provides some more details about the mathematical derivations as well as simulations.
A.1. Derivation of the Transfer Function
If the input and set-point are denoted by u0 and (x0,y0,z0), respectively, we compute the partial derivatives in the Jacobian  to getwhere Taking the Fourier transform of Equation 5 and collecting terms yield the transfer function To obtain the sensitivity function, we write H(iω) as where the first term is just H(iω) with the feedback term set to zero (k = 0).
If 1 − y0 ≫ kY and , then β ≈ 0, in which caseand
The simulations were computed using MATLAB (Mathworks, Natick, MA). The deterministic simulations of Figure 1(c) used the ode23.m solver, using parameters kUX = kXZ = 0.2s−1, k−X = k−Y = k−Z = 0.2s−1, , , and .
Stochastic simulations were carried out using Gillespie's stochastic simulation algorithm .
- Hartwell, L.H.; Hopfield, J.J.; Leibler, S.; Murray, A.W. From molecular to modular cell biology. Nature 1999, 402, 47–52, doi:10.1038/46972. 10573417
- Csete, M.E.; Doyle, J.C. Reverse engineering of biological complexity. Science 2002, 295, 1664–1669, doi:10.1126/science.1069981. 11872830
- Bennett, S. A History of Control Engineering, 1800–1930; Peregrinus: Stevenage, UK, 1979.
- Iglesias, P.; Ingalls, B. Control Theory and Systems Biology; MIT Press: Cambridge, MA, USA, 2010.
- Rato, C.; Amirova, S.R.; Bates, D.G.; Stansfield, I.; Wallace, H.M. Translational recoding as a feedback controller: Systems approaches reveal polyamine-specific effects on the antizyme ribosomal frameshift. Nucleic Acids Res. 2011, 39, 4587–4597, doi:10.1093/nar/gkq1349. 21303766
- Milias-Argeitis, A.; Summers, S.; Stewart-Ornstein, J.; Zuleta, I.; Pincus, D.; El-Samad, H.; Khammash, M.; Lygeros, J. In silico feedback for in vivo regulation of a gene expression circuit. Nat. Biotechnol. 2011, 29, 1114–1116, doi:10.1038/nbt.2018. 22057053
- Cosentino, C.; Salerno, L.; Passanti, A.; Merola, A.; Bates, D.G.; Amato, F. Structural bistability of the GAL regulatory network and characterization of its domains of attraction. J. Comput. Biol. 2012, 19, 148–162, doi:10.1089/cmb.2011.0251. 22300317
- Sheppard, P.W.; Rathinam, M.; Khammash, M. SPSens: A software package for stochastic parameter sensitivity analysis of biochemical reaction networks. Bioinformatics 2013, 29, 140–142, doi:10.1093/bioinformatics/bts642. 23104889
- Neuert, G.; Munsky, B.; Tan, R.Z.; Teytelman, L.; Khammash, M.; van Oudenaarden, A. Systematic identification of signal-activated stochastic gene regulation. Science 2013, 339, 584–587, doi:10.1126/science.1231456. 23372015
- Cooper, S.J. From claude bernard to walter cannon. Emergence of the concept of homeostasis. Appetite 2008, 51, 419–427, doi:10.1016/j.appet.2008.06.005. 18634840
- Cannon, W. The Wisdom of the Body; W.W. Norton: New York, NY, USA, 1932.
- Ma, W.; Trusina, A.; El-Samad, H.; Lim, W.A.; Tang, C. Defining network topologies that can achieve biochemical adaptation. Cell 2009, 138, 760–773, doi:10.1016/j.cell.2009.06.013. 19703401
- Åström, K.; Murray, R. Feedback Systems: An Introduction for Scientists and Engineers; Princeton University Press: Princeton, NJ, USA, 2008.
- Francis, B.; Wonham, W. Internal model principle of control-theory. Automatica 1975, 12, 457–465.
- Sontag, E. Adaptation and regulation with signal detection implies internal model. Syst. Control Lett. 2003, 50, 119–126, doi:10.1016/S0167-6911(03)00136-1.
- Alon, U.; Surette, M.G.; Barkai, N.; Leibler, S. Robustness in bacterial chemotaxis. Nature 1999, 397, 168–171, doi:10.1038/16483. 9923680
- Barkai, N.; Leibler, S. Robustness in simple biochemical networks. Nature 1997, 387, 913–917, doi:10.1038/43199. 9202124
- Yi, T.M.; Huang, Y.; Simon, M.I.; Doyle, J. Robust perfect adaptation in bacterial chemotaxis through integral feedback control. Proc. Natl. Acad. Sci. USA 2000, 97, 4649–4653, doi:10.1073/pnas.97.9.4649. 10781070
- Hohmann, S. Osmotic stress signaling and osmoadaptation in yeasts. Microbiol. Mol. Biol. Rev. 2002, 66, 300–372, doi:10.1128/MMBR.66.2.300-372.2002. 12040128
- Mettetal, J.T.; Muzzey, D.; Gomez-Uribe, C.; van Oudenaarden, A. The frequency dependence of osmo-adaptation in Saccharomyces cerevisiae. Science 2008, 319, 482–484, doi:10.1126/science.1151582. 18218902
- Muzzey, D.; Gomez-Uribe, C.A.; Mettetal, J.T.; van Oudenaarden, A. A systems-level analysis of perfect adaptation in yeast osmoregulation. Cell 2009, 138, 160–171, doi:10.1016/j.cell.2009.04.047. 19596242
- Mangan, S.; Alon, U. Structure and function of the feed-forward loop network motif. Proc. Natl. Acad. Sci. USA 2003, 100, 11980–11985, doi:10.1073/pnas.2133841100. 14530388
- Koshland, D.E. A response regulator model in a simple sensory system. Science 1977, 196, 1055–1063, doi:10.1126/science.870969. 870969
- Levchenko, A.; Iglesias, P.A. Models of eukaryotic gradient sensing: Application to chemotaxis of amoebae and neutrophils. Biophys. J. 2002, 82, 50–63, doi:10.1016/S0006-3495(02)75373-3. 11751295
- Iglesias, P.A.; Devreotes, P.N. Navigating through models of chemotaxis. Curr. Opin. Cell Biol. 2008, 20, 35–40, doi:10.1016/j.ceb.2007.11.011. 18207721
- Takeda, K.; Shao, D.; Adler, M.; Charest, P.G.; Loomis, W.F.; Levine, H.; Groisman, A.; Rappel, W.J.; Firtel, R.A. Incoherent feedforward control governs adaptation of activated ras in a eukaryotic chemotaxis pathway. Sci. Signal 2012, 5, ra2, doi:10.1126/scisignal.2002413.
- Krishnan, J.; Iglesias, P.A. Analysis of the signal transduction properties of a module of spatial sensing in eukaryotic chemotaxis. Bull. Math. Biol. 2003, 65, 95–128, doi:10.1006/bulm.2002.0323. 12597118
- Shoval, O.; Goentoro, L.; Hart, Y.; Mayo, A.; Sontag, E.; Alon, U. Fold-change detection and scalar symmetry of sensory input fields. Proc. Natl. Acad. Sci. USA 2010, 107, 15995–16000, doi:10.1073/pnas.1002352107. 20729472
- Bode, H. Network Analysis and Feedback Amplifier Design; D. Van Nostrand: New York, NY, USA, 1945.
- Serón, M.; Braslavsky, J.; Goodwin, G. Fundamental Limitations in Filtering and Control; Springer: London, UK, 1997.
- Iglesias, P. Tradeoffs in linear time-varying sysems: An analogue of Bode's sensitivity integral. Automatica 2001, 37, 1541–1550, doi:10.1016/S0005-1098(01)00103-0.
- Zang, G.; Iglesias, P. Nonlinear extension of Bode's integral based on an information-theoretic interpretation. Syst. Contr. Lett. 2003, 50, 11–19, doi:10.1016/S0167-6911(03)00119-1.
- Chandra, F.A.; Buzi, G.; Doyle, J.C. Glycolytic oscillations and limits on robust efficiency. Science 2011, 333, 187–192, doi:10.1126/science.1200705. 21737735
- Rao, C.V.; Wolf, D.M.; Arkin, A.P. Control, exploitation and tolerance of intracellular noise. Nature 2002, 420, 231–237, doi:10.1038/nature01258. 12432408
- Munsky, B.; Neuert, G.; van Oudenaarden, A. Using gene expression noise to understand gene regulation. Science 2012, 336, 183–187. 22499939
- Berg, H.C.; Purcell, E.M. Physics of chemoreception. Biophys. J. 1977, 20, 193–219, doi:10.1016/S0006-3495(77)85544-6. 911982
- Perkins, T.J.; Swain, P.S. Strategies for cellular decision-making. Mol. Syst. Biol. 2009, 5, 326, doi:10.1038/msb.2009.83.
- Libby, E.; Perkins, T.J.; Swain, P.S. Noisy information processing through transcriptional regulation. Proc. Natl. Acad. Sci. USA 2007, 104, 7151–7156, doi:10.1073/pnas.0608963104. 17420464
- Sartori, P.; Tu, Y. Noise filtering strategies in adaptive biochemical signaling networks: Application to E. coli chemotaxis. J. Stat. Phys. 2011, 142, 1206–1217, doi:10.1007/s10955-011-0169-z. 22977289
- Chou, C.S.; Bardwell, L.; Nie, Q.; Yi, T.M. Noise filtering tradeoffs in spatial gradient sensing and cell polarization response. BMC Syst. Biol. 2011, 5, 196, doi:10.1186/1752-0509-5-196.
- Strong, S.; Freedman, B.; Bialek, W.; Koberle, R. Adaptation and optimal chemotactic strategy in E. coli. Phys. Rev. E 1998, 57, 4604–4617.
- Samoilov, M.; Arkin, A.; Ross, J. Signal Processing by Simple Chemical Systems. J. Phys. Chem. A 2002, 106, 10205–10221, doi:10.1021/jp025846z.
- Andrews, B.W.; Yi, T.M.; Iglesias, P.A. Optimal noise filtering in the chemotactic response of Escherichia coli. PLoS Comput. Biol. 2006, 2, e154, doi:10.1371/journal.pcbi.0020154. 17112312
- Lan, G.; Sartori, P.; Neumann, S.; Sourjik, V.; Tu, Y. The energy-speed-accuracy tradeoff in sensory adaptation. Nat. Phys. 2012, 8, 422–428, doi:10.1038/nphys2276. 22737175
- Bialek, W. Physical limits to sensation and perception. Annu. Rev. Biophys. Biophys. Chem. 1987, 16, 455–478, doi:10.1146/annurev.bb.16.060187.002323. 3297091
- Kalman, R. A new approach to linear filtering and prediction problems. Trans. ASME—J. Basic Eng. Series D 1960, 82, 35–45, doi:10.1115/1.3662552.
- Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, doi:10.1002/j.1538-7305.1948.tb00917.x.
- Cover, T.; Thomas, J. Elements of Information Theory, 2nd. ed. ed.; Wiley: Hoboken, NJ, USA, 2006.
- Waltermann, C.; Klipp, E. Information theory based approaches to cellular signaling. Biochim. Biophys. Acta 2011, 1810, 924–932, doi:10.1016/j.bbagen.2011.07.009. 21798319
- Brennan, M.D.; Cheong, R.; Levchenko, A. Systems biology. How information theory handles cell signaling and uncertainty. Science 2012, 338, 334–335, doi:10.1126/science.1227946. 23087235
- Rhee, A.; Cheong, R.; Levchenko, A. The application of information theory to biochemical signaling systems. Phys. Biol. 2012, 9, 045011, doi:10.1088/1478-3975/9/4/045011.
- Tkačik, G.; Callan, C.G.; Bialek, W. Information flow and optimization in transcriptional regulation. Proc. Natl. Acad. Sci. USA 2008, 105, 12265–12270, doi:10.1073/pnas.0806077105. 18719112
- Hormoz, S. Cross talk and interference enhance information capacity of a signaling pathway. Biophys. J. 2013, 104, 1170–1180, doi:10.1016/j.bpj.2013.01.033. 23473500
- Janetopoulos, C.; Ma, L.; Devreotes, P.N.; Iglesias, P.A. Chemoattractant-induced phosphatidylinositol 3,4,5-trisphosphate accumulation is spatially amplified and adapts, independent of the actin cytoskeleton. Proc. Natl. Acad. Sci. USA 2004, 101, 8951–8956, doi:10.1073/pnas.0402152101. 15184679
- Andrews, B.W.; Iglesias, P.A. An information-theoretic characterization of the optimal gradient sensing response of cells. PLoS Comput. Biol. 2007, 3, e153, doi:10.1371/journal.pcbi.0030153. 17676949
- Cheong, R.; Rhee, A.; Wang, C.J.; Nemenman, I.; Levchenko, A. Information transduction capacity of noisy biochemical signaling networks. Science 2011, 334, 354–358, doi:10.1126/science.1204553. 21921160
- Lestas, I.; Vinnicombe, G.; Paulsson, J. Fundamental limits on the suppression of molecular fluctuations. Nature 2010, 467, 174–178, doi:10.1038/nature09333. 20829788
- Hamming, R. Error detecting and error correcting codes. Bell Syst. Tech. J. 1950, 29, 147–160, doi:10.1002/j.1538-7305.1950.tb00463.x.
- Porter, J.R.; Andrews, B.W.; Iglesias, P.A. A framework for designing and analyzing binary decision-making strategies in cellular systems. Integr. Biol. (Camb) 2012, 4, 310–317, doi:10.1039/c2ib00114d.
- Fuller, D.; Chen, W.; Adler, M.; Groisman, A.; Levine, H.; Rappel, W.J.; Loomis, W.F. External and internal constraints on eukaryotic chemotaxis. Proc. Natl. Acad. Sci. USA 2010, 107, 9656–9659, doi:10.1073/pnas.0911178107. 20457897
- Hu, B.; Chen, W.; Levine, H.; Rappel, W.J. Quantifying information transmission in eukaryotic gradient sensing and chemotactic response. J. Stat. Phys. 2011, 142, 1167–1186, doi:10.1007/s10955-011-0156-4. 21643513
- Mortimer, D.; Dayan, P.; Burrage, K.; Goodhill, G.J. Bayes-optimal chemotaxis. Neural Comput. 2011, 23, 336–373, doi:10.1162/NECO_a_00075. 21105826
- Yuan, J.; Chan, S.; Mortimer, D.; Nguyen, H.; Goodhill, G.J. Optimality and saturation in axonal chemotaxis. Neural Comput. 2013, 25, 833–853, doi:10.1162/NECO_a_00426. 23339614
- Thomas, P.; Matuschek, H.; Grima, R. Intrinsic noise analyzer: A software package for the exploration of stochastic biochemical kinetics using the system size expansion. PLoS One 2012, 7, e38518, doi:10.1371/journal.pone.0038518. 22723865
- Munsky, B.; Khammash, M. The finite state projection algorithm for the solution of the chemical master equation. J. Chem. Phys. 2006, 124, 044104:1–044104:13.
- Lillacci, G.; Khammash, M. Parameter estimation and model selection in computational biology. PLoS Comput. Biol. 2010, 6, e10000696.
- Porter, J.R.; Burg, J.S.; Espenshade, P.J.; Iglesias, PA. Identifying a static nonlinear structure in a biological system using noisy, sparse data. J. Theor. Biol. 2012, 300, 232–241, doi:10.1016/j.jtbi.2012.01.037. 22310068
- Neuert, G.; Munsky, B.; Tan, R.Z.; Teytelman, L.; Khammash, M.; van Oudenaarden, A. Systematic identification of signal-activated stochastic gene regulation. Science 2013, 339, 584–587, doi:10.1126/science.1231456. 23372015
- Del Vecchio, D.; Ninfa, A.J.; Sontag, E.D. Modular cell biology: Retroactivity and insulation. Mol. Syst. Biol. 2008, 4, 161–10.1038/msb4100204. 18277378
- Randall, A.; Guye, P.; Gupta, S.; Duportet, X.; Weiss, R. Design and connection of robust genetic circuits. Meth. Enzymol. 2011, 497, 159–186. 21601086
- Wu, K.; Rao, C.V. Computational methods in synthetic biology: Towards computer-aided part design. Curr. Opin. Chem. Biol. 2012, 16, 318–322, doi:10.1016/j.cbpa.2012.05.003. 22727029
- Kording, K. Decision theory: What “should” the nervous system do? Science 2007, 318, 606–610, doi:10.1126/science.1142998. 17962554
- Doyle, J.C.; Csete, M. Architecture, constraints, and behavior. Proc. Natl. Acad. Sci. USA 2011, 108 Suppl 3, 15624–15630, doi:10.1073/pnas.1103557108. 21788505
- Gillespie, D.T. Stochastic simulation of chemical kinetics. Annu. Rev. Phys. Chem. 2007, 58, 35–55, doi:10.1146/annurev.physchem.58.032806.104637. 17037977
© 2013 by the author; licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution license (http://creativecommons.org/licenses/by/3.0/).