Entropy vs. Energy Waveform Processing: A Comparison Based on the Heat Equation
AbstractVirtually all modern imaging devices collect electromagnetic or acoustic waves and use the energy carried by these waves to determine pixel values to create what is basically an “energy” picture. However, waves also carry “information”, as quantified by some form of entropy, and this may also be used to produce an “information” image. Numerous published studies have demonstrated the advantages of entropy, or “information imaging”, over conventional methods. The most sensitive information measure appears to be the joint entropy of the collected wave and a reference signal. The sensitivity of repeated experimental observations of a slowly-changing quantity may be defined as the mean variation (i.e., observed change) divided by mean variance (i.e., noise). Wiener integration permits computation of the required mean values and variances as solutions to the heat equation, permitting estimation of their relative magnitudes. There always exists a reference, such that joint entropy has larger variation and smaller variance than the corresponding quantities for signal energy, matching observations of several studies. Moreover, a general prescription for finding an “optimal” reference for the joint entropy emerges, which also has been validated in several studies. View Full-Text
Share & Cite This Article
Hughes, M.S.; McCarthy, J.E.; Bruillard, P.J.; Marsh, J.N.; Wickline, S.A. Entropy vs. Energy Waveform Processing: A Comparison Based on the Heat Equation. Entropy 2015, 17, 3518-3551.
Hughes MS, McCarthy JE, Bruillard PJ, Marsh JN, Wickline SA. Entropy vs. Energy Waveform Processing: A Comparison Based on the Heat Equation. Entropy. 2015; 17(6):3518-3551.Chicago/Turabian Style
Hughes, Michael S.; McCarthy, John E.; Bruillard, Paul J.; Marsh, Jon N.; Wickline, Samuel A. 2015. "Entropy vs. Energy Waveform Processing: A Comparison Based on the Heat Equation." Entropy 17, no. 6: 3518-3551.