Learning Entropy: Multiscale Measure for Incremental Learning
AbstractFirst, this paper recalls a recently introduced method of adaptive monitoring of dynamical systems and presents the most recent extension with a multiscale-enhanced approach. Then, it is shown that this concept of real-time data monitoring establishes a novel non-Shannon and non-probabilistic concept of novelty quantification, i.e., Entropy of Learning, or in short the Learning Entropy. This novel cognitive measure can be used for evaluation of each newly measured sample of data, or even of whole intervals. The Learning Entropy is quantified in respect to the inconsistency of data to the temporary governing law of system behavior that is incrementally learned by adaptive models such as linear or polynomial adaptive filters or neural networks. The paper presents this novel concept on the example of gradient descent learning technique with normalized learning rate. View Full-Text
A printed edition of this Special Issue is available here.
Share & Cite This Article
Bukovsky, I. Learning Entropy: Multiscale Measure for Incremental Learning. Entropy 2013, 15, 4159-4187.
Bukovsky I. Learning Entropy: Multiscale Measure for Incremental Learning. Entropy. 2013; 15(10):4159-4187.Chicago/Turabian Style
Bukovsky, Ivo. 2013. "Learning Entropy: Multiscale Measure for Incremental Learning." Entropy 15, no. 10: 4159-4187.