Entropy Measures for Data Analysis
Title | Entropy Measures for Data Analysis PDF eBook |
Author | Karsten Keller |
Publisher | MDPI |
Pages | 260 |
Release | 2019-12-19 |
Genre | Science |
ISBN | 3039280325 |
Entropies and entropy-like quantities play an increasing role in modern non-linear data analysis. Fields that benefit from this application range from biosignal analysis to econophysics and engineering. This issue is a collection of papers touching on different aspects of entropy measures in data analysis, as well as theoretical and computational analyses. The relevant topics include the difficulty to achieve adequate application of entropy measures and the acceptable parameter choices for those entropy measures, entropy-based coupling, and similarity analysis, along with the utilization of entropy measures as features in automatic learning and classification. Various real data applications are given.
Entropy Measures for Data Analysis: Theory, Algorithms and Applications
Title | Entropy Measures for Data Analysis: Theory, Algorithms and Applications PDF eBook |
Author | Karsten Keller |
Publisher | |
Pages | 260 |
Release | 2019 |
Genre | Engineering (General). Civil engineering (General) |
ISBN | 9783039280339 |
Entropies and entropy-like quantities play an increasing role in modern non-linear data analysis. Fields that benefit from this application range from biosignal analysis to econophysics and engineering. This issue is a collection of papers touching on different aspects of entropy measures in data analysis, as well as theoretical and computational analyses. The relevant topics include the difficulty to achieve adequate application of entropy measures and the acceptable parameter choices for those entropy measures, entropy-based coupling, and similarity analysis, along with the utilization of entropy measures as features in automatic learning and classification. Various real data applications are given.
Entropy Measures, Maximum Entropy Principle and Emerging Applications
Title | Entropy Measures, Maximum Entropy Principle and Emerging Applications PDF eBook |
Author | Karmeshu |
Publisher | Springer |
Pages | 300 |
Release | 2012-10-01 |
Genre | Technology & Engineering |
ISBN | 3540362126 |
The last two decades have witnessed an enormous growth with regard to ap plications of information theoretic framework in areas of physical, biological, engineering and even social sciences. In particular, growth has been spectac ular in the field of information technology,soft computing,nonlinear systems and molecular biology. Claude Shannon in 1948 laid the foundation of the field of information theory in the context of communication theory. It is in deed remarkable that his framework is as relevant today as was when he 1 proposed it. Shannon died on Feb 24, 2001. Arun Netravali observes "As if assuming that inexpensive, high-speed processing would come to pass, Shan non figured out the upper limits on communication rates. First in telephone channels, then in optical communications, and now in wireless, Shannon has had the utmost value in defining the engineering limits we face". Shannon introduced the concept of entropy. The notable feature of the entropy frame work is that it enables quantification of uncertainty present in a system. In many realistic situations one is confronted only with partial or incomplete information in the form of moment, or bounds on these values etc. ; and it is then required to construct a probabilistic model from this partial information. In such situations, the principle of maximum entropy provides a rational ba sis for constructing a probabilistic model. It is thus necessary and important to keep track of advances in the applications of maximum entropy principle to ever expanding areas of knowledge.
Proceedings of the 4th International Conference on Electronics, Biomedical Engineering, and Health Informatics
Title | Proceedings of the 4th International Conference on Electronics, Biomedical Engineering, and Health Informatics PDF eBook |
Author | Triwiyanto Triwiyanto |
Publisher | Springer Nature |
Pages | 690 |
Release | |
Genre | |
ISBN | 981971463X |
Neutrosophic Entropy Measures For The Normal Distribution: Theory And Applications
Title | Neutrosophic Entropy Measures For The Normal Distribution: Theory And Applications PDF eBook |
Author | Rehan Ahmad Khan Sherwani |
Publisher | Infinite Study |
Pages | 16 |
Release | |
Genre | Mathematics |
ISBN |
Entropy is a measure of uncertainty and often used in information theory to determine the precise testimonials about unclear situations. Different entropy measures available in the literature are based on the exact form of the observations and lacks in dealing with the interval-valued data. The interval-valued data often arises from the situations having ambiguity, imprecise, unclear, indefinite, or vague states of the experiment and is called neutrosophic data. In this research modified forms of different entropy measures for normal probability distribution have been proposed by considering the neutrosophic form data. The performance of the proposed neutrosophic entropies for normal distribution has been assessed via a simulation study. Moreover, the proposed measures are also applied to two real data sets for their wide applicability.
Entropy, Search, Complexity
Title | Entropy, Search, Complexity PDF eBook |
Author | Imre Csiszár |
Publisher | Springer Science & Business Media |
Pages | 262 |
Release | 2007-04-05 |
Genre | Mathematics |
ISBN | 3540327770 |
This book collects survey papers in the fields of entropy, search and complexity, summarizing the latest developments in their respective areas. More than half of the papers belong to search theory which lies on the borderline of mathematics and computer science, information theory and combinatorics, respectively. The book will be useful to experienced researchers as well as young scientists and students both in mathematics and computer science.
Universal Estimation of Information Measures for Analog Sources
Title | Universal Estimation of Information Measures for Analog Sources PDF eBook |
Author | Qing Wang |
Publisher | Now Publishers Inc |
Pages | 104 |
Release | 2009-05-26 |
Genre | Computers |
ISBN | 1601982305 |
Entropy, mutual information and divergence measure the randomness, dependence and dissimilarity, respectively, of random objects. In addition to their prominent role in information theory, they have found numerous applications, among others, in probability theory statistics, physics, chemistry, molecular biology, ecology, bioinformatics, neuroscience, machine learning, linguistics, and finance. Many of these applications require a universal estimate of information measures which does not assume knowledge of the statistical properties of the observed data. Over the past few decades, several nonparametric algorithms have been proposed to estimate information measures. Universal Estimation of Information Measures for Analog Sources presents a comprehensive survey of universal estimation of information measures for memoryless analog (real-valued or real vector-valued) sources with an emphasis on the estimation of mutual information and divergence and their applications. The book reviews the consistency of the universal algorithms and the corresponding sufficient conditions as well as their speed of convergence. Universal Estimation of Information Measures for Analog Sources provides a comprehensive review of an increasingly important topic in Information Theory. It will be of interest to students, practitioners and researchers working in Information Theory