Information theory (nonfiction)

From Gnomon Chronicles
Revision as of 17:59, 10 September 2016 by Admin (talk | contribs) (Created page with "'''Information theory''' studies the quantification, storage, and communication of information. It was originally proposed by Claude Shannon i...")
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to navigation Jump to search

Information theory studies the quantification, storage, and communication of information.

It was originally proposed by Claude Shannon in 1948 to find fundamental limits on signal processing and communication operations such as data compression, in a landmark paper entitled "A Mathematical Theory of Communication".

Now this theory has found applications in many other areas, including statistical inference, natural language processing, cryptography, neurobiology, the evolution and function of molecular codes, model selection in ecology, thermal physics, quantum computing, linguistics, plagiarism detection, pattern recognition, and anomaly detection.

The field is at the intersection of mathematics, statistics, computer science, physics, neurobiology, and electrical engineering.

A key measure in information theory is entropy. Entropy quantifies the amount of uncertainty involved in the value of a random variable or the outcome of a random process. For example, identifying the outcome of a fair coin flip (with two equally likely outcomes) provides less information (lower entropy) than specifying the outcome from a roll of a die (with six equally likely outcomes).

Some other important measures in information theory are mutual information, channel capacity, error exponents, and relative entropy.

Fiction cross-reference

Nonfiction cross-reference

External links: