Entropic Measures, Markov Information Sources and Complexity

Show simple item record

dc.contributor.author Calude, C.S en
dc.contributor.author Dumitrescu, M en
dc.date.accessioned 2009-04-16T23:12:23Z en
dc.date.available 2009-04-16T23:12:23Z en
dc.date.issued 2001-02 en
dc.identifier.citation CDMTCS Research Reports CDMTCS-150 (2001) en
dc.identifier.issn 1178-3540 en
dc.identifier.uri http://hdl.handle.net/2292/3658 en
dc.description.abstract The concept of entropy plays a major part in communication theory. The Shannon entropy is a measure of uncertainty with respect to a priori probability distribution. In algorithmic information theory the information content of a message is measured in terms of the size in bits of the smallest program for computing that message. This paper discusses the classical entropy and entropy rate for discrete or continuous Markov sources, with finite or continuous alphabets, and their relations to program-size complexity and algorithmic probability. The accent is on ideas, constructions and results; no proofs will be given. en
dc.publisher Department of Computer Science, The University of Auckland, New Zealand en
dc.relation.ispartofseries CDMTCS Research Report Series en
dc.rights.uri https://researchspace.auckland.ac.nz/docs/uoa-docs/rights.htm en
dc.source.uri http://www.cs.auckland.ac.nz/staff-cgi-bin/mjd/secondcgi.pl?serial en
dc.title Entropic Measures, Markov Information Sources and Complexity en
dc.type Technical Report en
dc.subject.marsden Fields of Research::280000 Information, Computing and Communication Sciences en
dc.rights.holder The author(s) en
dc.rights.accessrights http://purl.org/eprint/accessRights/OpenAccess en


Files in this item

Find Full text

This item appears in the following Collection(s)

Show simple item record

Share

Search ResearchSpace


Browse

Statistics