- From: <noah_mendelsohn@us.ibm.com>
- Date: Sat, 16 Oct 2004 16:04:21 -0400
- To: Dan Connolly <connolly@w3.org>
- Cc: www-tag@w3.org
- Message-ID: <OF400788F0.79C42DFB-ON85256F2F.006E3215@lotus.com>
Just to be clear: I am not expert in information theory and it's possible that Shannon's is not quite the right work to reference. Whatever the reference may be, the concept I have in mind is not statistical. One bit can convey a choice between two states. Two bits a choice between four states. Given some context, then I can associate some of these states with real world choices such as the choice between possible sequences of Unicode characters, countable numbers, etc. My probably over-simplified assumption is that Shannon's work allows one to say: given a channel with certain characteristics, what is the statistical probability that I can convey with full fidelity a message of n-bits. Perhaps I have misunderstood. In any case, as I've said before, I am NOT suggesting that we refer to the statistical aspects of Shannon's theory, which are in fact its essence. I am suggesting that we refer to what I had presumed was a precondition for applying his theory: I.e. that one has in hand a set of perfectly determined bits that we might wish to call information. This is surely closely associated with the "entropy" referenced in the Wiki entry, but I too have insufficient depth in this area to fill in the blanks. If I have misused Shannon's terminology or misunderstood the theory please accept my apologies for any confusion caused. Perhaps someone more knowledgeable in this area can help us out. -------------------------------------- Noah Mendelsohn IBM Corporation One Rogers Street Cambridge, MA 02142 1-617-693-4036 -------------------------------------- Dan Connolly <connolly@w3.org> Sent by: www-tag-request@w3.org 10/15/04 06:47 PM To: www-tag@w3.org cc: (bcc: Noah Mendelsohn/Cambridge/IBM) Subject: citing Shannon on Information Theory Further to... "NM suggested that Claude Shannon's work on information theory might provide a suitable definition" -- http://www.w3.org/2001/tag/2004/10/05-07-tag#infores2 I'm still not to the point where I understand well enough to cite Shannon's work, I think I know what the canonical work is now: "It is generally believed that the modern discipline of information theory began with the publication of Shannon's article "The Mathematical Theory of Communication" in the Bell System Technical Journal in July and October of 1948." -- http://en.wikipedia.org/wiki/Information_theory I'm slowly swapping in that part of wikipedia. -- Dan Connolly, W3C http://www.w3.org/People/Connolly/
Attachments
- application/octet-stream attachment: signature.asc
Received on Saturday, 16 October 2004 20:07:06 UTC