{"id":672,"date":"2008-08-18T11:17:17","date_gmt":"2008-08-18T11:17:17","guid":{"rendered":"http:\/\/scientopia.org\/blogs\/goodmath\/2008\/08\/18\/why-is-randomness-informative-correcting-the-clueless-creationists\/"},"modified":"2008-08-18T11:17:17","modified_gmt":"2008-08-18T11:17:17","slug":"why-is-randomness-informative-correcting-the-clueless-creationists","status":"publish","type":"post","link":"http:\/\/www.goodmath.org\/blog\/2008\/08\/18\/why-is-randomness-informative-correcting-the-clueless-creationists\/","title":{"rendered":"Why is randomness informative? Correcting the Clueless Creationists"},"content":{"rendered":"<p> Someone sent me some questions about information theory; or actually,<br \/>\nabout some questions raised by <a href=\"http:\/\/triablogue.blogspot.com\/2008\/08\/complex-sleight-of-hand.html\"> a bozo creationist arguing about information<br \/>\ntheory.<\/a> But the issues raised are interesting. So while I&#8217;m<br \/>\nnominally snarking at a clueless creationist, I&#8217;m really using it<br \/>\nas an excuse to talk about one of my favorite mathematical subjects.<\/p>\n<p> The creationist is confused by the information theory idea of information and complexity. That <em>is<\/em> somewhat confusing.<br \/>\nIntuitively, we think of information in terms of semantics and communication. We think that information is related to semantic content. And our understanding of semantic content comes from language. So we expect<br \/>\nthings that are rich in information to be highly structured, like language.<\/p>\n<p><!--more--><\/p>\n<p> But our initial intuition is wrong. wrong. As I explained <a href=\"http:\/\/scientopia.org\/blogs\/goodmath\/2006\/06\/an-introduction-to-information-theory-updated-from-blogspot\">when I was writing about information theory<\/a>, a totally unstructured random string of bits has a lot more information than an equal-length string of english text.<br \/>\nWhat we see as structure is really redundancy &#8211; multiple expressions of the same bits of information. <\/p>\n<p> One of the things that I find particularly interesting about information theory is that it&#8217;s an area where many of the basic ideas can be<br \/>\nexpressed in intuitive ways. There is a good intuition for why randomness has high information content &#8211; and once you hear it, it makes perfect sense, and you can easily see what&#8217;s wrong with the argument that structure implies high information content. And  the intuition is actually very nearly a direct rendering of the theoretical (Kolmogorov-Chaitin) definition of information!<\/p>\n<p> Here goes:<\/p>\n<p> Imagine you have a string of symbols. (In information theory, we describe everything as strings of symbols.) Now suppose you want<br \/>\nto tell someone else how to write down the same string of symbols. The<br \/>\namount of information in the string is the length of <em>the shortest description<\/em> you can give them that allows them to write down the same string.<\/p>\n<p> Think of a simple string like &#8220;1111111111111111&#8221;. To tell someone how to write that string, you just say &#8220;Write 16 ones&#8221;. For a slightly more interesting string, like &#8220;1010101010101010&#8221; &#8211; you could say &#8220;write &#8217;10&#8217; 8 times.&#8221;. For &#8220;ABCDEFGHIJKLMNOP&#8221;, you could say &#8220;write the first sixteen letters of the alphabet&#8221;. For each of those, there&#8217;s a nice simple description, because they&#8217;ve got a lot of structure. But that structure means that it doesn&#8217;t take much to describe them; and if there&#8217;s a short<br \/>\ndescription, then that means that there&#8217;s not a lot of information.<\/p>\n<p> Now think of a random string: &#8220;XEGAPC]GI(BAUL+&#8221;? How could you tell<br \/>\nsomeone to write down that string of symbols? There&#8217;s some amount of pattern there &#8211; the letters are all uppercase; there are only 13 distinct symbols out of the sixteen. But you&#8217;re not going to get away with a simple description. It&#8217;s harder to give someone instructions on how to reproduce that <em>exact<\/em> string than it is to describe a nicely structured string &#8211; which is why we say that the random string has more information in it.<\/p>\n<p> Think of a simple image: a triangle, drawn onto a 1000&#215;1000 bitmap. To allow someone to produce an identical image, you&#8217;d say &#8220;A 1000&#215;1000 bitmap, with a triangle with vertices (x1,y1), (x2,y2), (x3,y3)&#8221;. The bit map has<br \/>\na million bits. The description above would be about 90 characters long with real vertex positions, which is 720 bits &#8211; using an <em>extremely<\/em> inefficient representation of the text. Because it has structure, it takes very little description to allow someone to reproduce it exactly.<\/p>\n<p> Now, think of a random image. For every pixel in the image, you use<br \/>\na truly random process to determine whether it&#8217;s black or white. So you&#8217;ve got something that looks like a screen capture of static on an old TV. How can you describe that image to someone else to allow them to produce an exact duplicate? You might be able to do it with less that a million<br \/>\nbits &#8211; but you&#8217;re not going to be able to do it with less that 1,000 bits. There&#8217;s just too much information in the image!<\/p>\n<p> The hardest thing to describe &#8211; the string that needs the longest<br \/>\ndescription for its length &#8211; is something with absolutely no pattern.<br \/>\nSomething totally random, where there&#8217;s absolutely no pattern, no logic, no<br \/>\nway of predicting even a single bit without being told its value explicitly.<br \/>\nA string with maximum information context is a string where it&#8217;s so random<br \/>\nthat the complete string itself is its own shortest description. And the<br \/>\nstring with the most information is, by definition, the most complex. The<br \/>\nreason why should be absolutely clear by now: it takes longer to describe<br \/>\nsomething complex than something simple, and random stuff is the hardest to<br \/>\ndescribe. So random stuff has the most information.<\/p>\n<p> One last point, and I&#8217;ll get to a couple of quotes. The word &#8220;random&#8221; means something very specific in the context of information theory. A perfectly random string is a string where knowledge of what came before in the string gives you <em>absolutely<\/em> no knowledge of what comes later. It&#8217;s closely related to the probabilistic notion of random. The main difference is that from the viewpoint of probability, a string is &#8220;random&#8221; if it was generated by a random process &#8211; and a random process can (and in fact by necessity <em>will<\/em>) periodically generate a string that we wouldn&#8217;t call random. For example, a truly random letter generator, given enough time, will eventually create a string consisting of the letters of the alphabet in order. But from the viewpoint of information theory, if we looked at that string, we&#8217;d say it&#8217;s low information content, meaning non-random. On the other hand, if we take a highly compressed version of<br \/>\nthe bitmap of a triangle, it <em>looks<\/em> random &#8211; it&#8217;s got very high<br \/>\ninformation content for its size per information theory, but according<br \/>\nto probability theory, it isn&#8217;t random, because it&#8217;s generated by a deterministic process. The big difference here is that when information<br \/>\ntheory talks about randomness, it&#8217;s talking about unpredictability for someone viewing a string &#8211; it&#8217;s not talking about meaning or semantics. You can take any meaningful statement, and render it down to a high information content string via compression. High information content versus low information content isn&#8217;t a property of the statement or entity being encoded; it&#8217;s a property of the representation. An image of a triangle<br \/>\nis a representation of some information in a very inefficient form; a string of random-looking bits is a representation of <em>the same information<\/em> in a more compact form. The image is low information content because it&#8217;s got very little information <em>for its size<\/em>; the compressed bits are high information content because they&#8217;ve got a lot of information for their size. But it&#8217;s the same information.<\/p>\n<p> Now, let&#8217;s look at what the creationist in question actually had to say, so that we can see where he went wrong. The context was one of the theism\/atheism arguments. Theists argue that the universe\/life\/whatever is too complex to be the product of a random process, therefore there must be a God. The atheists respond that it&#8217;s just a pushback: if the universe\/life is too complex to come into existence itself, then how is it possible for an omniscient being to have come into being by himself? <\/p>\n<p> In the course of this discussion, the argument came around to<br \/>\nthe creationist pulling out one of the old canards about information theory,<br \/>\nto which the atheist explained that information theory really says<br \/>\nthat &#8220;structure&#8221; isn&#8217;t information &#8211; that randomness is really information.<\/p>\n<p> I&#8217;m going to stay away from the theism\/atheism aspect of this,<br \/>\nbecause it&#8217;s a tired old discussion, and I&#8217;ve got absolutely nothing to add. I&#8217;m a religious Jew, but I find the religious argument from complexity<br \/>\nto be unbearably stupid.<\/p>\n<p> I&#8217;m also going to take things a bit out of order, because<br \/>\nthere&#8217;s a stupid thing that I want to get out of the way before<br \/>\nI get to the interesting part.<\/p>\n<blockquote><p>\nFWIW, I disagree with T-Stone&#8217;s version of information and complexity. And despite what his post would lead you to believe, the idea that &#8220;maximal randomness = maximal complexity&#8221; is not true for all information theories. And in fact, if I were to use T-Stone&#8217;s definition of complexity then I would ask him to explain not why there is so much complexity in the universe, but rather why there is so little complexity. If complexity = randomness, then it doesn&#8217;t take a rocket scientist to realize that there&#8217;s a lot of the universe that is not random, and therefore there is a lot of this universe that is not complex. Under his information theory, randomness is the default. We do not need to explain random data. We do need to explain structured and ordered data. Therefore, we do not need to explain complexity; we need to explain non-complexity.\n<\/p><\/blockquote>\n<p> This isn&#8217;t something where there&#8217;s room for disagreement. This is sort<br \/>\nof like arguing that &#8220;I don&#8217;t agree that 2+2=4.&#8221; Informational complexity is<br \/>\nwell-defined by information, and it&#8217;s got a precise meaning. The precise<br \/>\ndefinitions vary between algorithmic information theory (Kolmogorov-Chaitin)<br \/>\nand communication information theory (Shannon), but the basic concept<br \/>\nunderlying both is the same, and they agree that complexity is related to<br \/>\ninformation content, and maximum information content (and thus maximum complexity) is perfect randomness.<\/p>\n<p> <em><b>There is no information theory that says randomness doesn&#8217;t<br \/>\nmaximize information content and complexity. None! <\/b><\/em> This<br \/>\nis something that you see frequently from the clueless about information theory: they really don&#8217;t like the idea that randomness contains<br \/>\nmaximum information, and they <em>assert<\/em> that not all information<br \/>\ntheory agrees with that &#8211; like the statement above&#8221; maximal randomness = maximal complexity is not true for all information theories&#8221;; but they<br \/>\n<em>never<\/em> actually cite any kind of information theory at all &#8211; because there is none that does what they want. They&#8217;re sure that there must be,<br \/>\nbecause K\/C and Shannon seem wrong. But there is no such theory, no<br \/>\nmatter how much you may want one to exist.<\/p>\n<p> Now, we&#8217;ll jump backwards to the interesting part.<\/p>\n<blockquote>\n<p>Unfortunately for T-Stone, if he paid attention to what he has written here he&#8217;d see that he&#8217;s soundly refuted Dawkins. After all, if maximal randomness is equivalent to maximal complexity, then it is easy for me to write a program that will generate completely random output. In other words, it is easy for me&#8211;a person who is not maximally complex&#8211;to produce a program with output that is maximally complex. Thus, if we want to play T-Stone&#8217;s game and use complexity in this sense, then Dawkin&#8217;s argument must be surrendered.<\/p>\n<p>If I can make a program that is more complex than I am, then God can create a universe that is more complex than He is.<\/p>\n<\/blockquote>\n<p> Our creationist friend is making several mistakes here. One of them is his claim that he can <em>easily<\/em> write a program that generates<br \/>\na random string.  Generating random data &#8211; true random data, not just data that appears random &#8211; is really difficult. What he&#8217;s thinking about, I&#8217;m almost sure, is running a program that calls a function like the<br \/>\nC &#8220;srandom&#8221; function. That&#8217;s <em>not<\/em> random; it&#8217;s completely deterministic. Start with a particular seed, and you&#8217;ll always get the<br \/>\nsame sequence of &#8220;random&#8221; numbers. In fact, there&#8217;s <em>only one<\/em><br \/>\nsequence of values that you&#8217;ll get: the function is a long-cycle<br \/>\nperiodic function. Every sequence generated using that function is a part of the same, long, fixed cycle. And you can perfectly reproduce any string<br \/>\nthat you generated in it with a simple set of instructions. The<br \/>\noriginal pseudo-random function used by &#8220;srandom&#8221;, in the Berkeley Unix C library is <em>one line long<\/em>; it&#8217;s basically the next &#8220;random&#8221; number is &#8220;<code>(lastrandom*1103515245L + 12345L) &amp; 0x7fffffff<\/code>&#8220;. So given that line of code, and the initial seed, you can generate the &#8220;random&#8221; sequence. So our friend&#8217;s &#8220;random&#8221; sequence is anything but random. It&#8217;s got a nicely random <em>distribution<\/em>, but it&#8217;s not random &#8211; it&#8217;s reproducible with a very short sequence of instructions. (Even if you use a better pseudo-random function, it goes up to around 64 lines of code for a really nice one.) Computers are deterministic: they <em>suck<\/em> at randomness. In fact true randomness must come from a data source outside of the computer. Real randomness comes from random external events, like radioactive decay. Building a true random generator is a very non-trivial task.<\/p>\n<p> That&#8217;s thoroughly clueless. But it&#8217;s not the worst of his errors. He&#8217;s confused about two really important things about how you do things in information theory. First, he&#8217;s confused about what we mean in information theory by having a program generate a string; and second, he&#8217;s confused about how to compare the complexity of two different strings.<\/p>\n<p> In information theory, when we talk about a program or a set of<br \/>\ninstructions for generating a string, we&#8217;re talking about a program that<br \/>\n<em>repeatably<\/em> and <em>deterministically<\/em> generates that string.<br \/>\nWhat makes a high complexity string is the difficulty of<br \/>\n<em>reproducing<\/em> it, not of <em>producing<\/em> it. If I&#8217;ve got a true<br \/>\nrandomness source, I can generate a random string from it with very little<br \/>\ndifficulty. But that&#8217;s not the important thing, from the viewpoint of<br \/>\ninformation theory. What information theory cares about is <em>how can I<br \/>\ndescribe the string in order to allow it to be reproduced.<\/em> If he&#8217;s<br \/>\nwritten a short program to generate a string that has high information<br \/>\ncontent, that&#8217;s not very impressive, and it says nothing about information<br \/>\ntheory. The short program that generates the random string doesn&#8217;t mean that<br \/>\nthe random string is easy to reproduce, or even that the string is really<br \/>\nproduceable by a short program. Because when information theory says a<br \/>\nstring is produceable by a short program, what it means is that it&#8217;s<br \/>\nproduceable by a short <em>deterministic<\/em> program. Run a random program<br \/>\ntwice, it will generate two different strings! So from the viewpoint of<br \/>\ninformation theory, the string isn&#8217;t a product of the program; it&#8217;s the<br \/>\nproduct of randomness which was the input to the program, and that says<br \/>\nnothing about its information content. What matters is how long a program<br \/>\nthat will <em>always<\/em> generate the random string, and <em>only<\/em> the<br \/>\nrandom string. <\/p>\n<p> Completely random strings are incredibly rare in the real world. Almost any finite string that you can imagine is at least somewhat compressible &#8211; which means that it&#8217;s got some amount of redundant structure. If you think of it by metaphor with numbers, theoretically, the vast majority of numbers are irrationals. But in the real physical world, encountering irrational<br \/>\nnumbers isn&#8217;t particularly common. Most things that we really encounter<br \/>\nhave finite precision. We don&#8217;t see perfect circles that really have a perfect ratio of &pi; between their measured circumference and diameter. We don&#8217;t see triangles so perfect that their edges have a 1\/1\/sqrt(2) ratio. We see finite approximations. Similarly, we don&#8217;t often see true randomness.<\/p>\n<p> Our friend doesn&#8217;t get that. And that&#8217;s a huge problem for his<br \/>\nargument.<\/p>\n<p> The second mistake in in his idea of comparison. Even if I could write a<br \/>\nprogram that generates terabytes of randomness, it won&#8217;t be as complex as<br \/>\na human being. Because for all of the structure that there is in us, for<br \/>\nall of the redundancy, for all the duplication, we are enormously<br \/>\ncomplex. To create a string of information that provided you<br \/>\nwith enough information to be able to accurately reproduce the precise<br \/>\nstate of my body at a moment in time is really mind-boggling. It&#8217;s<br \/>\nan amount of information that&#8217;s difficult to conceive. The error that<br \/>\nour friend is making is forgetting that complexity isn&#8217;t a percentage. A<br \/>\nstring that is 80% random (meaning 20% compressible) can be less complex<br \/>\nthan a string that is 10% random &#8211; provided the second one is<br \/>\nlonger. If the 80% random string is 1,000 characters long, and<br \/>\nthe 10% random string is 1,000,000 characters long, then the 10% random<br \/>\nstring is more complex!<\/p>\n<p> Of course, that&#8217;s not the end of his cluelessness. Following<br \/>\ndirectly on that error, he finishes up with:<\/p>\n<blockquote>\n<p> T-Stone is just giving a sleight of hand here. It would be like a mathematician saying &#8220;a &gt; b&#8221; and having T-Stone say, &#8220;The greater than sign is inverted with the less than sign, therefore &#8216;a &gt; b&#8217; means &#8216;a is less than b&#8217;.&#8221;<\/p>\n<p> But as soon as he engages in his sleight of hand, we respond: &#8220;If the greater than sign is inverted with the less than sign, then &#8216;a &gt; b&#8217; is no longer true, rather &#8216;a &lt; b&#8217; is.&#8221;<\/p>\n<p> Inverting the operator without inverting the operands does not refute the original expression.<\/p>\n<\/blockquote>\n<p> What he&#8217;s going on about here is based on his misunderstanding of what<br \/>\ncomplexity means in terms of information. He <em>wants<\/em> structured strings to have more information than random. He&#8217;s insisting that that&#8217;s the way things <em>really<\/em> work. And so, by his reasoning, an argument that<br \/>\nrandomness has more information than structure is reversing the correct comparison &#8211; more structure should be more complexity, so if you&#8217;re making an argument that something needs to have more complexity, it necessarily<br \/>\n<em>must<\/em> be an argument that it has more structure. If you insist that<br \/>\nrandomness is complex, then you&#8217;re reversing the argument: every statement that you make arguing for complexity needs to be switched to an argument<br \/>\nfor simplicity, because you&#8217;ve swapped things around.<\/p>\n<p> Unfortunately for him, the fact that he&#8217;s completely clueless changes<br \/>\nnothing about the nature of reality. Randomness <em>is<\/em> more<br \/>\ncomplex than structure. And when we talk about how complex the universe<br \/>\nis, we&#8217;re using <em>the same<\/em> definition of complexity. It doesn&#8217;t work the way he thinks it does &#8211; he&#8217;s just plain wrong.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Someone sent me some questions about information theory; or actually, about some questions raised by a bozo creationist arguing about information theory. But the issues raised are interesting. So while I&#8217;m nominally snarking at a clueless creationist, I&#8217;m really using it as an excuse to talk about one of my favorite mathematical subjects. The creationist [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"jetpack_post_was_ever_published":false,"_jetpack_newsletter_access":"","_jetpack_dont_email_post_to_subs":false,"_jetpack_newsletter_tier_id":0,"_jetpack_memberships_contains_paywalled_content":false,"_jetpack_memberships_contains_paid_content":false,"footnotes":"","jetpack_publicize_message":"","jetpack_publicize_feature_enabled":true,"jetpack_social_post_already_shared":false,"jetpack_social_options":{"image_generator_settings":{"template":"highway","default_image_id":0,"font":"","enabled":false},"version":2}},"categories":[16,30],"tags":[],"class_list":["post-672","post","type-post","status-publish","format-standard","hentry","category-debunking-creationism","category-information-theory"],"jetpack_publicize_connections":[],"jetpack_featured_media_url":"","jetpack_shortlink":"https:\/\/wp.me\/p4lzZS-aQ","jetpack_sharing_enabled":true,"jetpack_likes_enabled":true,"_links":{"self":[{"href":"http:\/\/www.goodmath.org\/blog\/wp-json\/wp\/v2\/posts\/672","targetHints":{"allow":["GET"]}}],"collection":[{"href":"http:\/\/www.goodmath.org\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"http:\/\/www.goodmath.org\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"http:\/\/www.goodmath.org\/blog\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"http:\/\/www.goodmath.org\/blog\/wp-json\/wp\/v2\/comments?post=672"}],"version-history":[{"count":0,"href":"http:\/\/www.goodmath.org\/blog\/wp-json\/wp\/v2\/posts\/672\/revisions"}],"wp:attachment":[{"href":"http:\/\/www.goodmath.org\/blog\/wp-json\/wp\/v2\/media?parent=672"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"http:\/\/www.goodmath.org\/blog\/wp-json\/wp\/v2\/categories?post=672"},{"taxonomy":"post_tag","embeddable":true,"href":"http:\/\/www.goodmath.org\/blog\/wp-json\/wp\/v2\/tags?post=672"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}