News:

PD.com: Taoism in a clown costume.

Main Menu

ATTN 000: information theory nerdspaggery

Started by Rococo Modem Basilisk, March 26, 2010, 08:18:54 PM

Previous topic - Next topic

Rococo Modem Basilisk

I wrote a short python script to generate a first order markov model of a document (tokenizes by whitespace), and after each token use the equation i=log2(anbn-1/an-1bn) to figure out the difference in information from one token to the next in all of the token pairs in the model.

I ran it on the first 4539 words in the phrack archives, and used google docs to graph it:


Edit: Whoops! I forgot to mention: I model the HMM in terms of ratios a:b, where a is the frequency of a given pair, and b is the frequency of all pairs with the same first token. an would be the frequency of the pair after the current token has been added, and an-1 would be its frequency before it has been added.

I can post the code up in a bit.
Edit: here it is


I am not "full of hate" as if I were some passive container. I am a generator of hate, and my rage is a renewable resource, like sunshine.

Triple Zero

I saw you post that image on twitter, wondered what it was about.

I have alcohol in my head right now, I will check this out later.
Ex-Soviet Bloc Sexual Attack Swede of Tomorrow™
e-prime disclaimer: let it seem fairly unclear I understand the apparent subjectivity of the above statements. maybe.

INFORMATION SO POWERFUL, YOU ACTUALLY NEED LESS.