Compress Wikipedia, Win 20,000 Euros

Wed 09 August 2006

tags: ML

Brilliant! The Hutter Prize for Lossless Compression ( takes as its challenge the task of compressing 100MB of Wikipedia text into the pre-competition best of \~17MB. The idea is that a chunk of Wikipedia text that big has characteristics relevant to compression that go beyond statistical analysis (i.e., "meaning"). The deliverable must be entirely self-contained, but it can be near 17MB in size in order to get in the money, so that's a lot of space for generative code (there are no restrictions on runtime speed or memory consumption).