Home » Archive

Articles Archive for June 2010

Data compression, Headline »

[5 Jun 2010 by Keshav Shetty | 7 Comments | 6,001 views]
Random data compression – One bit diff encoding

This is continuation of my previous article on Random data compression-How to use merge sort?.
As I mentioned earlier – it is hard or impossible(as of now) to compress million random digit. As per my analysis million random digit file contains around 90 – 110 bytes duplicates or repeated number within every block of 256 bytes. When input data is very pure or near pure(I mean uniqueness) or when input data is highly polluted or noise(I mean duplicates), then it is easy to compress. May be we can borrow the idea …

Data compression, Headline »

[3 Jun 2010 by Keshav Shetty | 7 Comments | 19,071 views]
Random data compression – Is it possible? How to use merge sort?

This is continuation of my previous article on Random data compression possibilities.
Some of the readers asked how reverse merge sort (merge unsort) can be used to represent 256 unique values using 128bytes(best case) or 224.25(worst case).
As illustrated in previous article using merge sort we sort the random input and store bit information of the list from where we picked the smaller number. Actually we are reshuffling the original position and stored bit information represents how a input changed its position from original list to sorted list. Lets take an example …