Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

This "experiment" is garbage. No description of test methodology other than "then outputted the compressed file back to plain text."

Last time I had to use Omnipage to convert some highly compressed .JPEGs, it blew me away with it's accuracy. Not only did it get the text nearly 100%, it often got the tables right too. The only thing it struggled with was advanced mathematical symbols.



The methodology seems obvious - interpret every three bytes in the text as a pixel in an image, compress the image, then reinterpret each (post-compression) pixel as three ASCII characters.

This is not about OCR.


Thanks. Great explanation. Considering that the ASCII character map is arbitrary and there are so many great ways to test lossy compression algorithms( PSNR to A/B testing), this seems like a pointless exercise, but at least I see what the author was trying to do now.

If I wanted to demonstrate lossy compression, I'd use Image Subtraction plugin in GIMP, not ASCII.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: