Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I believe almost all LLMs are trained using wikpedia these days. So compressing wikipedia well without including the size of the LLM in the compression result is a bit of a cheat. I guess one would argue it is a universal dataset representing understanding the English language and real-world relationships at this point but it is still a bit of a cheat.


There's a reason compression benchmarks often times include the size of the executable when benchmarking compression ratios. Although Matt Mahoney's large text compression benchmark[0] does currently have a transformer model at number 1.

[0] http://www.mattmahoney.net/dc/text.html


Which is also made by the same author as ts_zip (Fabrice Bellard): https://bellard.org/nncp/


Is there anything this man can’t do?




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: