Hacker News new | past | comments | ask | show | jobs | submit login

That’s 260 gigabyte, that’s actually a tiny dataset, you can query and index that on a normal workstation in seconds.

Even training models on that is possible in realistic times on normal systems with that.




[deleted due to wrongness]


Which one is which? Thanks in advance if you know, because i'd rather not download a huge torrent just to find out I should have been using that bandwidth to download a different one.


Never mind, misread, the comment data is indeed what I linked to.

However, uncompressed, it hits over 1+TB, if the BigQuery sizes are indication.


> However, uncompressed, it hits over 1+TB, if the BigQuery sizes are indication.

Even then, that’s easily doable on a consumer system.

I’ll download it in the night between friday and saturday, after I install my new HDD, and just run queries over it for fun. (far slower, but also far cheaper than BigQuery. Even at German electricity prices).




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: