Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Spark is going to try and ingest all the data, and it won’t fit in RAM. Wrong tool for the job basically.


Depends on exactly how you do it, I suppose, but it shouldn't necessarily. Most Hadoop-y work can also be accomplished in Spark without much fuss.




Consider applying for YC's Winter 2026 batch! Applications are open till Nov 10

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: