Hacker Newsnew | past | comments | ask | show | jobs | submit | sci-genie's commentslogin

TBH I kinda agree with the argument that distributed training is too hard. Its so architecture/compute-resources/network-topology dependent that when people open those can of worms, they quickly realize that the cost/benefit tradeoff is limited unless you are doing large-scale pre-training. its just so much easier to train as much as possible on a single node


how do I save a HN comment ? Someone give this person a medal!


Click the time it was posted (e.g. “2 hours ago”) and then “favorite”


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: