Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

> There are no great FaaS options for running GPU workloads, which is why we could never go fully FaaS.

I keep wondering when this is going to show up. We have a lot of service providers, but even more frameworks, and every vendor seems to have their own bespoke API.



I don’t think anybody should go “fully FaaS”, it’s like saying screwdrivers are useless, all you need is a hammer.

That being said, Cloudflare is on the path to offering a great GPU FaaS system for inference.

I believe it’s still in beta, but it’s the most promising option at the moment.


Right, I still find it faster to manually provision a specific instance type, install PyTorch on it, and deploy a little flask app for an inference server.


Check out beam.cloud. They’re impressing me with calling GPU runtimes as a FaaS


I just started playing with modal.com and so far it seems good. I haven't run anything in production yet, so YMMV.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: