"This internal AI project will power generative creation on our platform. Our 3D foundational model will be open source and multimodal, and it will power 3D generation through text, video, and 3D prompts. We see a powerful future where Roblox experiences will have extensive generative AI capabilities to power real-time creation integrated with gameplay. We’ll provide these capabilities in a resource-efficient way, so we can make them available to everyone on the platform."
Chub AI | REMOTE | Mid-level frontend engineer | Full-time or Contract | TypeScript
We are the largest generative AI gaming platform, entirely self-funded and profitable since August 2023.
Our frontend is React TypeScript, and the bulk of the backend APIs are Flask. We’re looking for someone with a relevant degree and a few years of experience, for exclusively or mostly frontend work. An eye for design and aesthetics is needed as we have none. Ever having worked with Ant Design, Capacitor, or react-three-fiber would be a big plus. Your most likely first project would be a better UI for our image and video generation API.
How to Apply: Send however you prefer to represent yourself — a resume, a site, a GitHub, all or none of the above — along with salary expectations to hiring@chub.ai.
Chub AI | REMOTE | Mid-level frontend engineer | Full-time or Contract | TypeScript
We are the largest generative AI gaming platform, entirely self-funded and profitable since August 2023.
Our frontend is React TypeScript, and the bulk of the backend APIs are Flask. We’re looking for someone with a relevant degree and/or a few years of experience, for exclusively or mostly frontend work. An eye for design and aesthetics is needed as we have none. Ever having worked with Ant Design, Capacitor, or react-three-fiber would be a big plus. Your most likely first project would be a better UI for our image and video generation API.
How to Apply:
Send however you prefer to represent yourself — a resume, a site, a GitHub, all or none of the above — along with salary expectations to hiring@chub.ai.
Very, very recently. In the past few days. I tried it out immediately because GPTQ-for-LLaMA and hunting for or making quantized models can be tedious, but it was disappointingly slow. On a 3090 where I was getting responses for a given 13B model in 10-30 seconds, just using transformers with load_in_4bit took about ten times that for each response.
There’s also the storage benefit of using actually quantized models.