Hacker News new | past | comments | ask | show | jobs | submit login

Hi there, I work for Groq. That's right. We love graphics processors for training but for inference our language processor (LPU) is by far the fastest and lowest latency. Feel free to ask me anything.



What's the scale of hardware behind this demo, in terms of watts, transistors and cost?


Are they only available on the cloud? Are you planning on releasing a consumer version?


Mostly available as a service via cloud API at the moment. The systems themselves are too big for consumers but we will sell systems to corporations.




Consider applying for YC's Spring batch! Applications are open till Feb 11.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: