Groq Chat
p/groq-chat
An LPU inference engine
Chris Messina
Groq® — Hyperfast LLM running on custom built GPUs
Featured
17
An LPU Inference Engine, with LPU standing for Language Processing Unit™, is a new type of end-to-end processing unit system that provides the fastest inference at ~500 tokens/second.
Replies
Johan Steneros
It is fast, that is for sure. Where can I get more information about the chips and hardware? Is there a GPU cloud service?
Cuong Vu
Groq is a promising product, and I believe your detailed insights could attract even more supporters, helping people better understand its value.
Peter Schout
Congratulations! speed/accuracy is incredible, no wonder NVDA took a dip 😯
Abhilash Chowdhary
Going to give this a try, team Groq®. Looks interesting.
Borja Soler
this will be incredible for the future of LLMs and all the products benefiting from them. super excited with all the new things that will come
Congrats team Groq® on your launch.
Kien Nguyen
Congrat on the launch? Do you have any plan when to support custom training?
Ian Nance
Man, that IS fast...Already loving it : )
Sourabh
Don't know why its ranked so low as of now, the speed is awesome. It does what it says.
Ivan Somov
Good luck! I am really excited about this hardware stuff for LLMs!
Daniel Rödler
Wow, love it. We are heavily relying on LLMs and the slowness of our agents is a constant annoyance. A 14x speed-up would be a real game changer. Can't wait to see LPUs in action and at scale. Keep going!
Mona Dey
This is helpful post.thanks
Aris Nakos
Wow, you guys are innovating. Congratulations! I tested it out and was blown away.
ROHIT GUPTA
amazing...
Junior Perassoli
It looks very promising. How can I find information on how to use the APIs?
Avi Basnet
This seems extremely interesting- I’m curious what you’ve seen to be the biggest use case for this LLM?