Chris Messina

Groq® - Hyperfast LLM running on custom built GPUs

An LPU Inference Engine, with LPU standing for Language Processing Unit™, is a new type of end-to-end processing unit system that provides the fastest inference at ~500 tokens/second.

Add a comment

Replies

Best
Congrats team Groq® on your launch.
It looks very promising. How can I find information on how to use the APIs?
Aris Nakos
Wow, you guys are innovating. Congratulations! I tested it out and was blown away.
Daniel Rödler
Wow, love it. We are heavily relying on LLMs and the slowness of our agents is a constant annoyance. A 14x speed-up would be a real game changer. Can't wait to see LPUs in action and at scale. Keep going!
Mona Dey
This is helpful post.thanks