The Shift to AI Inference

With the seismic shift in AI toward deploying or running models – known as inference – developers and enterprises alike can experience instant intelligence with Groq. We provide fast AI inference in the cloud and in on-prem AI compute centers. We power the speed of iteration, fueling a new wave of innovation, productivity, and discovery. Groq was founded in 2016 to build technology to advance AI because we saw this moment coming. 

Delivering Fast AI Inference with the LPU

The Groq Language Processing Unit, the LPU, is the technology that meets this moment. The LPU delivers instant speed, unparalleled affordability, and energy efficiency at scale. Fundamentally different from the GPU – originally designed for graphics processing – the LPU was designed for AI inference and language.

Making AI Accessible to All

Groq technology can be accessed by anyone via GroqCloud™, while enterprises and partners can choose between cloud or on-prem AI compute center deployment. We are committed to deploying millions of LPUs, providing access to the value of AI to the world. Our first-generation LPU is available and shipping today, with more innovations to come.

Exceptional Talent

At our core, Groq is made up of the best technology and the best people. We’re looking for exceptionally talented individuals who share our passion for pushing the boundaries of AI technology. In fact, our name comes from the word “grok” which means the ability “to understand profoundly and intuitively.” This is at the core of every Groqster – operating at the highest level of curiosity, courage, and humility. Learn more about a Groq career re.

The Latest on Groq

Never miss a Groq update! Sign up below for our latest news.