Groq is Fast AI Inference

0 K+
Developers using GroqCloud™
since Feb ‘24 launch

Instant Intelligence

Fast AI inference for openly-available models like Llama 3.1
I’m really excited to see Groq’s ultra-low-latency inference for cloud deployments of the Llama 3.1 models. This is an awesome example of how our commitment to open source is driving innovation and progress in AI…
Mark Zuckerberg
Founder & CEO, Meta

Use Groq, Fast.

Move seamlessly to Groq from other providers like OpenAI by changing three lines of code.

 
  1. With our OpenAI endpoint compatibility, simply set OPENAI_API_KEY to your Groq API Key.
  2. Set the base URL.
  3. Choose your model and run!

Groq Speed Is Instant

Don’t just take our word for it. Independent benchmarks from Artificial Analysis prove Groq Speed is instant for foundational openly-available models.

Groq Powers Leading
Openly-available AI Models

Llama
Mixtral
Gemma
Whisper
The Groq chip really goes for the jugular.
Yann LeCun
VP & Chief AI Scientist at Meta, Groq Technical Advisor

Groq in the News

Never miss a Groq update! Sign up below for our latest news.