Insights
For general press inquiries, reach out to our PR team.

Delivering Fast Inference with the Full 131k Context Window GroqCloud™ now supports Qwen3 32B, a cutting-edge, dense 32.8 billion parameter causal language model from Alibaba’s Qwen3 series. This integration brings the power of Qwen3 32B’s advanced multilingual capabilities to GroqCloud, enabling businesses to leverage complex reasoning and efficient dialogue across...

GroqCloud now supports Low-Rank Adaptation (LoRA) fine-tunes, exclusively by request, for our Enterprise tier customers. LoRA enables businesses to deploy adaptations of base models customized to their specific use cases on GroqCloud, offering a more efficient and cost-effective approach to model customization. As a part of this release, we are...

You know Groq runs small models. But did you know we run large models including MoE uniquely well? Here’s why. The Evolution of Advanced Openly-Available LLMs There’s no argument that Artificial intelligence (AI) has exploded, in part because of the advancements in large language models (LLMs). These models have shown...