Batch Processing with GroqCloud™ for AI Inference Workloads

Written by:
Groq
Share:

GroqCloud™ provides fast inference for complex AI solutions that require instant responsiveness. But what happens when your use cases expand and require features beyond speed? That’s where Batch Processing comes in –now you can use GroqCloud at scale to process massive workloads.

Let’s say you have some large-scale datasets you want to analyze or summarize. Or you have a large set of images that need captioning. Or run validation and quality tests on a new AI program. The GroqCloud Batch Processing API, available to Developer and Enterprise Tier customers, the perfect solution for these use cases and more. 

Batch Processing allows users to batch together non-time sensitive requests or submit large scale workloads and get a response back within 24 hours. For bulk processing made easy, it’s perfect for tasks like large scale data classification, translations, document summaries, and image to text workloads. All at a 25% discount to normal on-demand pricing without taxing rate limits. 

Experience a broader range of batch processing models with newly added support for Llama 3.3 70B, DeepSeek-R1-Distill-Llama-70B, and Llama 3.2 90B Vision models. And, for a limited time, our Batch Processing offering can help you build fast for even less! 

50% Discount Through April – Double the Savings!

Now through the end of April, we’re doubling our discount on Batch Processing to 50% off for all paid GroqCloud customers! Whether you’re running large-scale chat completions or handling massive transcription tasks, now’s the perfect time to scale up.

If you’re not on Developer Tier already, upgrade today. Need help getting started? Check out our reference docs.

New Features for Whisper Transcription Batch Processing

We’ve enhanced our Whisper transcription capabilities, adding support for Batch Processing alongside powerful new features:

  • All Three Whisper Versions Available for Batch Processing: Choose the model that best fits your needs across Whisper Large v3, Distill-Whisper, and Whisper Large v3 Turbo.
  • Word-Level Timestamps: Get precise timing with timestamps for every word for audio  transcriptions, ideal for YouTube videos, captions, and social media. Read more in our blog here.
  • Streamlined Workflows: Perfect for bulk video/audio transcriptions and generating subtitles.

While pipelining between Whisper and chat completions Batch Processing isn’t currently supported, you can easily run two separate batch workloads to achieve your desired results.

Build Fast, Now!

Access our Batch Processing features and enjoy the 50% discount for all Batch Processing workloads through the end of April. Whether you’re transcribing audio, translating huge documents, or summarizing large texts, GroqCloud has you covered.

The GroqCloud Batch Processing API is yet another feature we developed in response to the needs of our users. As you move from building AI solutions to launching and scaling them, GroqCloud is committed to providing you with the best tools to accelerate your work. Stay tuned for more exciting updates!

The latest Groq news. Delivered to your inbox.