Resources
The latest articles and
press releases on Groq
Groq Sets New Large Language Model Performance Record of 300 Tokens per Second per User on Meta AI Foundational LLM, Llama-2 70B
The Groq Language Processing Unit™ system is the AI assistance enablement technology poised to provide real-time, “low lag” experiences for users with its inference performance.
Groq to Feature World’s Fastest GenAI Inference Performance for Foundational LLMs at Supercomputing ’23 on Its LPU™ Systems
Groq and their team will be showcasing a demo of the world’s best low latency performance for Large Language Models (LLMs) running on a Language
Argonne Deploys New Groq System to ALCF AI Testbed, Providing AI Accelerator Access to Researchers Globally
Groq, an artificial intelligence (AI) solutions company, and the US Department of Energy’s (DOE) Argonne National Laboratory announced today that Groq hardware is now available
Groq to Showcase World’s Fastest Large Language Model Performance, Powered by Its LPU™ System, at the Global Emerging Technology Summit in Washington, DC
Groq, an AI solutions company announced today a record-breaking AI processing demo, powered by the ultra-low latency performance of their LPU™ system, to be delivered
Groq Smashes LLM Performance Record Again Using an LPU™ System With No Response From GPU Companies
Groq, an artificial intelligence (AI) solutions provider, today announced it has more than doubled its inference performance of the Large Language Model (LLM), Llama-2 70B, in
Groq Selects Samsung Foundry to Bring Next-gen LPU™ to the AI Acceleration Market
Groq, an artificial intelligence (AI) inference systems innovator, today announced it has contracted with Samsung’s growing Foundry business to be its next-gen silicon partner, solidifying
Groq’s Record-breaking LPU™ Hits 100 Tokens Per Second Per User On A Massive AI Model
Groq’s newly announced language processor, the Groq LPU, has demonstrated that it can run 70-billion-parameter enterprise-scale language models at a record speed of more than
Groq™ First to Achieve 100 Tokens Per Second Per User on Meta AI’s Llama-2 70B, Leading All Artificial Intelligence Solutions Providers in Inference Performance
Groq, an artificial intelligence (AI) solutions provider, today announced it now runs the Large Language Model (LLM), Llama-2 70B, at more than 100 tokens per second
AI Accelerator Groq™ Adapts and Runs LLaMA, the Meta™ Chatbot Model and Competitor to ChatGPT, for Its Systems
Facebook® parent, Meta, released LLaMA, which can be used by chatbots to generate human-like text, on February 24th. Three days later the Groq team downloaded the
Groq adapts Meta’s chatbot for its own chips in race against Nvidia
Groq modified LLaMA, a large language model released last month by Facebook parent Meta Platforms Inc that can be used to power bots to generate