News & Press
Groq is on a mission to set the standard for GenAI inference speed, helping real-time AI applications come to life today.
For general press inquiries, reach out to our PR team.
ArtificialAnalysis.ai Adjusts Chart Axes to Accommodate Groq Performance Levels MOUNTAIN VIEW, CA, February 13, 2024 – Groq®, a generative AI solutions company, is the clear
Groq® Opens API Access to Real-time Inference, the Magic Behind Instant Responses from Generative AI Products
Customer and Partner aiXplain Implements Game-changing Groq Technology to Bring the World’s Fastest AI Language Processing for Consumer Electronics to Market LAS VEGAS, CES® 2024,
Groq Sets New Large Language Model Performance Record of 300 Tokens per Second per User on Meta AI Foundational LLM, Llama-2 70B
The Groq Language Processing Unit™ system is the AI assistance enablement technology poised to provide real-time, “low lag” experiences for users with its inference performance.
Groq to Feature World’s Fastest GenAI Inference Performance for Foundational LLMs at Supercomputing ’23 on Its LPU™ Systems
Groq and their team will be showcasing a demo of the world’s best low latency performance for Large Language Models (LLMs) running on a Language
Argonne Deploys New Groq System to ALCF AI Testbed, Providing AI Accelerator Access to Researchers Globally
Groq, an artificial intelligence (AI) solutions company, and the US Department of Energy’s (DOE) Argonne National Laboratory announced today that Groq hardware is now available
Groq to Showcase World’s Fastest Large Language Model Performance, Powered by Its LPU™ System, at the Global Emerging Technology Summit in Washington, DC
Groq, an AI solutions company announced today a record-breaking AI processing demo, powered by the ultra-low latency performance of their LPU™ system, to be delivered
Groq, an artificial intelligence (AI) solutions provider, today announced it has more than doubled its inference performance of the Large Language Model (LLM), Llama-2 70B, in
Groq, an artificial intelligence (AI) inference systems innovator, today announced it has contracted with Samsung’s growing Foundry business to be its next-gen silicon partner, solidifying
Groq’s newly announced language processor, the Groq LPU, has demonstrated that it can run 70-billion-parameter enterprise-scale language models at a record speed of more than
Groq™ First to Achieve 100 Tokens Per Second Per User on Meta AI’s Llama-2 70B, Leading All Artificial Intelligence Solutions Providers in Inference Performance
Groq, an artificial intelligence (AI) solutions provider, today announced it now runs the Large Language Model (LLM), Llama-2 70B, at more than 100 tokens per second