EDGE Insights

EDGE Insights

icon
Filter
Older updates:
Generative AI Infrastructure

Generative AI Infrastructure

Groq: Designing specialized processors and firmware for AI workloads

Groq specializes in developing AI inference chips and cloud services designed to accelerate AI model deployment and execution. Groq's flagship product is its language processing unit (LPU), a custom chip architecture optimized for running large language models (LLMs) and other AI workloads at high speeds. It offers its users hardware and software solutions, including GroqChip processors, GroqCard accelerators, GroqNode servers, GroqRack compute clusters, and GroqWare suites.
  • Groq's LPU technology takes a different approach compared to traditional GPUs, using on-chip SRAM memory and a software-defined architecture to achieve faster processing speeds for AI inference tasks. The company claims its chips can run LLMs up to 10x faster than competing solutions while using less energy. Groq provides access to its tensor streaming processor (TSP) through GroqCloud, a cloud service that allows developers to run LLMs like Llama.
  • In August 2024, Groq raised USD 640 million to expand the capacity of its services and introduce new models and features to its platform.
  • In March 2024, Groq acquired Definitive Intelligence to expand its cloud platform and provide on-demand access to its AI chips.

Contact us

Gain access to all industry hubs, market maps, research tools, and more
Get a demo
arrow
menuarrow

By using this site, you agree to allow SPEEDA Edge and our partners to use cookies for analytics and personalization. Visit our privacy policy for more information about our data collection practices.