All Updates

All Updates

icon
Filter
Product updates
Groq achieves over 300 tokens per second per user on Meta AI’s Llama-2 70B
Generative AI Infrastructure
Nov 7, 2023
This week:
Partnerships
T-Mobile partners with OpenAI to develop AI-powered customer service platform
Generative AI Applications
Today
Partnerships
Runway partners with Lionsgate to develop AI video tools using studio's movie catalog
Generative AI Applications
Yesterday
Funding
QMill raises EUR 4 million in seed funding to provide quantum computing industrial applications
Quantum Computing
Yesterday
Product updates
QuiX Quantum launches 'Bia' quantum cloud computing service for quantum solutions
Quantum Computing
Yesterday
Partnerships
Oxford Ionics and Infineon Technologies partner to build portable quantum computer for Cyberagentur
Quantum Computing
Yesterday
Partnerships
Product updates
Tencent Ai Lab launches EzAudio AI for text-to-audio generation with Johns Hopkins University
Foundation Models
Yesterday
Funding
TON secures USD 30 million in investment from Bitget and Foresight Ventures
Web3 Ecosystem
Yesterday
Funding
Hemi Labs raises USD 15 million in funding to launch blockchain network
Web3 Ecosystem
Yesterday
Product updates
Fivetran launches Hybrid Deployment for data pipeline management
Machine Learning Infrastructure
Yesterday
Product updates
Fivetran launches Hybrid Deployment for data pipeline management
Data Infrastructure & Analytics
Yesterday
Generative AI Infrastructure

Generative AI Infrastructure

Nov 7, 2023

Groq achieves over 300 tokens per second per user on Meta AI’s Llama-2 70B

Product updates

  • Groq, an AI solutions company, achieved over 300 tokens per second per user on Meta AI's Llama-2 70B LLM using its Language Processing Unit™ (LPU) system.

  • Groq's LPU system addresses LLMs' sequential and compute-intensive nature, offering ultra-low latency crucial for natural conversation rhythm in AI interfaces; traditional solutions like GPUs fall short in handling incumbent latency and scale-related issues.

  • Groq is an AI solutions company and the inventor of the Language Processing Unit accelerator, which is purpose-built and software-driven to power large language models (LLMs) for the exploding AI market.

Contact us

Gain access to all industry hubs, market maps, research tools, and more
Get a demo
arrow
menuarrow

By using this site, you agree to allow SPEEDA Edge and our partners to use cookies for analytics and personalization. Visit our privacy policy for more information about our data collection practices.