.

Technology

LPU

The Language Processing Unit (LPU) is Groq's specialized processor: a custom-built chip architected for deterministic, high-speed, low-latency inference on Large Language Models (LLMs).

The LPU (Language Processing Unit) is a purpose-built accelerator from Groq, designed to resolve the compute and memory bandwidth bottlenecks that plague LLM inference on traditional hardware. Unlike general-purpose GPUs, the LPU utilizes a software-defined, single-core architecture with hundreds of megabytes of integrated SRAM, which acts as primary weight storage (not cache). This design enables deterministic performance and superior sequential processing, a critical factor for language models. The result is industry-leading speed and low latency: benchmarks confirm the LPU Inference Engine can achieve throughputs of up to 241 tokens per second, making real-time, large-scale generative AI applications practical for deployment.

https://groq.com/
1 project · 1 city

Related technologies

Recent Talks & Demos

Showing 1-1 of 1

Members-Only

Sign in to see who built these projects