Skip to content
Graphcore IPU Server

IPU Systems

Designed from the ground up for machine intelligence

Get started
 

Introducing the world's first Graphcore IPU Server

The Intelligence Processing Unit (IPU) has been designed from the ground up by Graphcore to support new breakthroughs in machine intelligence. Together with our production ready Poplar® software stack, it gives developers a powerful, efficient, scalable, and high performance solution which enables new innovations in AI. Customers can tackle their most difficult AI workloads by accelerating more complex models and developing entirely new techniques.

Bert-2

Natural Language Processing - BERT

We have achieved state of the art performance and accuracy with the BERT language model, training BERT Base in 56 hours with seven C2 IPU-Processor PCIe cards, each with two IPUs, in an IPU Server system. With BERT inference, we see 3x higher throughput with over 20% improvement in latency to serve up results faster than ever.

ResNext (1)-1

Image Classification - ResNext

Graphcore C2 IPU-Processor PCIe card achieves 3.4x higher throughput at 20x lower latency compared to a leading alternative processor. High throughput at the lowest possible latency is key in many of the important use cases today.

Faster Time to Results

Delivering a new level of fine-grained, parallel processing across thousands of independent processing threads on each individual IPU. The whole machine intelligence model is held inside the IPU with In-Processor Memory to maximise memory bandwidth and deliver high throughput for faster time to train and the lowest latency inference.

Enabling innovation

State of the art performance with today's large language models and conventional CNNs and dramatic breakthroughs with new higher accuracy models, like ResNext and probabilistic systems. Legacy processors struggle with non-aligned and sparse data accesses which are critical for next generation models. The IPU has been designed to support complex data access efficiently and at much higher speeds.

Training and Inference Flexibility

High performance training and low latency inference on the same hardware, improving utilisation and flexibility in the cloud and on-premise, vastly improving the total cost of ownership.

Designed for the Future

The IPU is designed to scale. Models are getting larger and demand for AI compute is scaling exponentially. High bandwidth IPU-Links™ allow multiple IPUs to be clustered, supporting huge models. Legacy architectures struggle on non-aligned and sparse data accesses. The IPU has been designed to support complex data access efficiently and at much higher speeds, which will be critical to run gigantic, next generation models efficiently

Azure-1

Azure IPU Preview

The Graphcore IPU preview on Microsoft Azure is now open for customer sign-up. We will prioritize customers focused on pushing the boundaries of NLP and developing new breakthroughs in machine intelligence. 

Register now
DSS-8440

Dell IPU Server

The Dell EMC DSS 8440 machine intelligence server with Graphcore technology is designed to address the needs of enterprise customers building out on-premise machine intelligence compute.

Register now
cirrascale_offeringblock

Cirrascale IPU Cloud

IPUs are offered in two different ways from Cirrascale: as part of an IPU bare-metal cloud service or with the Dell EMC DSS8440 machine intelligence server with Graphcore technology for on-premise customer applications

Register now
IPU

The Intelligence Processing Unit (IPU)

Learn more
Poplar1

Poplar® Software Stack

Learn more

Get the latest Graphcore news

Subscribe now