gc-webinar-ipu-tech

IPU Technology Webinar

Explore new ways to make machine intelligence breakthroughs

Watch On-Demand
 

The Dell DSS8440 Graphcore IPU Server

The Intelligence Processing Unit (IPU) has been designed from the ground up to support new breakthroughs in machine intelligence. Together with our production ready Poplar® software stack, it gives developers a powerful, efficient, scalable, and high performance solution which enables new innovations in AI. Customers can tackle their most difficult AI workloads by accelerating more complex models and developing entirely new techniques.

Read the product brief
Azure-1

Azure IPU Preview

The Graphcore IPU preview on Microsoft Azure is now open for customers focused on developing new breakthroughs in machine intelligence. 

Register now
DSS-8440

Dell IPU Server

The Dell EMC DSS 8440 machine intelligence server with Graphcore technology for enterprise customers building out on-premise AI compute.

Register now
cirrascale_offeringblock

Cirrascale IPU Cloud

Cirrascale offers an IPU bare-metal cloud service and Dell EMC DSS8440 IPU servers for sale for on-premise customer applications

Buy now
EfficientNet Training Click to Zoom

Computer Vision – EfficientNet

Graphcore IPU delivers 5x higher throughput on the standard EfficientNet-B0 model compared to a leading alternative GPU at equivalent power. This advantage extends to 7x with a modified version of the EfficientNet model which further showcases the benefit of the unique architecture of the IPU.

ResNext Click to Zoom

Computer Vision - ResNext

Graphcore C2 IPU-Processor PCIe card achieves 7x higher throughput at 24x lower latency compared to a leading alternative processor. High throughput at the lowest possible latency is key in many of the important use cases today.

Faster Time to Results

Delivering a new level of fine-grained, parallel processing across thousands of independent processing threads on each individual IPU. The whole machine intelligence model is held inside the IPU with In-Processor Memory to maximise memory bandwidth and deliver high throughput for faster time to train and the lowest latency inference.

Enabling innovation

State of the art performance with today's large language models and conventional CNNs and dramatic breakthroughs with new higher accuracy models, like ResNext and probabilistic systems. Legacy processors struggle with non-aligned and sparse data accesses which are critical for next generation models. The IPU has been designed to support complex data access efficiently and at much higher speeds.

Training and Inference Flexibility

High performance training and low latency inference on the same hardware, improving utilisation and flexibility in the cloud and on-premise, vastly improving the total cost of ownership.

Designed for the Future

The IPU is designed to scale. Models are getting larger and demand for AI compute is scaling exponentially. High bandwidth IPU-Links™ allow multiple IPUs to be clustered, supporting huge models. Legacy architectures struggle on non-aligned and sparse data accesses. The IPU has been designed to support complex data access efficiently and at much higher speeds, which will be critical to run gigantic, next generation models efficiently

IPU

The Intelligence Processing Unit (IPU)

Learn more
Poplar1

Poplar® Software Stack

Learn more
×