California-based Cerebras Systems has launched the Wafer Scale Engine 3 (WSE-3), an AI chip that delivers twice the performance of its 2021 predecessor the WSE-2.
The 5nm-based, 4 trillion transistor WSE-3 chip includes 900,000 AI optimised compute cores and is composed of a silicon wafer measuring 8.5 by 8.5 inches.
This third-generation chip will be used to power the Cerebras CS-3 AI supercomputer, delivering 125 petaflops of peak AI performance and training AI models up to 24 trillion parameters.
According to Cerebras, these 24 trillion parameter models can be stored in a single logical memory space without partitioning or refactoring, dramatically simplifying training workflow and accelerating developer productivity. It claims that training a one trillion parameter model on the CS-3 is as straightforward as training a one billion parameter model on GPUs.
“When we started on this journey eight years ago, everyone said wafer-scale processors were a pipe dream. We could not be more proud...