They are doing it with custom silicon with several times more area than 8x H100s. Iām sure they are doing some sort of optimization at execution/runtime, but the primary difference is the sheer transistor count.
It is worth splitting out the stacked memory silicon layers on both too (if Cerebras is set up with external DRAM memory). HBM is over 10 layers now so the die area is a good bit more than the chip area, but different process nodes are involved.
https://cerebras.ai/product-chip/