Demand for AI compute is growing faster than conventional systems architecture can match, so companies like Cerebras Systems are building massive special-purpose processing units. In this episode, Andy Hock, VP of Product for Cerebras Systems, joins Frederic Van Haren and Stephen Foskett to discuss this new class of hardware. The Cerebras Wafer-Scale Engine (WSE-2) has 850,000 processors on a single chip the size of a dinner plate, along with 40 GB of SRAM and supporting interconnects. But Cerebras also has a software stack that integrates with standard ML frameworks like PyTorch and TensorFlow. Although the trillion-parameter model is a real need for certain applications, platforms need to be flexible to support both massive-scale and more mainstream workloads, and this is a focus for Cerebras as well.
- Frederic’s Question: How small can ML get? Will we have ML-powered household appliances? Toys? Disposable devices?
- Stephen’s Question: Will we ever see a Hollywood-style “artificial mind” like Mr. Data or other characters?
- Leon Adato, host of the Technically Religious Podcast: I’m curious, what responsibility do you think IT folks have to insure the things that we build are ethical?
Andy Hock, VP of Product at Cerebras Systems. Connect with Andy on LinkedIn. Follow Cerebras Systems on Twitter at @CerebrasSystems.
- Frederic Van Haren, Founder at HighFens Inc., Consultancy & Services. Connect with Frederic on Highfens.com or on Twitter at @FredericVHaren
- Stephen Foskett, Publisher of Gestalt IT and Organizer of Tech Field Day. Find Stephen’s writing at GestaltIT.com and on Twitter at @SFoskett
For your weekly dose of Utilizing AI, subscribe to our podcast on your favorite podcast app through Anchor FM and check out more Utilizing AI podcast episodes on the dedicated website https://utilizing-ai.com/