All Utilizing AI

Running AI Everywhere and In Everything with Intel | Utilizing AI 2×19

AI processing is appearing everywhere, running on just about any kind of infrastructure, from the cloud to the edge to end-user devices. Although we might think AI processing requires massive centralized resources, this is not necessarily the case. Deep learning training might need centralized resources, but the topic goes way beyond this, and it is likely that most production applications will use CPUs to process data in-place. Simpler machine learning applications don’t need specialized accelerators and Intel has been building specialized hardware support into their processors for a decade. DL Boost on Xeon is competitive with discrete GPUs thanks to specialized instructions and optimized software libraries.

Three Questions

  1. How long will it take for a conversational AI to pass the Turing test and fool an average person?
  2. Is it possible to create a truly unbiased AI?
  3. How small can ML get? Will we have ML-powered household appliances? Toys? Disposable devices?

Guests and Hosts

For your weekly dose of Utilizing AI, subscribe to our podcast on your favorite podcast app through Anchor FM and watch more Utilizing AI podcast videos on the dedicated website

About the author

Stephen Foskett

Stephen Foskett is an active participant in the world of enterprise information technology, currently focusing on enterprise storage, server virtualization, networking, and cloud computing. He organizes the popular Tech Field Day event series for Gestalt IT and runs Foskett Services. A long-time voice in the storage industry, Stephen has authored numerous articles for industry publications, and is a popular presenter at industry events. He can be found online at,, and on Twitter at @SFoskett.

Leave a Comment