Ray Lucchesi of RayOnStorage Blog comments:
At Google IO conference this week, they revealed (see Google supercharges machine learning tasks …) that they had been designing and operating their own processor chips in order to optimize machine learning. They called the new chip, a Tensor Processing Unit (TPU). According to Google, the TPU provides an order of magnitude more power efficient machine learning over what’s achievable via off the shelf GPU/CPUs. TensorFlow is Google’s open sourced machine learning software.
When it comes to machine learning, hardware is still a necessity to do some of the really complicated things fast.
Read more at: TPU and hardware vs. software innovation (round 3)
- Rogue Device Detection Thanks To PathSolutions - May 27, 2020
- Validating Identity with Identiq - May 21, 2020
- Pensando Places Programmability First - May 18, 2020
- Stopping Stoplight Risk Analysis with Brinqa - May 14, 2020
- Leaving Legacy Behind to Build Better Networks with DriveNets - May 13, 2020
- Tomversations: Episode 2 – Wi-Fi 6 and 6E - May 11, 2020
- Assured IoT Reporting with Jitsuin Archivst - May 7, 2020
- Ensuring Code Quality with Arista - May 7, 2020
- Monitoring Application Performance from the Inside Out with Solarwinds AppOptics - May 6, 2020
- Feature-Based Licensing for Infrastructure is a Good Thing - May 5, 2020