Intel Chip Chat

Accelerating AI Inference with Intel® Deep Learning Boost – Intel® Chip Chat episode 632

Informações:

Synopsis

When Intel previewed an array of data-centric innovations in August 2018, one that captured media attention was Intel® Deep Learning Boost, an embedded AI accelerator in the CPU designed to speed deep learning inference workloads. Intel DL Boost will make its initial appearance in the upcoming generation of Intel® Xeon® Scalable processors code-named Cascade Lake. In this Chip Chat podcast, Intel Data-centric Platform Marketing Director Jason Kennedy shares details about the optimization behind some impressive test results. The key to Intel DL Boost – and its performance kick – is augmentation of the existing Intel® Advanced Vector Extensions 512 (Intel® AVX-512) instruction set. This innovation significantly accelerates inference performance for deep learning workloads optimized to use vector neural network instructions (VNNI). Image classification, language translation, object detection, and speech recognition are just a few examples of workloads that can benefit. Early tests have shown image recognition