When Intel previewed an array of data-centric innovations in August 2018, one that captured media attention was Intel\xae Deep Learning Boost, an embedded AI accelerator in the CPU designed to speed deep learning inference workloads.\n\nIntel DL Boost will make its initial appearance in the upcoming generation of Intel\xae Xeon\xae Scalable processors code-named Cascade Lake. In this Chip Chat podcast, Intel Data-centric Platform Marketing Director Jason Kennedy shares details about the optimization behind some impressive test results.\n\nThe key to Intel DL Boost \u2013 and its performance kick \u2013 is augmentation of the existing Intel\xae Advanced Vector Extensions 512 (Intel\xae AVX-512) instruction set. This innovation significantly accelerates inference performance for deep learning workloads optimized to use vector neural network instructions (VNNI). Image classification, language translation, object detection, and speech recognition are just a few examples of workloads that can benefit.\n\nEarly tests have shown image recognition 11 times faster using a similar configuration than with current-generation Intel Xeon Scalable processors when launched in July 2017. Current projections estimate 17 times faster inference throughput benefit with Intel\xae Optimized Caffe ResNet-50 and Intel Deep Learning Boost that can be achieved with a new class of advance performance CPUs debuting in the upcoming generation.\n\nFor more information about AI activities across Intel visit ai.intel.com.\n\nSoftware and workloads used in performance tests may have been optimized for performance only on Intel microprocessors. Performance tests, such as SYSmark and MobileMark, are measured using specific computer systems, components, software, operations and functions. Any change to any of those factors may cause the results to vary. You should consult other information and performance tests to assist you in fully evaluating your contemplated purchases, including the performance of that product when combined with other products. For more information go to www.intel.com/benchmarks. \n\nPerformance results are based on testing or projections as of 6/2017 to 11/7/2018 and may not reflect all publicly available security updates. See configuration disclosures in https://intel.ly/2Mw7KKE and https://intel.ly/2VUvY2I for details. No product can be absolutely secure. \n\n[1] Future Intel Xeon Scalable processors codenamed Cascade Lake provide up to 11x inference performance with Intel\xae Optimized Caffe, ResNet-50, and Intel\xae Deep Learning Boost (VNNI) in comparison to current generation Intel Xeon Scalable Platinum 8180 Processor with FP32 instructions (at launch, measured July, 2017), for details see https://intel.ly/2Mw7KKE, slide 33.\n\n[2] Future Intel Xeon Scalable processors codenamed Cascade Lake advanced performance provide up to 17x images per second (deep learning inference) in comparison to current generation Intel Xeon Scalable Platinum 8180 Processor (at launch, measured July, 2017), for details see https://intel.ly/2VUvY2I.