Intel held its inaugural Artificial Intelligence (AI) Developers Conference in San Francisco on May 23-24th, presenting its leadership, technologies, and customers to a capacity audience of some 800 AI geeks and media. The company now has a rich portfolio of AI technologies, after acquiring Movidius and MobileEye for real-time processing, Altera for reprogrammable FPGA acceleration hardware, and Nervana for the training workloads currently served by NVIDIA GPUs. Intel’s primary focus on inference processing for the production use of trained neural networks is a sound strategy, as inference is likely to become a much larger market that the training segment over the next few years. While Intel does not yet have a brawny ASIC in its portfolio to build AI networks, it can create a sizable position in inference, alongside companies such as Apple , Qualcomm , Xilinx, and NVIDIA.
That being said, Intel has not abandoned the AI training market, where NVIDIA is enjoying tremendous success with a $3B run rate. Intel stressed Xeon’s strengths in training at the event, while pointing to the future where it hopes to leverage Nervana to compete more directly with NVIDIA’s big silicon. Unfortunately for Intel, Nervana now appears to be at least 18 months away—a result of an even larger redesign, which I forecasted here. Here's what I learnedAt the event keynote, Naveen Rao, Intel’s SVP for AI, articulated the company’s strategy for AI: essentially, to provide the full range of general purpose and specialized devices for AI supported by a unified suite of optimizing development software. As Mr. Rao pointed out, running AI apps is not a market where one size fits all, and Intel has products offering a broad range of performance, latency, and power envelopes for inference processing

The event also gave the company the opportunity to share some customer progress with Intel hardware for AI, including Google, Amazon, Microsoft , Novartis , and Facebook. Novartis is a good example where the GPU memory may be inadequate for processing large datasets without thrashing between HBM and the CPU-controlled DDR4 memory. This also points to the benchmarking issues that Intel is complaining about. Specifically, the performance of a chip in training the ImageNet database, where images are only 224x224x3 bits, is irrelevant if you, like Novartis, are processing 1024x1280x3 bit images of molecules. Novartis also recently scaled its AI training to eight nodes using OmniPath, reducing its training time from 11 hours to just 31 minutes. Intel noted that the Skylake generation Xeons have specific instructions such as reduced precision math operations that help Xeon performance for AI.
Facebook shared some interesting data showing that it uses (Xeon) CPUs for all inference work and select training jobs, and GPUs for training Convolutional Neural Networks (CNNs) for image processing and Recurrent Neural Networks (RNN) for speech and language translation. It is workloads such as these for which Intel needs Nervana, but more on that in a minute. Per the subtitle in Figure 2, Facebook makes use of CPUs in part because it already has lots (millions) of them.

Intel quite rightly pointed out that most enterprises have surplus CPU capacity, especially at night. Intel has eased the software burden to use those existing resources for many machine learning workloads. As Mr. Rao said, enterprises can run AI on the chips they already have.
In my mind, the biggest news of the event was an update on the highly anticipated Nervana Neural Network Processor (NNP) roadmap. Prior to its acquisition by Intel, Nervana was expected to deliver a fabric-enabled NNP accelerator that might beat an unspecified GPU by a purported 10X. Intel has been sampling that first-generation chip to major AI customers and plans to incorporate their input and enhancement requests into the first production NNP part, due in late 2019. I expected Nervana to ship its first part to generate revenue last year, but now it feels it can afford the time to get the first commercial product right. Intel gave us the tidbits we can use to project where the company might land with the NNP in figure 3 and 4.
