Artificial intelligence, machine and deep learning are some of the hottest areas in all of high-tech today. We’ve had a few generations of AI over the last 50 years, but in 2010, IBM kicked off the latest cycle with Watson, using brute-force, Big Data techniques to win jeopardy. The University of Toronto in 2012 pioneered Imagenet using deep learning to identify pictures. NVIDIA then began to drive the GPU-accelerated training technology of deep neural nets, and in the course of that, huge service providers opened up and announced initiatives beginning with Microsoft, Google, Apple, Samsung, and then Amazon. Chinese giants Baidu, Alibaba and Tencent are of course, involved. Intel recently held an entire AI day outlining their strategy and announcing their roadmap. Both Qualcomm and Xilinx are active as well.
“Tight call” for Tesla self-driving cars electronics
Given all this activity and Advanced Micro Devices’ stature in GPUs, vital to leading-edge DNN training, everyone was wondering if and when AMD would enter the market. The public got a sneak peek of AMD’s involvement in a conference call with Tesla founder and CEO Elon Musk, where he remarked that AMD was a “tight call” between AMD and NVIDIA for the self-driving electronics. My jaw dropped as I knew while AMD could enter the market if they invested the resources, but didn’t think they would add another thing to their plate. Today, they added DNN accelerators and software to their plate, and it’s all upside for them from a market perspective.
Radeon Instinct accelerator cards
Advanced Micro Devices announced three accelerators, two for inference, and one for training. They are branding these “Radeon Instinct”, representing just what machine learning is supposed to do, which is get machines to have human instincts, the pinnacle of AI.
Here is the Radeon card lineup announced today, which AMD expects to be available in 1H 2017, meaning Q2 2017:
- Instinct MI6: For inference, based on the shipping Polaris architecture, delivering 5.7 TFLOPS (FP16 per AMD), in a full-length board form factor, pulling 150 watts.
- Instinct MI8: For inference, based on the shipping Fiji architecture with HBM, delivering 8.2 TFLOPS (FP16 per AMD), in a short board form factor, pulling less than 175 watts.
- Instinct MI25: For training, based on the upcoming Vega architecture, performance TBD, in a full-length board form-factor, pulling less than 300 watts.
- SuperMicro “SuperServer: 1U rack-mount server, dual-socket Xeon E5 -2600 v4, 3 full-length Radeon Instinct cards
- Inventec K888 G3: 2U rack-mount server, dual socket Xeon E5- 2600 v3, 4 full-length MI25 Radeon Instinct cards, up to 100 TFLOPS (FP16 per AMD)
- “Falconwitch”: disaggregated PCI-E design with support for 16 full-length MI25 cards, up to 400 TFLOPS (FP16 per AMD)
- Inventec Rack: 39U rack-design, 6x Falconwitch switches, 2xToR (top of rack switch), 120 full-length MI25 cards, up to 3,000 TFLOPS (FP16 per AMD)