Presented By O’Reilly and Intel Nervana
Put AI to work
September 17-18, 2017: Training
September 18-20, 2017: Tutorials & Conference
San Francisco, CA

Choosing a high-performance computing development direction for original algorithms

Art Popp (ServiceNow)
4:00pm–4:40pm Tuesday, September 19, 2017
Implementing AI
Location: Yosemite BC Level: Intermediate
Secondary topics:  Algorithms, Architectures
Average rating: *....
(1.00, 1 rating)

Prerequisite Knowledge

  • A familiarity with parallel architectures (e.g., Hadoop) and multithreaded programming models (pthreads, OpenMP, etc.)

What you'll learn

  • Explore the tools available for original algorithm development, using both SIMD and SIMT designs

Description

There are many good approaches for designing a new deep learning or AI cluster. If the algorithms you are using are composed entirely of calls to standard AI libraries, then it’s easy to design a few architectures, try them in the cloud, and pick the best design for your needs. However, if there are no supporting libraries, the task of creating two or three prototype architectures can be dauntingly cost and time prohibitive.

Art Popp walks you through a “from scratch" implementation of two algorithms to demonstrate the tools available for original algorithm development, using both SIMD and SIMT designs, the leading hardware architectures of which are Xeon Phi and NVIDIA Cuda. Along the way, Art explores the performance per watt, performance per dollar (initial cost), and performance per dollar (TCO) of each. Each computation camp has its merits. Art’s goal is to give you a peek down each fork in the road and a plan as to how to determine the best direction with the least amount of wasted effort.

Photo of Art Popp

Art Popp

ServiceNow

Art Popp is the senior hardware test engineer at ServiceNow. Previously, Art spent 25 years in the telecommunications industry, the last eight as the principal architect of a large telco carrier’s engineering data warehouse ecosystem, which grew to a mixed environment with 4 PB of IBM TwinFins (formerly Netezzas), 20+ racks of Hadoop gear, and dozens of racks of reporting and presentation systems. (Over time the focus of this environment shifted from reporting to predicting, which is where it got fun.)