Presented By
O’Reilly + Cloudera
Make Data Work
29 April–2 May 2019
London, UK

Deep learning with TensorFlow and Spark using GPUs and Docker containers

Thomas Phelan (HPE BlueData)
16:3517:15 Thursday, 2 May 2019
Average rating: ***..
(3.29, 7 ratings)

Who is this presentation for?

  • Engineers and data scientists

Level

Beginner

What you'll learn

  • Learn how to spin up and tear down GPU-enabled TensorFlow and Spark
  • Explore quota management of GPU resources for better manageability, isolating GPUs to specific clusters to avoid resource conflict, attaching and detaching GPU resources from clusters, and transient use of GPUs for the duration of the job

Description

Organizations understand the need to keep pace with newer technologies and methodologies when it comes to doing data science with machine learning and deep learning. Data volume and complexity is increasing by the day, so it’s imperative that companies understand their business better and stay on par with or ahead of the competition. Thanks to applications such as Apache Spark, H2O, and TensorFlow, these organizations no longer have to lock in to a specific vendor technology or proprietary solutions. These rich deep learning applications are available in the open source community, with many different algorithms and options for various use cases.

However, one of the biggest challenges is how to get all these open source tools up and running in an easy and consistent manner (keeping in mind that some of them have OS kernel and software components). For example, TensorFlow can leverage NVIDIA GPU resources, but installing TensorFlow for GPU requires users to set up NVIDIA CUDA libraries on the machine and install and configure TensorFlow to make use of the GPU computing ability. The combination of device drivers, libraries, and software versions can be daunting and may be a nonstarter for many users.

Since GPUs are a premium resource, organizations that want to leverage this capability need to bring up clusters with these resources on demand and then relinquish their use after computation is complete. Docker containers can be used to set up this instant cluster provisioning and deprovisioning and can help ensure reproducible builds and easier deployment.

Thomas Phelan demonstrates how to deploy a TensorFlow and Spark with NVIDIA CUDA stack on Docker containers in a multitenant environment. Using GPU-based services with Docker containers does require some careful consideration, so Thomas shares best practices specifically related to the pros and cons of using NVIDIA-Docker versus regular Docker containers, CUDA library usage in Docker containers, Docker run parameters to pass GPU devices to containers, storing results for transient clusters, and integration with Spark.

Photo of Thomas Phelan

Thomas Phelan

HPE BlueData

Thomas Phelan is cofounder and chief architect of BlueData. Previously, a member of the original team at Silicon Graphics that designed and implemented XFS, the first commercially availably 64-bit file system; and an early employee at VMware, a senior staff engineer and a key member of the ESX storage architecture team where he designed and developed the ESX storage I/O load-balancing subsystem and modular pluggable storage architecture as well as led teams working on many key storage initiatives such as the cloud storage gateway and vFlash.