Presented By O'Reilly and Cloudera
Make Data Work
September 25–26, 2017: Training
September 26–28, 2017: Tutorials & Conference
New York, NY

Schedule: Hadoop platform & applications sessions

1:30pm5:00pm Tuesday, September 26, 2017
Location: 1E 12/13 Level: Advanced
Secondary topics:  Architecture
Jonathan Seidman (Cloudera), Gwen Shapira (Confluent), Mark Grover (Lyft)
Average rating: ****.
(4.11, 9 ratings)
Using Customer 360 and the IoT as examples, Jonathan Seidman, Mark Grover, and Gwen Shapira explain how to architect a modern, real-time big data platform leveraging recent advancements in the open source software world, using components like Kafka, Impala, Kudu, Spark Streaming, and Spark SQL with Hadoop to enable new forms of data processing and analytics. Read more.
1:15pm1:55pm Wednesday, September 27, 2017
Location: 1A 23/24 Level: Beginner
Secondary topics:  Platform, Telecom
Travis Bakeman (T-Mobile)
Average rating: **...
(2.00, 1 rating)
Travis Bakeman shares how T-Mobile ported its large-scale network performance management platform, T-PIM, from a legacy database to a big data platform with Impala as the main reporting interface, covering the migration journey, including the challenges the team faced, how the team evaluated new technologies, lessons learned along the way, and the efficiencies gained as a result. Read more.
2:55pm3:35pm Wednesday, September 27, 2017
Location: 1E 09 Level: Beginner
marc carlson (Seattle Children's Research Institute), Sean Taylor (Seattle Children's Research Institute)
Average rating: *****
(5.00, 1 rating)
Marc Carlson and Sean Taylor offer an overview of Project Rainier, which leverages the power of HDFS and the Hadoop and Spark ecosystem to help scientists at Seattle Children’s Research Institute quickly find new patterns and generate predictions that they can test later, accelerating important pediatric research and increasing scientific collaboration by highlighting where it is needed most. Read more.
4:35pm5:15pm Wednesday, September 27, 2017
Location: 1E 09 Level: Intermediate
Secondary topics:  Architecture, Platform, Streaming
Stephen Devine (Big Fish Games), kalah brown (Big Fish Games)
Companies are increasingly interested in processing and analyzing live-streaming data. The Hadoop ecosystem includes platforms and software library frameworks to support this work, but these components require correct architecture, performance tuning, and customization. Stephen Devine and Kalah Brown explain how they used Spark, Flume, and Kafka to build a live-streaming data pipeline. Read more.
5:25pm6:05pm Wednesday, September 27, 2017
Location: 1E 14 Level: Beginner
Sean Kandel (Trifacta), Kaushal Gandhi (Trifacta)
Sean Kandel and Kaushal Gandhi share best practices for building and deploying Hadoop applications to support large-scale data exploration and analysis across an organization. Read more.
1:15pm1:55pm Thursday, September 28, 2017
Location: 1A 23/24 Level: Intermediate
Tony McAllister (Be the Match (National Marrow Donor Program))
The National Marrow Donor Program (Be the Match) recently moved its core transplant matching platform onto Cloudera Hadoop. Tony McAllister explains why the program chose Cloudera Hadoop and shares its big data goals: to increase the number of donors and matches, make the process more efficient, and make transplants more effective. Read more.