Ian Wrigley demonstrates how to leverage the capabilities of Apache Kafka to collect, manage, and process stream data for both big data projects and general-purpose enterprise data integration—no prior knowledge of Kafka required. Ian covers system architecture and use cases and walks you through hands-on exercises where you’ll publish data to, and subscribe to data from, Kafka and investigate Kafka’s Java and REST APIs. Ian also explores other elements of the broader Kafka ecosystem, such as Kafka Connect and Kafka Streams.
This tutorial is ideal for application developers, ETL (extract, transform, load) developers, or data scientists who need to interact with Kafka clusters as a source of, or destination for, stream data.
Ian Wrigley is a Technical Director at StreamSets, the company behind the industry’s first data operations platform. Over his 25-year career, Ian has taught tens of thousands of students subjects ranging from C programming to Hadoop development and administration.
Comments on this page are now closed.
©2016, O'Reilly Media, Inc. • (800) 889-8969 or (707) 827-7019 • Monday-Friday 7:30am-5pm PT • All trademarks and registered trademarks appearing on oreilly.com are the property of their respective owners. • firstname.lastname@example.org
Apache Hadoop, Hadoop, Apache Spark, Spark, and Apache are either registered trademarks or trademarks of the Apache Software Foundation in the United States and/or other countries, and are used with permission. The Apache Software Foundation has no affiliation with and does not endorse, or review the materials provided at this event, which is managed by O'Reilly Media and/or Cloudera.