Transparency, auditability, and stability of predictive models and results are typically key differentiators in effective machine learning applications. Patrick Hall, Avni Wadhwa, and Mark Chan share tips and techniques learned through implementing interpretable machine learning solutions in industries like financial services, telecom, and health insurance. Using a set of publicly available and highly annotated examples, Patrick, Avni, and Mark teach several holistic approaches to interpretable machine learning. The examples use the well-known University of California Irvine (UCI) credit card dataset and popular open source packages to train constrained, interpretable machine learning models and visualize, explain, and test more complex machine learning models in the context of an example credit-risk application. Along the way, Patrick, Avni, and Mark draw on their applied experience to highlight crucial success factors and common pitfalls not typically discussed in blog posts and open source software documentation, such as the importance of both local and global interpretability and the approximate nature of nearly all machine learning interpretability techniques.
Outline:
Enhancing transparency in machine learning models with Python and XGBoost (example Jupyter notebook)
Increasing transparency and accountability in your machine learning project with Python (example Jupyter notebook)
Explaining your predictive models to business stakeholders with local interpretable model-agnostic explanations (LIME) using Python and H2O (example Jupyter Notebook)
Testing machine learning models for accuracy, trustworthiness, and stability with Python and H2O (example Jupyter notebook)
Patrick Hall is principle scientist at bnh.ai, a boutique law firm focused on AI and analytics; a senior director of product at H2O.ai, a leading Silicon Valley machine learning software company; and a lecturer in the Department of Decision Sciences at George Washington University, where he teaches graduate classes in data mining and machine learning.
At both bnh.ai and H2O.ai, he works to mitigate AI risks and advance the responsible practice of machine learning. Previously, Patrick held global customer-facing and R&D research roles at SAS. He holds multiple patents in automated market segmentation using clustering and deep neural networks. Patrick is the 11th person worldwide to become a Cloudera Certified Data Scientist. He studied computational chemistry at the University of Illinois before graduating from the Institute for Advanced Analytics at North Carolina State University.
Avni Wadhwa is an analytics and marketing hacker at H2O.ai, where she does a mix of marketing and sales engineering. She holds a BS in management science from the University of California, San Diego.
Mark Chan is a hacker and data scientist at H2O.ai. Previously, he was a quantitative research developer at Thomson Reuters and Nipun Capital and a data scientist at an IoT startup, where he built a web-based machine learning platform and developed predictive models. Mark holds an MS in financial engineering from UCLA and a BS in computer engineering from the University of Illinois Urbana-Champaign. In his spare time, Mark likes competing on Kaggle and cycling.
Comments on this page are now closed.
For exhibition and sponsorship opportunities, email strataconf@oreilly.com
For information on trade opportunities with O'Reilly conferences, email partners@oreilly.com
View a complete list of Strata Data Conference contacts
©2018, O'Reilly Media, Inc. • (800) 889-8969 or (707) 827-7019 • Monday-Friday 7:30am-5pm PT • All trademarks and registered trademarks appearing on oreilly.com are the property of their respective owners. • confreg@oreilly.com
Comments
Hi Juan – Either a Gold or Silver pass would be sufficient to attend this talk.
Which pass would I need to attend this talk?