A model you can interpret and understand is one you can more easily improve. It is also one you, regulators, and society can better trust to be safe and nondiscriminatory. An accurate, interpretable model can also offer insights that can be used to change real-world outcomes for the better. There is a central tension, however, between accuracy and interpretability: the most accurate models are necessarily the hardest to understand.
Michael Lee Williams examines the growing business case for interpretability, explores concrete applications including churn, finance, and healthcare, and discusses LIME, an open source, model-agnostic tool that gets around the tension between accuracy and interpretability by allowing you to peer inside black-box models. Michael concludes by sharing a practical prototype that brings these concepts to life: a working web application that uses LIME to explain why customers churn and raises the possibility of intervening to prevent their loss.
Mike Lee Williams is a research engineer at Cloudera Fast Forward Labs, where he builds prototypes that bring the latest ideas in machine learning and AI to life and helps Cloudera’s customers understand how to make use of these new technologies. Mike holds a PhD in astrophysics from Oxford.
For exhibition and sponsorship opportunities, email strataconf@oreilly.com
For information on trade opportunities with O'Reilly conferences, email partners@oreilly.com
View a complete list of Strata Data Conference contacts
©2018, O’Reilly UK Ltd • (800) 889-8969 or (707) 827-7019 • Monday-Friday 7:30am-5pm PT • All trademarks and registered trademarks appearing on oreilly.com are the property of their respective owners. • confreg@oreilly.com