Episode Thumbnail
Episode 74  |  28:14 min

Episode 74: How to Avoid Bias in Your Machine Learning Models with Clare Corthell

Episode 74  |  28:14 min  |  11.25.2019

Episode 74: How to Avoid Bias in Your Machine Learning Models with Clare Corthell

00:00
00:00
This is a podcast episode titled, Episode 74: How to Avoid Bias in Your Machine Learning Models with Clare Corthell. The summary for this episode is: Bias exists everywhere. It factors into everything that we do and into virtually every decision that we make. An interesting, but problematic side effect of this is that bias can also easily slip into our machine models. In this episode, Jon Prial talks with Clare Corthell, a well-known and respected data scientist and engineer, and the founder of Luminant Data, about the issues that can arise when bias enters your models and how to avoid it in the first place. Plus, check out our show notes to find out how to access our first episode of Extra Impact, where Jon and Clare go deeper into this topic, talking about bias in AI-powered services like Airbnb and the controversial policing stop-and-frisk program. You’ll hear about: -- The nature of bias in machine learning models and what causes it -- Cynthia Dwork’s work on transparency -- How companies should be thinking about data -- Implementing fairness into AI and machine learning models -- Developing an AI code of ethics Access the show notes here: http://bit.ly/2GLyjsU
Bias exists everywhere. It factors into everything that we do and into virtually every decision that we make. An interesting, but problematic side effect of this is that bias can also easily slip into our machine models. In this episode, Jon Prial talks with Clare Corthell, a well-known and respected data scientist and engineer, and the founder of Luminant Data, about the issues that can arise when bias enters your models and how to avoid it in the first place. Plus, check out our show notes to find out how to access our first episode of Extra Impact, where Jon and Clare go deeper into this topic, talking about bias in AI-powered services like Airbnb and the controversial policing stop-and-frisk program. You’ll hear about: -- The nature of bias in machine learning models and what causes it -- Cynthia Dwork’s work on transparency -- How companies should be thinking about data -- Implementing fairness into AI and machine learning models -- Developing an AI code of ethics Access the show notes here: http://bit.ly/2GLyjsU

More Episodes

An Introduction to Self-Sovereign Identity with Northern Block CEO Mathieu Glaude

An Introduction to Self-Sovereign Identity with Northern Block CEO Mathieu Glaude

Building  Effective Cybersecurity Defenses with eSentire's CTO Dustin Hillard

Building Effective Cybersecurity Defenses with eSentire's CTO Dustin Hillard

Detecting Cybersecurity Threats with Devo's CEO Marc van Zadelhoff

Detecting Cybersecurity Threats with Devo's CEO Marc van Zadelhoff

What's Identity Authentication and How Does It Work? ft. Trusona's CTO John Summers

What's Identity Authentication and How Does It Work? ft. Trusona's CTO John Summers

Navigating the Cybersecurity Landscape with CISO Alex Manea

Navigating the Cybersecurity Landscape with CISO Alex Manea

Rewind CEO Mike Potter On Backing up the Cloud, the Risk 3rd Party Applications Pose & Building a Time Machine for SaaS Apps

Rewind CEO Mike Potter On Backing up the Cloud, the Risk 3rd Party Applications Pose & Building a Time Machine for SaaS Apps