What is bad data?

Media Thumbnail
00:00
00:00
1x
  • 0.5
  • 1
  • 1.25
  • 1.5
  • 1.75
  • 2
This is a podcast episode titled, What is bad data?. The summary for this episode is: <p>This episode of Techsplainers explores the concept of ""bad data"" - information that compromises decision-making because it's inaccurate, incomplete, inconsistent, outdated, duplicate, invalid, or biased. We examine why bad data is particularly dangerous due to its stealthy nature, often going undetected until significant damage occurs. Through real-world examples like Unity Technologies' $110 million loss from bad data in their AI algorithms, we illustrate the severe consequences across industries from healthcare to finance. The discussion covers the diverse causes of data quality problems - from system failures and data decay to human error and integration challenges - and provides a comprehensive approach to prevention through governance, monitoring, cleansing, and data literacy. As organizations increasingly rely on AI systems, understanding that ""garbage in, garbage out"" applies more than ever becomes crucial for success in data-driven initiatives.&nbsp;</p><p><br></p><p><strong>Find more information at</strong> https://www.ibm.com/think/topics/bad-data&nbsp;&nbsp;</p><p><br></p><p><strong>Find more episodes at </strong>https://www.ibm.biz/techsplainers-podcast&nbsp;</p><p><br></p><p><strong>Narrated by Amanda Downie&nbsp;</strong></p>

DESCRIPTION

This episode of Techsplainers explores the concept of ""bad data"" - information that compromises decision-making because it's inaccurate, incomplete, inconsistent, outdated, duplicate, invalid, or biased. We examine why bad data is particularly dangerous due to its stealthy nature, often going undetected until significant damage occurs. Through real-world examples like Unity Technologies' $110 million loss from bad data in their AI algorithms, we illustrate the severe consequences across industries from healthcare to finance. The discussion covers the diverse causes of data quality problems - from system failures and data decay to human error and integration challenges - and provides a comprehensive approach to prevention through governance, monitoring, cleansing, and data literacy. As organizations increasingly rely on AI systems, understanding that ""garbage in, garbage out"" applies more than ever becomes crucial for success in data-driven initiatives. 


Find more information at https://www.ibm.com/think/topics/bad-data  


Find more episodes at https://www.ibm.biz/techsplainers-podcast 


Narrated by Amanda Downie