This episode of Techsplainers explores the machine learning pipeline—the systematic process of designing, developing, and deploying machine learning models. We break down the entire workflow into three distinct stages: data processing (covering ingestion, preprocessing, exploration, and feature engineering), model development (including algorithm selection, hyperparameter tuning, training approaches, and performance evaluation), and model deployment (addressing serialization, integration, architecture, monitoring, updates, and compliance). The podcast also emphasizes the critical "Stage 0" of project commencement, where stakeholders define clear objectives, success metrics, and potential obstacles before starting technical work. Throughout the discussion, we highlight how each stage contributes to creating effective, high-performing ML models while examining various training methodologies—from supervised and unsupervised learning to reinforcement and continual learning approaches. Special attention is given to model monitoring and maintenance, acknowledging that deployment is not the end but rather the beginning of a model's productive life cycle.
Find more information at https://www.ibm.com/think/podcasts/techsplainers
Narrated by Ian Smalley