Lecture 8 - Lifecycle of A Data Science Project - Part 2
Lecture 8 - Lifecycle of A Data Science Project - Part 2
[email protected]
CRISP-DM: with a use case
3
What is CRISP-DM
Cross-industry standard process for data
mining - CRISP-DM
Source
4
What is CRISP-DM
Cross-industry standard process for data
mining - CRISP-DM
Maintenace and
An open standard developed in 1996 by leading monitoring
companies in data analysis
Deployment
• Pre-deployment
• Testing online
• Monitoring and logging
Maintenace and
• Active feedback monitoring
• On-call responsibilities
• Set aside enough time for this phase
8
Deployment
• Three Vs of MLOps: Velocity, Validation and Version
Maintenace and
monitoring
9
• Deployable code
• Robust
• Standardized code quality
• Hard to make unintended changes
• Infrastructure constraints
• Easy to maintain
• Well tested
• Easy to maintain
• Production environment
10
Production readiness
12
Breck, Eric, et al. "The ML test score: A rubric for ML production readiness and technical debt
reduction." 2017 IEEE International Conference on Big Data (Big Data). IEEE, 2017.
13
Breck, Eric, et al. "The ML test score: A rubric for ML production readiness and technical debt
reduction." 2017 IEEE International Conference on Big Data (Big Data). IEEE, 2017.
14
Breck, Eric, et al. "The ML test score: A rubric for ML production readiness and technical debt
reduction." 2017 IEEE International Conference on Big Data (Big Data). IEEE, 2017.
15
Breck, Eric, et al. "The ML test score: A rubric for ML production readiness and technical debt
reduction." 2017 IEEE International Conference on Big Data (Big Data). IEEE, 2017.
16
Breck, Eric, et al. "The ML test score: A rubric for ML production readiness and technical debt
reduction." 2017 IEEE International Conference on Big Data (Big Data). IEEE, 2017.
17
Infrastructure tests
Breck, Eric, et al. "The ML test score: A rubric for ML production readiness and technical debt
reduction." 2017 IEEE International Conference on Big Data (Big Data). IEEE, 2017.
18
Maintenace and
monitoring
19
Monitoring
• By definition, your system is making predictions on previously unseen data
• Crucial to know that the system continues to work correctly over time
• Using Dashboards displaying relevant graphs and statistics
• Monitoring the system, pipelines and input data
• Alerting the team when metrics deviate significantly for expectations
20
Monitoring
Source
21
Monitoring
22
Monitoring tests
Breck, Eric, et al. "The ML test score: A rubric for ML production readiness and technical debt
reduction." 2017 IEEE International Conference on Big Data (Big Data). IEEE, 2017.
24
Alert fatigue
• A surplus of false-positive alerts led to fatigue and silencing of alerts, which could miss
actual performance drops.
“Recently we've noticed that some of these alerts have been rather noisy and not
necessarily reflective of events that we care about triaging and fixing. So we've recently
taken a close look at those alerts and are trying to figure out, how can we more precisely
specify that query such that it's only highlighting the problematic events?”
“You typically ignore most alerts...I guess on record I'd say 90% of them aren't immediate.
You just have to acknowledge them internally, like just be aware that there is something
happening.”
Maintenance
Source
26
Maintenance
Source
27
•https://www.kaggle.com/trnderenergikraft/grid-loss-time-series-dataset
29
• Market changes
• Technology updates
• Hardware/Software updates
Simplicity
Simplicity is an advantage but sadly, complexity sells better (Source)
• Rule #4: Keep the first model simple and get the infrastructure right.
• Rule #5: Test the infrastructure independently from the machine learning.
• Rule #26: Look for patterns in the measured errors, and create new features.
Source
35
“One thing that I've noticed is, especially when you have as many resources as
large companies do, that there's a compulsive need to leverage all the
resources that you have. And just, you know, get all the experiments out there.
Come up with a bunch of ideas; run a bunch of stuff. I actually think that's bad.
You can be overly concerned with keeping your GPUs warm, so much so that
you don't actually think deeply about what the highest value experiment is. I
think you can end up saving a lot more time—and obviously GPU cycles, but
mostly end-to-end completion time—if you spend more efforts choosing the
right experiment to run instead of spreading yourself thin. All these different
experiments have their own frontier to explore, and all these frontiers have
different options. I basically will only do the most important thing from each
project's frontier at a given time, and I found that the net throughput for
myself has been much higher.”
Important to know!
• Communication is the key (Stakeholders, Management, domain experts, end users, data scientists/engineers).
• Model performance depends less on the model than the data we feed the model.
• It is important to not dive right in and think about the problem and get feedback from experts.
• Best predicting model might not be the best value creating model.
• Important to test the systems for: software, data science pipeline and value creation.
• An imperfect deployed system is more valuable than a perfect undeployed system (80-20 rule).
40
Resources
41
Important Deadlines
When you will need to deliver or complete a task
Lecture Plan
Unpacking the course syllabus
1 23/8 Lecture 1: Introduction [Nisha Dalal] 8 11/10 Lecture 6: Data Visualization & Storytelling
[Manos Papagiannidis]
5 20/9 No lecture
8/11 Lecture 10: Experiences from Industry
12
[Thomas Thorensen]
6 27/9 Lecture 5: Lifecycle of a Data Science project I
[Nisha Dalal]
Nisha Dalal
Questions & Discussion [email protected]