- ML is still reserch - you shouldn’t aim for 100% succss rate
- But many are doomed to fail:
- Technically infeasible or poorly scoped
- Never make the leap to production
- Unclear success criteria
- Poor team management
- Lifechcle
- Prioritizing Projects
- Archetypes
- Metrics
- 최적화하는 과정에서 뽑아보는 single number
- Baselines
- model이 well performing하고 있는지 확인하기 위해
- How to think about all of the activities in an ML project
- Understand state of the art in your domain
- Understand what’s possible
- Know what to try next
- most promising research areas
- Assessing the feasibility and impact of your projects
A) High-impact ML problems
- Friction in your product
- Complex parts of your pipeline
- Places where cheap prediction is valuable
- What else are people doing?
B) Cost of ML project is driven by data avilability. Also consider accuracy requirements and intrinsic difficulty of the problem
- Where can you take advantage of cheap prediction?
- Where is there friction in your product?
- Where can you automate complicated manual processes?
- What are other people doing?
- Unsupervised learning
- Reinforcement learning
→ Both are showing promise in limited domains where tons of data and compute are available
-
Are you sure you need ML at all?
-
Put in the work up-front to define success criteria with all of the stakeholders
-
Consider the ethics of using ML
-
Do a literature review
-
Try to rapidly build a labeled benchmark dataset
-
Build a “minimal” viable product (e.g., manual rules)
-
Are you sure you need ML at all?
- The main categories of ML projects, and the implications for project management
- How to pick a single number to optimize
A. The real world is messy ; you usually care about lots of metrics
B. However, ML systems work best when optimizing a single number
C. As a result, you need to pick a formula for combining metrics
D. This formula can and will change
- Simple average / weighted average
- Threshold n-1 metrics, evaluate the nth
- More complex / domain-specific formula
- How to know if your model is performing well
A. Baselilnes give you a lower bound on expected model perfrormance
B. The tighter the lower bound, the more useful the baseline(e.g, published results, carefully tuned pipelines, & human baselines are better)
- Andrew Ng’s “Machine Learning Yearning”
- Andrej Karpathy’s “Software 2.0”
- Agrawal’s “The Economics of AI”
- Chip Huyen’s “Introdution to Machine Learning Systems Design”
- Apple’s “Human Interface Guidelines for Machine Learning”
- Google’s “Rules of Machine Learning”