What's the trade-off between bias and variance?
Bias is error due to erroneous or overly simplistic assumptions in the learning algorithm you're using. This can lead to the model underfitting your data, making it hard for it to have high predictive accuracy and for you to generalize your knowledge from the training set to the test set.
Variance is error due to too much complexity in the learning algorithm you're using. This leads to the algorithm being highly sensitive to high degrees of variation in your training data, which can lead your model to overfit the data. You'll be carrying too much noise from your training data for your model to be very useful for your test data.
The bias-variance decomposition essentially decomposes the learning error from any algorithm by adding the bias, the variance and a bit of irreducible error due to noise in the underlying dataset. Essentially, if you make the model more complex and add more variables, you'll lose bias but gain some variance — in order to get the optimally reduced amount of error, you'll have to tradeoff bias and variance. You don't want either high bias or high variance in your model.
Learn More :
Data Science
- What features would you use to predict the Uber ETA for ride requests?
- How would you evaluate the predictions of an Uber ETA model?
- Describe how you would build a model to predict Uber ETAs after a rider requests a ride.
- Suppose you're working as a data scientist at Facebook. How would you measure the success of private stories on Instagram, where only certain chosen friends can see the story?
- Precision vs Accuracy Vs Recall?
- Error vs variance vs bias?
- False negatives vs false positives? When is either one worse than the other?
- Describe your data science process start to finish?
- Data science vs machine learning vs AI?
- How would you find correlation between a categorical variable and a continuous variable?
- How do you treat null/missing values? Name 3 methodologies.
- How can outlier values be treated?
- What is data normalization? Name 2 normalization methodologies.
- What is the role/importance of data cleaning?
- What are success metrics vs tracking metrics?
- What kind of metric would you make to measure success of a program (marketing) and how do you define them?
- Let's say an app was getting a redesign. How do you know if the redesign was successful?
- We noticed a steep decline in users in a certain area of the world, how would you address/asses?
- What are the two methods used for the calibration in Supervised Learning?
- Which method is frequently used to prevent overfitting?
- What is the difference between heuristic for rule learning and heuristics for decision trees?
- What is Perceptron in Machine Learning?
- Explain the two components of Bayesian logic program?
- What are Bayesian Networks (BN) ?
- Why instance based learning algorithm sometimes referred as Lazy learning algorithm?