Estimating the bias and variance of your learning algorithm really helps you prioritize what to work on next. But the way you analyze bias and variance changes when your training set comes from a different distribution than your dev and test sets.
– | error |
---|---|
Human | 0% |
training | 1% |
dev | 10% |
To examine, make a sub set of data.
Training-dev set: same distribution as trainng set, but not used for training.
We started with following setup where the dev and test sets had the same distribution
Randomly shuffle training set
and create a train-dev set
So training and train-dev set have the same distribution, and dev and test have the same distribution.
Now, new training set can be used for train NN, and train-dev is used for error analyis as shown below.
Now we have a variance problem
– | error |
---|---|
Human | 4% |
training | 7% |
train-dev | 10% |
Dev | 6% |
Test | 6% |
– | general speech recognition | rearview mirror speech data |
---|---|---|
Human | “Human Level” 4% | 6% |
error on examples trained on | “Training error” 7% | 6% |
errors on examples not trained on | “Traing - dev error” 10% | “Dev/Test Error” 6% |
How do you address data mismatch?