Quote:
Original post by kirkd
I don't really understand your question.
What I tried to say is that if you train your NN to 100% accuracy/performance on the training set, the net will likely find random correlations which allow it to get to that level. Then when you use the net in a new situation outside those used for training, it may not perform well. I would bet that it would perform quite poorly.
The method I described is referred to as cross-validation. Take you dataset of 50 patterns, as you called them. Take out 10 at random and use the remaining 40 to train your net. The training protocol is trying to maximize performance on thost 40. If you watch the performance on teh 10 that were left out at random during the training, what you'll typically see is that the net starts off terrible (as expected), gradually it improves in performance, and then it starts to get worst. All the while it is gettting better and better on the 40 training patterns. This is where overtraining starts to set in. The net has captured most of the detail of the problem and is now starting to find random correlations in the training data. These leads to a loss in performance on the 10 left out patterns.
What typically is done is that the process is repeated a number of times in order to make sure that each training pattern was left out at some point, and also to ensure that the random selection of 10 to leave out was not biased in some way. By repeating the process and then averaging the performance results, you get a better estimate of how well the net will perform in a new situation.
I hope that helps.
-Kirk
Basically I'm trying to understand what you called cross-validation. I understand that I should train a pattern of 40 inputs, but I don't understand the use of the 10 inputs patterns.
Sorry guy, you're doing your best to explain me, but when I don't understand something, I cannot stay without asking...
thanks again for your explaning