Does it make sense to validate a NN after training?
Let A, B be training sets of n-vectors. Vectors in A are examples, or vectors with expected output 1. Vectors in B are counterexamples, with expected output 0. After training, a perceptron NN
P:R^n \rightarrow \{1,0\}
has outputs 1 on A, and 0 on B. So, it is 100\% accurate on the training data.
Recall that training is performed, in most cases, taking into account only the numerical values of the data vectors. Usually the training algorithms have no place for the 'meaning', 'interpretation', or 'empirical content', of the data.
Next, in order to 'validate' P, two sets of test data are given, A', B' and A'', B''.
Evaluating P on A', B' turns out to be 0\% accurate, that is, P(A')=0 and P(B')=1. The NN completely fails the first validating data set.
Finally, evaluating P on A'', B'' is 100\% accurate. Validation is completely successful in this case.
Many intermediate percentages of accuracy occur in practice.
All taken into account, the question arises: Was P well trained?
The criteria the neuroscientist used to choose both the training and the test vector sets ---in principle a subjective judgment--- must be taken into account. But the literature on NNs seldom carries an explicit discussion of these criteria.
Another way to put it: How to validate validation? This can be referred to as the problem of metavalidation.
References and opinions on this topic are welcomed.