Logo Questions Linux Laravel Mysql Ubuntu Git Menu
 

The correctness of neural networks

I have asked other AI folk this question, but I haven't really been given an answer that satisfied me.

For anyone else that has programmed an artificial neural network before, how do you test for its correctness?

I guess, another way to put it is, how does one debug the code behind a neural network?

like image 367
supercheetah Avatar asked May 07 '09 16:05

supercheetah


4 Answers

With neural networks, generally what is happening is you are taking an untrained neural network, and you are training it up using a given set of data, so that it responds in the way you expect. Here's the deal; usually, you're training it up to a certain confidence level for your inputs. Generally (and again, this is just generally; your mileage may vary), you cannot get neural networks to always provide the right answer; rather, you are getting the estimation of the right answer, to within a confidence range. You know that confidence range by how you have trained the network.

The question arises as to why you would want to use neural networks if you cannot be certain that the conclusion they come to is verifiably correct; the answer is that neural networks can arrive at high-confidence answers for certain classes of problems (specifically, NP-Complete problems) in linear time, whereas verifiably correct solutions of NP-Complete problems can only be arrived at in polynomial time. In layman's terms, neural networks can "solve" problems that normal computation can't; but you can only be a certain percentage confident that you have the right answer. You can determine that confidence by the training regimen, and can usually make sure that you will have at least 99.9% confidence.

like image 124
Paul Sonier Avatar answered Oct 22 '22 08:10

Paul Sonier


Correctness is a funny concept in most of "soft computing." The best I can tell you is: "a neural network is correct when it consistently satisfies the parameters of it's design." You do this by training it with data, and then verifying with other data, and having a feedback loop in the middle which lets you know if the neural network is functioning appropriately.

This is of-course the case only for neural networks that are large enough where a direct proof of correctness is not possible. It is possible to prove that a neural network is correct through analysis if you are attempting to build a neural network that learns XOR or something similar, but for that class of problem an aNN is seldom necessary.

like image 21
earino Avatar answered Oct 22 '22 09:10

earino


You're opening up a bigger can of worms here than you might expect.

NN's are perhaps best thought of as universal function approximators, by the way, which may help you in thinking about this stuff.

Anyway, there is nothing special about NN's in terms of your question, the problem applies to any sort of learning algorithm.

The confidence you have in the results it is giving is going to rely on both the quantity and the quality (often harder to determine) of the training data that you have.

If you're really interested in this stuff, you may want to read up a bit on the problems of overtraining, and ensemble methods (bagging, boosting, etc.).

The real problem is that you usually aren't actually interested in the "correctness" (cf quality) of an answer on a given input that you've already seen, rather you care about predicting the quality of answer on an input you haven't seen yet. This is a much more difficult problem. Typical approaches then, involve "holding back" some of your training data (i.e. the stuff you know the "correct" answer for) and testing your trained system against that. It gets subtle though, when you start considering that you may not have enough data, or it may be biased, etc. So there are many researchers who basically spend all of their time thinking about these sort of issues!

like image 3
simon Avatar answered Oct 22 '22 10:10

simon


I've worked on projects where there is test data as well as training data, so you know the expected outputs for a set of inputs the NN hasn't seen.

One common way of analysing the result of any classifier is use of an ROC curve; an introduction to the statistics of classifiers and ROC curves can be found at Interpreting Diagnostic Tests

like image 2
Pete Kirkham Avatar answered Oct 22 '22 10:10

Pete Kirkham