Data replication Neural Networks Matlab

9 vues (au cours des 30 derniers jours)
Andreas
Andreas le 7 Avr 2016
Commenté : Andreas le 19 Avr 2016
Hello world.! I have recently been studying neural networks, so I may ask something obvious, but I figured out that when I replicate my inputs and outputs and then train the network for pattern recognition,it has far more accuracy than with the original data. I thought of that in order to replicate some of the extreme values I have. Can that make my network overfit? Thank you everyone

Réponse acceptée

Greg Heath
Greg Heath le 10 Avr 2016
Modifié(e) : Greg Heath le 19 Avr 2016
1. I don't understand your question.
2. a. OVERFITTING means there are more unknown weights, Nw, than independent training equations, Ntrneq ( i.e., Nw > Ntrneq).
b. OVERTRAINING an overfit net CAN LEAD to loss of performance on NONTRAINING data.
3. There are several remedies to prevent OVERTRAINING AN OVERFIT NET. So, in general, overfitting need not be disastrous.
4. Methods for preventing loss of generalization via overtraining an overfit net
a. Do not overfit: Nw < Ntrneq. Preferrably,
Ntrneq >> Nw which yields design Stability and
robustness w.r.t. noise and measurement error.
For example:
i. Increase the number of training examples
ii. Reduce the number of hidden nodes
b. Use VALIDATION STOPPING to prevent overtraining
c. Use the BAYESIAN REGULARIZATION
training function TRAINBR with MSEREG
as a default.
d. Replace the default performance function
MSE with the regularized
modification MSEREG
Hope this helps.
Thank you for formally accepting my answer
Greg
  5 commentaires
Greg Heath
Greg Heath le 19 Avr 2016
No.
You add the random noise to the replicated data. Just make sure that the resulting signal to noise ratio is sufficiently high.
Obviously, one good way to approach the problem is as a function of RNG state (which determines the initial random weights) and SNR, given the number of hidden nodes.
However, I have done the reverse, i.e., trained and validated with noisy duplicated data and then used the original data for testing. Again results are presented in terms of the parameter SNR.
Hope this helps.
Greg
Andreas
Andreas le 19 Avr 2016
Dear Dr. Heath,
thank you very much for your time.
Best regards,
Andreas Kampianakis

Connectez-vous pour commenter.

Plus de réponses (0)

Catégories

En savoir plus sur Deep Learning Toolbox dans Help Center et File Exchange

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by