Hi all,
I create .mat files out of my 3D dicom images. This is one step of my pipeline to work with a convolutional NN.
In the begining, I noticed that creating the .mat files once saves me time to train the NN with different kernel sizes.
Ex: for a 3x3 kernel size, I first create the imds with brand new .mat files.
Then, I clear all the variables and make the correspondent changes in my network to approach the imds with 5x5 kernel.
I only have to indicate the path to the imds (with the already created .mat files). Here, I noticed that the training step, even by clear all variables, it is faster than for 3x3.
Would it be the case MATLAB 'learned' the previous created .mat files, leading to an overfitting issue?
Thanks in advance.