-
Notifications
You must be signed in to change notification settings - Fork 1.4k
Accuracy show the wrong result on graph #625
Comments
Hi @ssilphf, I can't reproduce your problem yet. I just merged #608, which should give you some more information when you classify a list of images. Try it out and see if you can get any more information. Also, if you can give me some standard information about your machine I may be able to spot something fishy:
|
Hi @ssilphf can you be more specific on the versions you are using:
|
I'm sorry, but maybe because I didn't use the comment
CuCNN: the file name is I find that the problem maybe not from DIGITS, and maybe from caffe or torch or itorch. |
I had reinstall my all system and my new version is: DIGITS: but the problem still here. And I use DIGITS2.3dev(other all is same as DIGITS3) on the same dataset, |
Hello, can you try to upgrade to more recent software? There is a bug in On Tue, May 3, 2016 at 1:11 PM, rajulanand notifications@github.com wrote:
|
Can you please suggest me how to upgrade cudnn to 4.0.7? I broke my existing digits, nv-caffe installation while trying to do that? I am on 15.04 so I have to build everything from source (I can't get to install 14.04 on my machine for reasons unknown to me). The cuda-7.5 installation contains default 4.0.4 and even after install 4.0.7 from deb file, the lib64 only contains files from 4.0.4. cmake of nv-caffe installations doesn't seem to find cudnn and variables CUDA_cublas_LIBRARY, CUDA_curand_LIBRARY are also set to not found. Any suggestions? Update: I was able to resolve cudnn to latest version, however now I can't compile nv-caffe despite satisfying all the dependencies mentioned on github for compling nv-caffe. Update II: I was able to compile nv-caffe but now digits fail with Intel MKL FATAL ERROR: Cannot load libmkl_avx2.so or libmkl_def.so Update III: I was able to resolve all the issues and having working version of digits, nv-caffe with latest cudnn. |
@gheinrich After upgrading and everything, I am still getting the wrong prediction result on that particular model. |
Can you tell us what the validation accuracy was for this model and what the confusion matrix looks like when you are doing "classify many"? To get a confusion matrix you may use the |
Closing due to inactivity. |
Dear Greg, I am getting a significant difference between the accuracy shown on the plot / in the final log file and the one from Classify Many module. Specially, this discrepancy happens in the GoogleNet trained models. I see this issue in the LeNet models as well but can be tolerated. If you have found the solution , please advise as I really need to get this job done. |
Does "classify one" show the expected result? Can you give numbers (accuracy on validation set v.s. accuracy during inference)? |
Greg - regarding the "classify one" and "classify many" , I have too many samples so that I tested a few of them randomly and the results are identical. It seems both produce the same accuracy. |
@gheinrich Any update to your end ? |
@gheinrich @rajulanand The issue came from cropping images by Caffe or Digits. For GoogleNet model, when you use 256x256 images, Caffe can implicitly handle cropping during training and validation somehow. But when you use the same samples for testing / predicition , it doesn't work. There are a few sources of the issue, in addition to the potential problem of cropping, the mean of training samples of 256x256 is different with the cropped images at 224 or cropping and it might not be fully correct. Because the mean is calculated for samples of 256 but is used for samples of 224. |
I am surprised that this has such a dramatic impact. That indicates that your network might not generalize well to new examples. Don't you have too much correlation between your training and validation sets? Either way, you might want to use mean pixel subtraction instead of mean image subtraction. |
Hi,
I have a problem when I use digits3.3dev.
I train a mode using GoogLeNet.
The graph show that accuracy go to 100% ,but when I use the "Test a list of images" on the same job's val.txt file. Result is all class in one category (my mode have 4 categories ,so the accuracy must be about 25%).
And I also use the same dataset and same parameters on digits2.3dev ,it go well. I don't know what is the different of digits2.3dev and digits3.3dev.
Can anyone help me ,please?
I have used "classification" under ~/digits/example/ to test, the result is different with "Test a list of images".
The text was updated successfully, but these errors were encountered: