-
Notifications
You must be signed in to change notification settings - Fork 88
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Problem of reproduce Camelyon16 result #54
Comments
Any update? I have the same training curve for camelyon simclr as @sy2es94098 that gets stuck at around 5 during training. Unfortunately I can only fit 280 images in batch size as I have only 2x 2080tis as gpus. |
Based on the loss curve of SimCLR. I think you should let the model further converge. Make sure you have loaded the model weights correctly, and that the type of the normalization layer is consistent. |
Hello, I would like to ask why all my test results are black when I am trying to draw attention maps. Thank you |
You can try two things:
|
How was the ground truth of tumor_026.tif obtained? |
I incorporated the training/testing into the same pipeline in the latest commit. This change allows you to read the evaluation results on a reserved test set. I also incorporated a simple weights initialization method which helps stabilize the training. You can set --eval_scheme=5-fold-cv-standalone-test which will perform a train/valid/test like this:
You can also simply run a 5-fold cv --eval_scheme=5-fold-cv There were some issues with the testing script when loading pretrained weights (i.e., sometimes the weights are not fully loaded or there are missing weights, setting strict=False can reveal the problems.). The purpose of the testing script is to generate the heatmap, you should now read the performance directly from the training script. I will fix the issues in a couple of days. |
Hello, thank you for your excellent work!
Earlier I tried to reproduce the results of Camelyon16, I used a total of 271 training sets, batch size 512 to train simclr for 3 days and train the aggregator, but the results are not as good as the simclr weights you provided for 3 days of training (model-v0 in google drive).
Like #46 , the accuracy of the aggregator will be stuck at about 60% and cannot be improved, and I found that in this case, each patch will produce the same attention score.
Can you provide relevant training parameters, such as the -o or -t parameters in deepzoom_tiler.py, and the learning rate, batch size, epoch, etc. of simclr.
The text was updated successfully, but these errors were encountered: