You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We try to reproduce arxiv task with f1 92 as shown in the paper, we are using default hyperparameters defined in bigbird/classifier/base_size.sh, pretrained checkpoint here, but with batch size = 2 due to memory limitation (total batch size = 8gpu * 2 = 16), after 16k steps (16000 * 16 / 30034 = 8.5 epoch), but only get f1 84 in the end, which is too low compare to the paper which is trained by 10 epochs.
Did we missing something? preprocessing of Arxiv? or just because of the batch size is too small?
Will you release the checkpoint of Arxiv in the future?
About the difference of dataset, we have finetune roberta on the same arxiv dataset and get f1 86, pretty close the the paper.
The text was updated successfully, but these errors were encountered:
liuyang148
changed the title
how to reproduce arxiv classification task
reproduce arxiv classification task
Aug 18, 2021
Yes, I am experiencing a similar result. Maybe the authors forgot to remove the leaked labels in the original data of the text from the scraped PDF text? (i.e., some samples have the label directly in the text, so classifying it is trivial)
We try to reproduce arxiv task with
f1 92
as shown in the paper, we are using default hyperparameters defined inbigbird/classifier/base_size.sh
, pretrained checkpoint here, but withbatch size = 2
due to memory limitation (total batch size = 8gpu * 2 = 16), after 16k steps (16000 * 16 / 30034 = 8.5 epoch), but only getf1 84
in the end, which is too low compare to the paper which is trained by 10 epochs.Did we missing something? preprocessing of Arxiv? or just because of the batch size is too small?
Will you release the checkpoint of Arxiv in the future?
About the difference of dataset, we have finetune roberta on the same arxiv dataset and get
f1 86
, pretty close the the paper.The text was updated successfully, but these errors were encountered: