Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Thanks! #24

Open
srush opened this issue Apr 25, 2020 · 11 comments
Open

Thanks! #24

srush opened this issue Apr 25, 2020 · 11 comments

Comments

@srush
Copy link

srush commented Apr 25, 2020

Thanks so much for this repo it has been amazingly useful. We used it to build ICLR 2020 virtual addition and added all the pictures this way!

https://twitter.com/srush_nlp/status/1253788694739386371

@zhxgj
Copy link
Contributor

zhxgj commented Apr 25, 2020

Hi @srush, Thanks for using PubLayNet. Glad to hear it helped!

@zhxgj
Copy link
Contributor

zhxgj commented Apr 26, 2020

Hi @srush We happen to be preparing a blog about PubLayNet and want to add this great news to the blog. Do you have an estimate of how much time you think PubLayNet saved you? Any type of metric will be greatly useful for us. Thanks very much.

@srush
Copy link
Author

srush commented Apr 26, 2020

Infinity time, I would have given up. I tried every other direct PDF extraction method and they all had intractable issues, e.g. couldn't extract PDF images or were too low res or were just bad. I was about to give up until I found this tool, and in 20 lines of code, and 2 hours on Google Colab (sorry) , I had every image from 700 papers at high res with precise enough accuracy (it's unfortunately bad at columns? Guessing that is because of pubmed).

@zhxgj
Copy link
Contributor

zhxgj commented Apr 27, 2020

Thanks @srush for your feedback. Do you have any examples where the model is bad at columns? I can have a look if that is because of any bias or annotation errors in the data, if I can fix it.

@srush
Copy link
Author

srush commented Apr 27, 2020 via email

@zhxgj
Copy link
Contributor

zhxgj commented Apr 27, 2020

Ah, yes. The dataset does not have many samples with text wraps around images. Most journals do not typeset in that way. And I also think our automated annotation algorithm does not handle this case well, and poor annotations are excluded, which further reduces samples with this appearance.

@zhxgj
Copy link
Contributor

zhxgj commented Apr 27, 2020

Infinity time, I would have given up. I tried every other direct PDF extraction method and they all had intractable issues, e.g. couldn't extract PDF images or were too low res or were just bad. I was about to give up until I found this tool, and in 20 lines of code, and 2 hours on Google Colab (sorry) , I had every image from 700 papers at high res with precise enough accuracy (it's unfortunately bad at columns? Guessing that is because of pubmed).

Hi @srush , could I please quote your above feedback in our blog post?

@srush
Copy link
Author

srush commented Apr 27, 2020 via email

@zhxgj
Copy link
Contributor

zhxgj commented Apr 27, 2020

Thanks @srush Yes, it is a good idea to have a small fine-tuning set for a specific template. The set can be pre-annotated with our model then manually curated, which will save some time.

@tshrjn
Copy link

tshrjn commented Jun 9, 2021

Hi there,

This is great!
However, I just wanted to clarify why something like pyMupdf or pdfminer (& it's cli script for images) couldn't be used?

They seem straightforward. Am I missing something?
Does publaynet do something additional like extract main image or so?

@srush
Copy link
Author

srush commented Jun 9, 2021

Publaynet extracts the images as they are shown in the paper (cropped, captioned etc) which is more interesting and harder.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants