You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Some code already exists for processing newspapers at scale, on the H-Top repo under :
/H-Top/generate_toponym_dataset/
In particular, the script apply_to_news_modular.py appears to be the template for the example given in the docs on running T-Res at scale.
The task here is to build something like this into the API.
Must be able to process a csv file of input data (one row per article text) as in the apply_to_news_modular.py script
If necessary, should also support a zip file of input data (one txt file per article text, structured in the usual way e.g. 0003548/1904/0616/0003548_19040616_art0053.txt). This may or may not be required, depending on whether the complete open newspapers collections (HMD & LwM) are already available in csv format (above).
The text was updated successfully, but these errors were encountered:
PR open into
276-refactor
.Some code already exists for processing newspapers at scale, on the H-Top repo under :
In particular, the script
apply_to_news_modular.py
appears to be the template for the example given in the docs on running T-Res at scale.The task here is to build something like this into the API.
Must be able to process a csv file of input data (one row per article text) as in the
apply_to_news_modular.py
scriptIf necessary, should also support a zip file of input data (one txt file per article text, structured in the usual way e.g.
0003548/1904/0616/0003548_19040616_art0053.txt
). This may or may not be required, depending on whether the complete open newspapers collections (HMD & LwM) are already available in csv format (above).The text was updated successfully, but these errors were encountered: