You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The next release of Universal Dependencies distinguishes between Eastern and Western Armenian. Is it suitable to use the word vectors you host for both dialects?
If I count the words from the UD datasets that are present in the Glove 200 file, for example, 97% of the words in the Eastern Armenian dataset appear here, and only 88% of the words in the Western Armenian dataset appear in the word vectors. This makes me think there's a bit of an issue with the coverage of these word vectors. If these vectors are not ideal, do you have any recommendations for others that would be, or do you have any intention of adding more Western Armenian words to the dataset?
Thanks!
The text was updated successfully, but these errors were encountered:
I'm working on preparing models for the next release of Stanford's Stanza python software, https://stanfordnlp.github.io/stanza
The next release of Universal Dependencies distinguishes between Eastern and Western Armenian. Is it suitable to use the word vectors you host for both dialects?
https://github.com/UniversalDependencies/UD_Armenian-ArmTDP/
https://github.com/UniversalDependencies/UD_Western_Armenian-ArmTDP/
If I count the words from the UD datasets that are present in the Glove 200 file, for example, 97% of the words in the Eastern Armenian dataset appear here, and only 88% of the words in the Western Armenian dataset appear in the word vectors. This makes me think there's a bit of an issue with the coverage of these word vectors. If these vectors are not ideal, do you have any recommendations for others that would be, or do you have any intention of adding more Western Armenian words to the dataset?
Thanks!
The text was updated successfully, but these errors were encountered: