You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
any guide/code on how can I do continue pre-training on nomic v1.5 to align with a specific domain area? like finance or medicine.
will just do Masked Language Modeling Pretraining on medical data works in this case? (based on this paper's idea: https://arxiv.org/pdf/2004.10964)
since Nomic embed have additional Unsupervised and Supervised Contrastive Pretraining, I'm not sure if I can still do additional masked langauge modeling pre-training on a domain data........
Best,
Keyu
The text was updated successfully, but these errors were encountered:
keyuchen21
changed the title
continue pretraining nomic
continue pretraining nomic embed on domain data
Dec 18, 2024
Best,
Keyu
The text was updated successfully, but these errors were encountered: