Training sharing transformer layer #7513
fcggamou
started this conversation in
Help: Best practices
Replies: 1 comment
-
If you have data annotated with multiple tasks, there is no reason why you can't train the transformer simultaneously on all these tasks. You can even set the If you're training in multiple stages, then yes one solution is to first train a transformer+component, then freeze and train a second component on top of the frozen transformer. It's difficult to say upfront what the effect on accuracy would be though. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
What is the recommended approach to train multiple pipeline components that would share the same transformer layer?
Should I freeze the transformer layer after training one component and then train the rest of them?
Thanks for any feedback
Beta Was this translation helpful? Give feedback.
All reactions