Skip to content

Does it make sense to change the # hidden layers of a pretrained transformer? #12264

Discussion options

You must be logged in to vote

Hi @bhartm3, no, it doesn't make sense to do that. At this point the model architecture is fixed.

Replies: 1 comment

Comment options

You must be logged in to vote
0 replies
Answer selected by bhartm3
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
feat / training Feature: Training utils, Example, Corpus and converters feat / transformer Feature: Transformer
2 participants