Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Request: New LM Adapted checkpoints for T5 #12384

Open
3 tasks done
Xirider opened this issue Jun 27, 2021 · 0 comments
Open
3 tasks done

Request: New LM Adapted checkpoints for T5 #12384

Xirider opened this issue Jun 27, 2021 · 0 comments

Comments

@Xirider
Copy link

Xirider commented Jun 27, 2021

🌟 New LM Adapted checkpoints for T5

Description

Google released a new set of checkpoints for T5 v1.1. here:
https://github.com/google-research/text-to-text-transfer-transformer/blob/master/released_checkpoints.md#t511

Especially interesting for most people will be the checkpoints below, as finetuning T5 with a prompt or using T5 for conditional text generation is quite common and these checkpoints promise better performance. The default T5 v1.1 checkpoints have never seen sequences without sentinel tokens.

LM-Adapted: t5.1.1.lm100k (copied from the readme)

These "LM adapted" models are initialized from t5.1.1 (above) and train for an
additional 100K steps on the LM objective discussed in the [T5 paper][paper].
This adaptation improves the ability of the model to be used for prompt
tuning
.

Open source status

  • the model implementation is available: t5 v1.1. with geglu
  • the model weights are available: see links above
  • who are the authors: Brian Lester, Rami Al-Rfou, Noah Constant
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

1 participant