Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Is there anyway to train "big data" using transformer? #275

Open
zqp2009happy opened this issue Apr 1, 2020 · 1 comment
Open

Is there anyway to train "big data" using transformer? #275

zqp2009happy opened this issue Apr 1, 2020 · 1 comment
Labels
question Further information is requested

Comments

@zqp2009happy
Copy link

It sames that Transformer reads training data into the Memory. So it easily got OOM Error with "big training data" like 10G (about 50 million text pairs). Is there some solution for this problem?

@gpengzhi gpengzhi added the question Further information is requested label Apr 28, 2020
@ZhitingHu
Copy link
Member

By "Transformer" do you mean the example code under examples/ or the transformer modules in the library?

The transformer modules are independent of how you manage training data (either in memory or disk), as long as you pass to it a data minibatch each iteration.

The transformer example code does load the whole training data into the memory beforehand (code here). To avoid this, you may want to use Texar data module that loads data sequentially. Here is an example use of the Texar data module.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

3 participants