Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support BERT finetuned on SQuAD #1

Open
Pierrci opened this issue Nov 18, 2019 · 3 comments
Open

Support BERT finetuned on SQuAD #1

Pierrci opened this issue Nov 18, 2019 · 3 comments
Assignees
Labels
enhancement New feature or request

Comments

@Pierrci
Copy link
Member

Pierrci commented Nov 18, 2019

Since the tokenizer is the same than MobileBERT/DistilBERT, would be pretty straightforward to add once this TensorFlow issue is solved: tensorflow/tensorflow#34210

@Pierrci Pierrci added the enhancement New feature or request label Nov 18, 2019
@Pierrci Pierrci self-assigned this Nov 18, 2019
@ucalyptus
Copy link

@Pierrci is this closing now as 34210 is closed?

@Pierrci
Copy link
Member Author

Pierrci commented Jan 6, 2020

The non-quantized TFLite version is around 1GB so way too big for a mobile app.
I'll close this once the FP16 quantization works so we can use a model with a reduced size and good performance, but it's not the case for now (at least when I tried last Friday)

@csarron
Copy link

csarron commented Mar 24, 2020

Hi @Pierrci, were you able to quantize the BERT for TFLite? I tried a few options but failed to get the quantized model.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

3 participants