Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Method for Faster Inference #614

Closed
fakerybakery opened this issue Oct 1, 2023 · 5 comments
Closed

Method for Faster Inference #614

fakerybakery opened this issue Oct 1, 2023 · 5 comments

Comments

@fakerybakery
Copy link
Contributor

Hi, might it be possible to merge this repository upstream? It's slightly faster than the original version. Thanks!

@jhowilbur
Copy link

@fakerybakery, can you help me understand the main point of making this other repo faster?
I tried to find it here in the commits, but I need help finding out why.

@fakerybakery
Copy link
Contributor Author

fakerybakery commented Oct 13, 2023

Hi @jhowilbur
In the API repo, the authors mentioned that in their fork they made it 30% faster:

We have made some improvements to Tortoise to make the inference ~30% faster, and welcome contributions on our repo to improve it further!

Source

The commit messages say they made it faster by not moving the models between CPU and GPU.

Hope this helps!

@fakerybakery
Copy link
Contributor Author

Hi @manmay-nakhashi, has this been resolved?

@manmay-nakhashi
Copy link
Collaborator

manmay-nakhashi commented Oct 19, 2023

Yes we don't do offloading to cpu in api_fast.py

@fakerybakery
Copy link
Contributor Author

The api_fast.py can't work on MPS (#658) since the fast API requires autocast and autocast isn't yet supported on MPS.

Yes we don't do offloading to cpu in api_fast.py

Does this mean support for the regular API is not planned @manmay-nakhashi?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants