-
Notifications
You must be signed in to change notification settings - Fork 496
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Development Roadmap (Deprecated) #157
Comments
Are there still plans for a high level pythonic interface? #39 (comment) |
Hi @AriMKatz, thanks for the reference. This is very important, I just added it. |
For the vision models support, is it possible to align with the openai gpt4v API? |
Are there plans for loading models in 8bit or 4bit? |
@nivibilla Yes, it is already aligned with the openai gpt4v API, see here. |
@aliencaocao Thanks for the question! The AWQ and GPTQ are already supported. But we do not support an automatic dtype translation yet. You are welcome to submit a PR for that. |
I'm looking to load llava 1.6 in 8bit, but it does not seem that llava series has AWQ or GPTQ quants, or did I miss out anything here? EDIT: I saw 1.5 has but not 1.6 yet. Perhaps its just too new and no one did a calibration yet. |
Hi all - is anyone working on the S-LoRA integration currently? I see the branch, but it looks a few months old. Would love to see this, happy to pick up from existing work or start fresh. |
Hi @qeternity, I was working on it but have been blocked by other affairs. You are welcome to contribute, either continue on the branch or start fresh! I'll be happy to review and collaborate. |
Tools support is very important, which is necessary for many use cases. |
Is TinyLlama supported? TinyLlama/TinyLlama-1.1B-Chat-v1.0 |
I see llama.cpp integration is on the roadmap. When will this feature be delivered? It would be very nice to have it , since it will support running local LLMs, such as llama models, on Mac computers and experiment them with the powerful and expressive SGLang. |
I'd request to include support for Phi-3-mini |
Hi which branch is it? looks like better start fresh |
I can help by getting rid of the vLLM in the dependencies. |
Moved to #634 |
Is the “tools” ok now? |
Function Calling
tools
argument insgl.gen
. See also guidance toolsHigh-level Pythonic Interface
Inference Optimizations
Structured Decoding
Compiler
sgl.fork
LoRA Support
Model Coverage
Device Coverage
The text was updated successfully, but these errors were encountered: