-
Notifications
You must be signed in to change notification settings - Fork 23
Closed
Description
Now that the llama.cpp server is running correctly, would it be possible to have this model tested?
https://huggingface.co/Infinimol/miiqu-gguf
using ChatML format and context length >= 1024, please :)
It is a model I been working on it for some time, and I think it's interesting. It is not a fine-tune, but a merge, and I find it consistently scores higher than the base model (miqu), which I think is a first for a pure merge model. Eq-bench runs in about 15 mins on an A100.
The model is GGUF, but split to fit under the 50Gb limit on Huggingface, but the model card give the one-liner to reassemble the file.
sam-paech
Metadata
Metadata
Assignees
Labels
No labels