You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
vLLM sampling parameters include a richer set of values, among which logprobs has a wider utility.
When testing by adding the logpobs option to the request payload, the model output schema was unchanged
{"generated text": "model_output"} suggesting it was not propagated to the output
Will this change the current api? How?
Probably by enriching the output schema.
Who will benefit from this enhancement?
Anyone who wants logprobs extracted from model predictions.
References
list known implementations
This thread provides a starting point for tackling this issue.
The text was updated successfully, but these errors were encountered:
We have slightly different names for some of the generation/sampling parameters - our API unifies different inference backends like vllm, tensorrt-llm, huggingface accelerate, and transformers-neuronx.
Description
vLLM sampling parameters include a richer set of values, among which
logprobs
has a wider utility.When testing by adding the logpobs option to the request payload, the model output schema was unchanged
{"generated text": "model_output"} suggesting it was not propagated to the output
Will this change the current api? How?
Probably by enriching the output schema.
Who will benefit from this enhancement?
Anyone who wants logprobs extracted from model predictions.
References
This thread provides a starting point for tackling this issue.
The text was updated successfully, but these errors were encountered: