You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/parameters.md
+1Lines changed: 1 addition & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -48,6 +48,7 @@ When using FastDeploy to deploy models (including offline inference and service
48
48
|```dynamic_load_weight```|`int`| Whether to enable dynamic weight loading, default: 0 |
49
49
|```enable_expert_parallel```|`bool`| Whether to enable expert parallel |
50
50
|```enable_logprob```|`bool`| Whether to enable return log probabilities of the output tokens or not. If true, returns the log probabilities of each output token returned in the content of message.If logrpob is not used, this parameter can be omitted when starting |
51
+
|```logprobs_mode```|`str`| Indicates the content returned in the logprobs. Supported mode: `raw_logprobs`, `processed_logprobs`, `raw_logits`, `processed_logits`. Raw means the values before applying logit processors, like bad words. Processed means the values after applying such processors. |
51
52
|```served_model_name```|`str`| The model name used in the API. If not specified, the model name will be the same as the --model argument |
52
53
|```revision```|`str`| The specific model version to use. It can be a branch name, a tag name, or a commit id. If unspecified, will use the default version. |
53
54
|```chat_template```|`str`| Specify the template used for model concatenation, It supports both string input and file path input. The default value is None. If not specified, the model's default template will be used. |
0 commit comments