From 571741097c1497bb428d5f6140fe5dfc00906dcc Mon Sep 17 00:00:00 2001 From: Erin Kavanaugh Date: Thu, 28 Mar 2024 13:34:14 -0700 Subject: [PATCH 1/2] update temp value --- openapi.yaml | 17 +++++++++++++++++ 1 file changed, 17 insertions(+) diff --git a/openapi.yaml b/openapi.yaml index 9433e6da..54f6db8a 100644 --- a/openapi.yaml +++ b/openapi.yaml @@ -8605,6 +8605,15 @@ components: type: object x-oaiTypeLabel: map nullable: true + temperature: + type: number + minimum: 0 + maximum: 2 + default: 1 + example: 1 + nullable: true + description: &run_temperature_description | + What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic. stream: type: boolean nullable: true @@ -8733,6 +8742,14 @@ components: type: object x-oaiTypeLabel: map nullable: true + temperature: + type: number + minimum: 0 + maximum: 2 + default: 1 + example: 1 + nullable: true + description: *run_temperature_description stream: type: boolean nullable: true From 9fcd900b58005549ab77d3e8af20f58473ce4d07 Mon Sep 17 00:00:00 2001 From: Erin Kavanaugh Date: Fri, 29 Mar 2024 12:08:36 -0700 Subject: [PATCH 2/2] pull latest changes in --- openapi.yaml | 40 ++++++++++++++++++++++++++++------------ 1 file changed, 28 insertions(+), 12 deletions(-) diff --git a/openapi.yaml b/openapi.yaml index 54f6db8a..d8d76d14 100644 --- a/openapi.yaml +++ b/openapi.yaml @@ -3766,7 +3766,8 @@ paths: "tools": [], "file_ids": [], "metadata": {}, - "usage": null + "usage": null, + "temperature": 1 } - title: Streaming @@ -4147,7 +4148,8 @@ paths: "prompt_tokens": 123, "completion_tokens": 456, "total_tokens": 579 - } + }, + "temperature": 1 }, { "id": "run_abc456", @@ -4178,7 +4180,8 @@ paths: "prompt_tokens": 123, "completion_tokens": 456, "total_tokens": 579 - } + }, + "temperature": 1 } ], "first_id": "run_abc123", @@ -4277,7 +4280,8 @@ paths: "file-abc456" ], "metadata": {}, - "usage": null + "usage": null, + "temperature": 1 } - title: Streaming request: @@ -4612,7 +4616,8 @@ paths: "prompt_tokens": 123, "completion_tokens": 456, "total_tokens": 579 - } + }, + "temperature": 1 } post: operationId: modifyRun @@ -4725,7 +4730,8 @@ paths: "prompt_tokens": 123, "completion_tokens": 456, "total_tokens": 579 - } + }, + "temperature": 1 } /threads/{thread_id}/runs/{run_id}/submit_tool_outputs: @@ -4862,7 +4868,8 @@ paths: ], "file_ids": [], "metadata": {}, - "usage": null + "usage": null, + "temperature": 1 } - title: Streaming @@ -5061,7 +5068,8 @@ paths: ], "file_ids": [], "metadata": {}, - "usage": null + "usage": null, + "temperature": 1 } /threads/{thread_id}/runs/{run_id}/steps: @@ -8522,6 +8530,10 @@ components: nullable: true usage: $ref: "#/components/schemas/RunCompletionUsage" + temperature: + description: The sampling temperature used for this run. If not set, defaults to 1. + type: number + nullable: true required: - id - object @@ -8568,7 +8580,8 @@ components: "prompt_tokens": 123, "completion_tokens": 456, "total_tokens": 579 - } + }, + "temperature": 1 } CreateRunRequest: type: object @@ -8918,7 +8931,7 @@ components: type: string nullable: true run_id: - description: If applicable, the ID of the [run](/docs/api-reference/runs) associated with the authoring of this message. + description: The ID of the [run](/docs/api-reference/runs) associated with the creation of this message. Value is `null` when messages are created manually using the create message or create thread endpoints. type: string nullable: true file_ids: @@ -9041,8 +9054,11 @@ components: properties: role: type: string - enum: ["user"] - description: The role of the entity that is creating the message. Currently only `user` is supported. + enum: ["user", "assistant"] + description: | + The role of the entity that is creating the message. Allowed values include: + - `user`: Indicates the message is sent by an actual user and should be used in most cases to represent user-generated messages. + - `assistant`: Indicates the message is generated by the assistant. Use this value to insert messages from the assistant into the conversation. content: type: string minLength: 1