external help file | Module Name | online version | schema |
---|---|---|---|
PSOpenAI-help.xml |
PSOpenAI |
2.0.0 |
Set the realtime session's configuration.
Set-RealtimeSessionConfiguration
[-EventId <String>]
[-Instructions <String>]
[-Modalities <String[]>]
[-Voice <String>]
[-InputAudioFormat <String>]
[-OutputAudioFormat <String>]
[-EnableInputAudioTranscription <Boolean>]
[-InputAudioTranscriptionModel <String>]
[-InputAudioTranscriptionLanguage <String>]
[-InputAudioTranscriptionPrompt <String>]
[-EnableTurnDetection <Boolean>]
[-TurnDetectionType <String>]
[-TurnDetectionThreshold <Single>]
[-TurnDetectionPrefixPadding <UInt16>]
[-TurnDetectionSilenceDuration <UInt16>]
[-CreateResponseOnTurnEnd <Boolean>]
[-InterruptResponse <Boolean>]
[-Tools <IDictionary[]>]
[-ToolChoice <String>]
[-Temperature <Single>]
[-MaxResponseOutputTokens <Int32>]
Set the realtime session's configuration.
PS C:\> Set-RealtimeSessionConfiguration `
-Modalities 'text','audio' `
-Voice 'shimmer' `
-EnableInputAudioTranscription $true `
-EnableTurnDetection $true
-Temperature 1.0
Enables input audio transcription.
Type: Boolean
Required: False
Position: Named
Enables the server VAD mode. In this mode, the server will run voice activity detection (VAD) over the incoming audio and respond after the end of speech.
Type: Boolean
Required: False
Position: Named
Optional client-generated ID used to identify this event.
Type: String
Required: False
Position: Named
The format of input audio. Options are pcm16
, g711_ulaw
, or g711_alaw
.
Type: String
Required: False
Position: Named
The model to use for transcription, whisper-1
is the only currently supported model.
Type: String
Required: False
Position: Named
Default value: whisper-1
The language of the input audio. Supplying the input language in ISO-639-1 (e.g. en) format will improve accuracy and latency.
Type: String
Required: False
Position: Named
An optional text to guide the model's style or continue a previous audio segment. The prompt should match the audio language.
Type: String
Required: False
Position: Named
The default system instructions (i.e. system message) prepended to model calls. This field allows the client to guide the model on desired responses.
Type: String
Required: False
Position: Named
Maximum number of output tokens for a single assistant response. Provide an integer between 1 and 4096 to limit output tokens, or -1 for no limitations.
Type: Int32
Required: False
Position: Named
Default value: -1
The set of modalities the model can respond with.
Type: String[]
Accepted values: text, audio
Required: False
Position: Named
The format of output audio. Options are pcm16
, g711_ulaw
, or g711_alaw
.
Type: String
Required: False
Position: Named
Sampling temperature for the model, limited to [0.6, 1.2].
Type: Single
Required: False
Position: Named
How the model chooses tools. Options are auto
, none
, required
, or specify a function.
Type: String
Required: False
Position: Named
Tools (functions) available to the model.
Type: IDictionary[]
Required: False
Position: Named
Amount of audio to include before the VAD detected speech (in milliseconds).
Type: UInt16
Required: False
Position: Named
Duration of silence to detect speech stop (in milliseconds). With shorter values the model will respond more quickly, but may jump in on short pauses from the user.
Type: UInt16
Required: False
Position: Named
Activation threshold for VAD (0.0 to 1.0), this defaults to 0.5. A higher threshold will require louder audio to activate the model, and thus might perform better in noisy environments.
Type: Single
Required: False
Position: Named
Type of turn detection, only server_vad
is currently supported.
Type: String
Required: False
Position: Named
Default value: server_vad
Whether or not to automatically generate a response when VAD is enabled. true by default.
Type: Boolean
Required: False
Position: Named
Default value: True
Whether or not to automatically interrupt any ongoing response with output to the default conversation when a VAD start event occurs. true by default.
Type: Boolean
Required: False
Position: Named
Default value: True
The voice the model uses to respond. Cannot be changed once the model has responded with audio at least once.
Type: String
Required: False
Position: Named
https://platform.openai.com/docs/api-reference/realtime-client-events/session/update