增加环绕侦察场景适配

This commit is contained in:
2026-01-08 15:44:38 +08:00
parent 3eba1f962b
commit 10c5bb5a8a
5441 changed files with 40219 additions and 379695 deletions

View File

@@ -125,8 +125,10 @@ class Calls(SyncAPIResource):
"gpt-4o-mini-realtime-preview-2024-12-17",
"gpt-realtime-mini",
"gpt-realtime-mini-2025-10-06",
"gpt-realtime-mini-2025-12-15",
"gpt-audio-mini",
"gpt-audio-mini-2025-10-06",
"gpt-audio-mini-2025-12-15",
],
]
| Omit = omit,
@@ -199,15 +201,20 @@ class Calls(SyncAPIResource):
limit, the conversation be truncated, meaning messages (starting from the
oldest) will not be included in the model's context. A 32k context model with
4,096 max output tokens can only include 28,224 tokens in the context before
truncation occurs. Clients can configure truncation behavior to truncate with a
lower max token limit, which is an effective way to control token usage and
cost. Truncation will reduce the number of cached tokens on the next turn
(busting the cache), since messages are dropped from the beginning of the
context. However, clients can also configure truncation to retain messages up to
a fraction of the maximum context size, which will reduce the need for future
truncations and thus improve the cache rate. Truncation can be disabled
entirely, which means the server will never truncate but would instead return an
error if the conversation exceeds the model's input token limit.
truncation occurs.
Clients can configure truncation behavior to truncate with a lower max token
limit, which is an effective way to control token usage and cost.
Truncation will reduce the number of cached tokens on the next turn (busting the
cache), since messages are dropped from the beginning of the context. However,
clients can also configure truncation to retain messages up to a fraction of the
maximum context size, which will reduce the need for future truncations and thus
improve the cache rate.
Truncation can be disabled entirely, which means the server will never truncate
but would instead return an error if the conversation exceeds the model's input
token limit.
extra_headers: Send extra headers
@@ -445,8 +452,10 @@ class AsyncCalls(AsyncAPIResource):
"gpt-4o-mini-realtime-preview-2024-12-17",
"gpt-realtime-mini",
"gpt-realtime-mini-2025-10-06",
"gpt-realtime-mini-2025-12-15",
"gpt-audio-mini",
"gpt-audio-mini-2025-10-06",
"gpt-audio-mini-2025-12-15",
],
]
| Omit = omit,
@@ -519,15 +528,20 @@ class AsyncCalls(AsyncAPIResource):
limit, the conversation be truncated, meaning messages (starting from the
oldest) will not be included in the model's context. A 32k context model with
4,096 max output tokens can only include 28,224 tokens in the context before
truncation occurs. Clients can configure truncation behavior to truncate with a
lower max token limit, which is an effective way to control token usage and
cost. Truncation will reduce the number of cached tokens on the next turn
(busting the cache), since messages are dropped from the beginning of the
context. However, clients can also configure truncation to retain messages up to
a fraction of the maximum context size, which will reduce the need for future
truncations and thus improve the cache rate. Truncation can be disabled
entirely, which means the server will never truncate but would instead return an
error if the conversation exceeds the model's input token limit.
truncation occurs.
Clients can configure truncation behavior to truncate with a lower max token
limit, which is an effective way to control token usage and cost.
Truncation will reduce the number of cached tokens on the next turn (busting the
cache), since messages are dropped from the beginning of the context. However,
clients can also configure truncation to retain messages up to a fraction of the
maximum context size, which will reduce the need for future truncations and thus
improve the cache rate.
Truncation can be disabled entirely, which means the server will never truncate
but would instead return an error if the conversation exceeds the model's input
token limit.
extra_headers: Send extra headers

View File

@@ -232,7 +232,7 @@ class AsyncRealtimeWithStreamingResponse:
class AsyncRealtimeConnection:
"""Represents a live websocket connection to the Realtime API"""
"""Represents a live WebSocket connection to the Realtime API"""
session: AsyncRealtimeSessionResource
response: AsyncRealtimeResponseResource
@@ -421,7 +421,7 @@ class AsyncRealtimeConnectionManager:
class RealtimeConnection:
"""Represents a live websocket connection to the Realtime API"""
"""Represents a live WebSocket connection to the Realtime API"""
session: RealtimeSessionResource
response: RealtimeResponseResource
@@ -829,7 +829,7 @@ class RealtimeConversationItemResource(BaseRealtimeConnectionResource):
class RealtimeOutputAudioBufferResource(BaseRealtimeConnectionResource):
def clear(self, *, event_id: str | Omit = omit) -> None:
"""**WebRTC Only:** Emit to cut off the current audio response.
"""**WebRTC/SIP Only:** Emit to cut off the current audio response.
This will trigger the server to
stop generating audio and emit a `output_audio_buffer.cleared` event. This
@@ -1066,7 +1066,7 @@ class AsyncRealtimeConversationItemResource(BaseAsyncRealtimeConnectionResource)
class AsyncRealtimeOutputAudioBufferResource(BaseAsyncRealtimeConnectionResource):
async def clear(self, *, event_id: str | Omit = omit) -> None:
"""**WebRTC Only:** Emit to cut off the current audio response.
"""**WebRTC/SIP Only:** Emit to cut off the current audio response.
This will trigger the server to
stop generating audio and emit a `output_audio_buffer.cleared` event. This